From ec88aa967212fbfe996048d8aba3beccafd68f71 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Thu, 24 Oct 2024 06:32:18 -0400 Subject: [PATCH 001/591] fix(profiling): Update active thread for asgi (#3669) Ensure the handling thread is set on the transaction for asgi transactions not just main thread. --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/django/asgi.py | 4 + sentry_sdk/integrations/django/views.py | 4 + sentry_sdk/integrations/fastapi.py | 5 + sentry_sdk/integrations/quart.py | 13 +- sentry_sdk/integrations/starlette.py | 5 +- sentry_sdk/tracing.py | 8 +- tests/integrations/django/asgi/test_asgi.py | 31 +++-- tests/integrations/fastapi/test_fastapi.py | 14 +- tests/integrations/quart/test_quart.py | 121 +++++++++++------- .../integrations/starlette/test_starlette.py | 14 +- 10 files changed, 150 insertions(+), 69 deletions(-) diff --git a/sentry_sdk/integrations/django/asgi.py b/sentry_sdk/integrations/django/asgi.py index 71b69a9bc1..73a25acc9f 100644 --- a/sentry_sdk/integrations/django/asgi.py +++ b/sentry_sdk/integrations/django/asgi.py @@ -172,6 +172,10 @@ def wrap_async_view(callback): @functools.wraps(callback) async def sentry_wrapped_callback(request, *args, **kwargs): # type: (Any, *Any, **Any) -> Any + current_scope = sentry_sdk.get_current_scope() + if current_scope.transaction is not None: + current_scope.transaction.update_active_thread() + sentry_scope = sentry_sdk.get_isolation_scope() if sentry_scope.profile is not None: sentry_scope.profile.update_active_thread_id() diff --git a/sentry_sdk/integrations/django/views.py b/sentry_sdk/integrations/django/views.py index cb81d3555c..0a9861a6a6 100644 --- a/sentry_sdk/integrations/django/views.py +++ b/sentry_sdk/integrations/django/views.py @@ -76,6 +76,10 @@ def _wrap_sync_view(callback): @functools.wraps(callback) def sentry_wrapped_callback(request, *args, **kwargs): # type: (Any, *Any, **Any) -> Any + current_scope = sentry_sdk.get_current_scope() + if current_scope.transaction is not None: + current_scope.transaction.update_active_thread() + sentry_scope = sentry_sdk.get_isolation_scope() # set the active thread id to the handler thread for sync views # this isn't necessary for async views since that runs on main diff --git a/sentry_sdk/integrations/fastapi.py b/sentry_sdk/integrations/fastapi.py index c3816b6565..8877925a36 100644 --- a/sentry_sdk/integrations/fastapi.py +++ b/sentry_sdk/integrations/fastapi.py @@ -88,9 +88,14 @@ def _sentry_get_request_handler(*args, **kwargs): @wraps(old_call) def _sentry_call(*args, **kwargs): # type: (*Any, **Any) -> Any + current_scope = sentry_sdk.get_current_scope() + if current_scope.transaction is not None: + current_scope.transaction.update_active_thread() + sentry_scope = sentry_sdk.get_isolation_scope() if sentry_scope.profile is not None: sentry_scope.profile.update_active_thread_id() + return old_call(*args, **kwargs) dependant.call = _sentry_call diff --git a/sentry_sdk/integrations/quart.py b/sentry_sdk/integrations/quart.py index ac58f21175..51306bb4cd 100644 --- a/sentry_sdk/integrations/quart.py +++ b/sentry_sdk/integrations/quart.py @@ -1,6 +1,5 @@ import asyncio import inspect -import threading from functools import wraps import sentry_sdk @@ -122,11 +121,13 @@ def decorator(old_func): @ensure_integration_enabled(QuartIntegration, old_func) def _sentry_func(*args, **kwargs): # type: (*Any, **Any) -> Any - scope = sentry_sdk.get_isolation_scope() - if scope.profile is not None: - scope.profile.active_thread_id = ( - threading.current_thread().ident - ) + current_scope = sentry_sdk.get_current_scope() + if current_scope.transaction is not None: + current_scope.transaction.update_active_thread() + + sentry_scope = sentry_sdk.get_isolation_scope() + if sentry_scope.profile is not None: + sentry_scope.profile.update_active_thread_id() return old_func(*args, **kwargs) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index 03584fdad7..52c64f6843 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -487,8 +487,11 @@ def _sentry_sync_func(*args, **kwargs): if integration is None: return old_func(*args, **kwargs) - sentry_scope = sentry_sdk.get_isolation_scope() + current_scope = sentry_sdk.get_current_scope() + if current_scope.transaction is not None: + current_scope.transaction.update_active_thread() + sentry_scope = sentry_sdk.get_isolation_scope() if sentry_scope.profile is not None: sentry_scope.profile.update_active_thread_id() diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 7ce577b1d0..3868b2e6c8 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -329,8 +329,7 @@ def __init__( self._span_recorder = None # type: Optional[_SpanRecorder] self._local_aggregator = None # type: Optional[LocalAggregator] - thread_id, thread_name = get_current_thread_meta() - self.set_thread(thread_id, thread_name) + self.update_active_thread() self.set_profiler_id(get_profiler_id()) # TODO this should really live on the Transaction class rather than the Span @@ -732,6 +731,11 @@ def get_profile_context(self): "profiler_id": profiler_id, } + def update_active_thread(self): + # type: () -> None + thread_id, thread_name = get_current_thread_meta() + self.set_thread(thread_id, thread_name) + class Transaction(Span): """The Transaction is the root element that holds all the spans diff --git a/tests/integrations/django/asgi/test_asgi.py b/tests/integrations/django/asgi/test_asgi.py index f6cfae0d2c..063aed63ad 100644 --- a/tests/integrations/django/asgi/test_asgi.py +++ b/tests/integrations/django/asgi/test_asgi.py @@ -104,14 +104,16 @@ async def test_async_views(sentry_init, capture_events, application): @pytest.mark.skipif( django.VERSION < (3, 1), reason="async views have been introduced in Django 3.1" ) -async def test_active_thread_id(sentry_init, capture_envelopes, endpoint, application): +async def test_active_thread_id( + sentry_init, capture_envelopes, teardown_profiling, endpoint, application +): with mock.patch( "sentry_sdk.profiler.transaction_profiler.PROFILE_MINIMUM_SAMPLES", 0 ): sentry_init( integrations=[DjangoIntegration()], traces_sample_rate=1.0, - _experiments={"profiles_sample_rate": 1.0}, + profiles_sample_rate=1.0, ) envelopes = capture_envelopes() @@ -121,17 +123,26 @@ async def test_active_thread_id(sentry_init, capture_envelopes, endpoint, applic await comm.wait() assert response["status"] == 200, response["body"] - assert len(envelopes) == 1 - profiles = [item for item in envelopes[0].items if item.type == "profile"] - assert len(profiles) == 1 + assert len(envelopes) == 1 + + profiles = [item for item in envelopes[0].items if item.type == "profile"] + assert len(profiles) == 1 + + data = json.loads(response["body"]) + + for item in profiles: + transactions = item.payload.json["transactions"] + assert len(transactions) == 1 + assert str(data["active"]) == transactions[0]["active_thread_id"] - data = json.loads(response["body"]) + transactions = [item for item in envelopes[0].items if item.type == "transaction"] + assert len(transactions) == 1 - for profile in profiles: - transactions = profile.payload.json["transactions"] - assert len(transactions) == 1 - assert str(data["active"]) == transactions[0]["active_thread_id"] + for item in transactions: + transaction = item.payload.json + trace_context = transaction["contexts"]["trace"] + assert str(data["active"]) == trace_context["data"]["thread.id"] @pytest.mark.asyncio diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index 93d048c029..97aea06344 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -184,7 +184,7 @@ def test_legacy_setup( def test_active_thread_id(sentry_init, capture_envelopes, teardown_profiling, endpoint): sentry_init( traces_sample_rate=1.0, - _experiments={"profiles_sample_rate": 1.0}, + profiles_sample_rate=1.0, ) app = fastapi_app_factory() asgi_app = SentryAsgiMiddleware(app) @@ -203,11 +203,19 @@ def test_active_thread_id(sentry_init, capture_envelopes, teardown_profiling, en profiles = [item for item in envelopes[0].items if item.type == "profile"] assert len(profiles) == 1 - for profile in profiles: - transactions = profile.payload.json["transactions"] + for item in profiles: + transactions = item.payload.json["transactions"] assert len(transactions) == 1 assert str(data["active"]) == transactions[0]["active_thread_id"] + transactions = [item for item in envelopes[0].items if item.type == "transaction"] + assert len(transactions) == 1 + + for item in transactions: + transaction = item.payload.json + trace_context = transaction["contexts"]["trace"] + assert str(data["active"]) == trace_context["data"]["thread.id"] + @pytest.mark.asyncio async def test_original_request_not_scrubbed(sentry_init, capture_events): diff --git a/tests/integrations/quart/test_quart.py b/tests/integrations/quart/test_quart.py index 321f07e3c6..f15b968ac5 100644 --- a/tests/integrations/quart/test_quart.py +++ b/tests/integrations/quart/test_quart.py @@ -1,8 +1,8 @@ import json import threading +from unittest import mock import pytest -import pytest_asyncio import sentry_sdk from sentry_sdk import ( @@ -28,8 +28,7 @@ auth_manager = AuthManager() -@pytest_asyncio.fixture -async def app(): +def quart_app_factory(): app = Quart(__name__) app.debug = False app.config["TESTING"] = False @@ -73,8 +72,9 @@ def integration_enabled_params(request): @pytest.mark.asyncio -async def test_has_context(sentry_init, app, capture_events): +async def test_has_context(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() client = app.test_client() @@ -99,7 +99,6 @@ async def test_has_context(sentry_init, app, capture_events): ) async def test_transaction_style( sentry_init, - app, capture_events, url, transaction_style, @@ -111,6 +110,7 @@ async def test_transaction_style( quart_sentry.QuartIntegration(transaction_style=transaction_style) ] ) + app = quart_app_factory() events = capture_events() client = app.test_client() @@ -126,10 +126,10 @@ async def test_errors( sentry_init, capture_exceptions, capture_events, - app, integration_enabled_params, ): sentry_init(**integration_enabled_params) + app = quart_app_factory() @app.route("/") async def index(): @@ -153,9 +153,10 @@ async def index(): @pytest.mark.asyncio async def test_quart_auth_not_installed( - sentry_init, app, capture_events, monkeypatch, integration_enabled_params + sentry_init, capture_events, monkeypatch, integration_enabled_params ): sentry_init(**integration_enabled_params) + app = quart_app_factory() monkeypatch.setattr(quart_sentry, "quart_auth", None) @@ -170,9 +171,10 @@ async def test_quart_auth_not_installed( @pytest.mark.asyncio async def test_quart_auth_not_configured( - sentry_init, app, capture_events, monkeypatch, integration_enabled_params + sentry_init, capture_events, monkeypatch, integration_enabled_params ): sentry_init(**integration_enabled_params) + app = quart_app_factory() assert quart_sentry.quart_auth @@ -186,9 +188,10 @@ async def test_quart_auth_not_configured( @pytest.mark.asyncio async def test_quart_auth_partially_configured( - sentry_init, app, capture_events, monkeypatch, integration_enabled_params + sentry_init, capture_events, monkeypatch, integration_enabled_params ): sentry_init(**integration_enabled_params) + app = quart_app_factory() events = capture_events() @@ -205,13 +208,13 @@ async def test_quart_auth_partially_configured( async def test_quart_auth_configured( send_default_pii, sentry_init, - app, user_id, capture_events, monkeypatch, integration_enabled_params, ): sentry_init(send_default_pii=send_default_pii, **integration_enabled_params) + app = quart_app_factory() @app.route("/login") async def login(): @@ -242,10 +245,9 @@ async def login(): [quart_sentry.QuartIntegration(), LoggingIntegration(event_level="ERROR")], ], ) -async def test_errors_not_reported_twice( - sentry_init, integrations, capture_events, app -): +async def test_errors_not_reported_twice(sentry_init, integrations, capture_events): sentry_init(integrations=integrations) + app = quart_app_factory() @app.route("/") async def index(): @@ -265,7 +267,7 @@ async def index(): @pytest.mark.asyncio -async def test_logging(sentry_init, capture_events, app): +async def test_logging(sentry_init, capture_events): # ensure that Quart's logger magic doesn't break ours sentry_init( integrations=[ @@ -273,6 +275,7 @@ async def test_logging(sentry_init, capture_events, app): LoggingIntegration(event_level="ERROR"), ] ) + app = quart_app_factory() @app.route("/") async def index(): @@ -289,13 +292,17 @@ async def index(): @pytest.mark.asyncio -async def test_no_errors_without_request(app, sentry_init): +async def test_no_errors_without_request(sentry_init): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() + async with app.app_context(): capture_exception(ValueError()) -def test_cli_commands_raise(app): +def test_cli_commands_raise(): + app = quart_app_factory() + if not hasattr(app, "cli"): pytest.skip("Too old quart version") @@ -312,8 +319,9 @@ def foo(): @pytest.mark.asyncio -async def test_500(sentry_init, app): +async def test_500(sentry_init): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() @app.route("/") async def index(): @@ -330,8 +338,9 @@ async def error_handler(err): @pytest.mark.asyncio -async def test_error_in_errorhandler(sentry_init, capture_events, app): +async def test_error_in_errorhandler(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() @app.route("/") async def index(): @@ -358,8 +367,9 @@ async def error_handler(err): @pytest.mark.asyncio -async def test_bad_request_not_captured(sentry_init, capture_events, app): +async def test_bad_request_not_captured(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() @app.route("/") @@ -374,8 +384,9 @@ async def index(): @pytest.mark.asyncio -async def test_does_not_leak_scope(sentry_init, capture_events, app): +async def test_does_not_leak_scope(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() sentry_sdk.get_isolation_scope().set_tag("request_data", False) @@ -402,8 +413,9 @@ async def generate(): @pytest.mark.asyncio -async def test_scoped_test_client(sentry_init, app): +async def test_scoped_test_client(sentry_init): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() @app.route("/") async def index(): @@ -417,12 +429,13 @@ async def index(): @pytest.mark.asyncio @pytest.mark.parametrize("exc_cls", [ZeroDivisionError, Exception]) async def test_errorhandler_for_exception_swallows_exception( - sentry_init, app, capture_events, exc_cls + sentry_init, capture_events, exc_cls ): # In contrast to error handlers for a status code, error # handlers for exceptions can swallow the exception (this is # just how the Quart signal works) sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() @app.route("/") @@ -441,8 +454,9 @@ async def zerodivision(e): @pytest.mark.asyncio -async def test_tracing_success(sentry_init, capture_events, app): +async def test_tracing_success(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0, integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() @app.before_request async def _(): @@ -474,8 +488,9 @@ async def hi_tx(): @pytest.mark.asyncio -async def test_tracing_error(sentry_init, capture_events, app): +async def test_tracing_error(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0, integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() @@ -498,8 +513,9 @@ async def error(): @pytest.mark.asyncio -async def test_class_based_views(sentry_init, app, capture_events): +async def test_class_based_views(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) + app = quart_app_factory() events = capture_events() @app.route("/") @@ -523,39 +539,56 @@ async def dispatch_request(self): @pytest.mark.parametrize("endpoint", ["/sync/thread_ids", "/async/thread_ids"]) -async def test_active_thread_id(sentry_init, capture_envelopes, endpoint, app): - sentry_init( - traces_sample_rate=1.0, - _experiments={"profiles_sample_rate": 1.0}, - ) +@pytest.mark.asyncio +async def test_active_thread_id( + sentry_init, capture_envelopes, teardown_profiling, endpoint +): + with mock.patch( + "sentry_sdk.profiler.transaction_profiler.PROFILE_MINIMUM_SAMPLES", 0 + ): + sentry_init( + traces_sample_rate=1.0, + profiles_sample_rate=1.0, + ) + app = quart_app_factory() - envelopes = capture_envelopes() + envelopes = capture_envelopes() - async with app.test_client() as client: - response = await client.get(endpoint) - assert response.status_code == 200 + async with app.test_client() as client: + response = await client.get(endpoint) + assert response.status_code == 200 + + data = json.loads(await response.get_data(as_text=True)) - data = json.loads(response.content) + envelopes = [envelope for envelope in envelopes] + assert len(envelopes) == 1 - envelopes = [envelope for envelope in envelopes] - assert len(envelopes) == 1 + profiles = [item for item in envelopes[0].items if item.type == "profile"] + assert len(profiles) == 1, envelopes[0].items - profiles = [item for item in envelopes[0].items if item.type == "profile"] - assert len(profiles) == 1 + for item in profiles: + transactions = item.payload.json["transactions"] + assert len(transactions) == 1 + assert str(data["active"]) == transactions[0]["active_thread_id"] - for profile in profiles: - transactions = profile.payload.json["transactions"] + transactions = [ + item for item in envelopes[0].items if item.type == "transaction" + ] assert len(transactions) == 1 - assert str(data["active"]) == transactions[0]["active_thread_id"] + + for item in transactions: + transaction = item.payload.json + trace_context = transaction["contexts"]["trace"] + assert str(data["active"]) == trace_context["data"]["thread.id"] @pytest.mark.asyncio -async def test_span_origin(sentry_init, capture_events, app): +async def test_span_origin(sentry_init, capture_events): sentry_init( integrations=[quart_sentry.QuartIntegration()], traces_sample_rate=1.0, ) - + app = quart_app_factory() events = capture_events() client = app.test_client() diff --git a/tests/integrations/starlette/test_starlette.py b/tests/integrations/starlette/test_starlette.py index 1ba9eb7589..fd47895f5a 100644 --- a/tests/integrations/starlette/test_starlette.py +++ b/tests/integrations/starlette/test_starlette.py @@ -885,7 +885,7 @@ def test_legacy_setup( def test_active_thread_id(sentry_init, capture_envelopes, teardown_profiling, endpoint): sentry_init( traces_sample_rate=1.0, - _experiments={"profiles_sample_rate": 1.0}, + profiles_sample_rate=1.0, ) app = starlette_app_factory() asgi_app = SentryAsgiMiddleware(app) @@ -904,11 +904,19 @@ def test_active_thread_id(sentry_init, capture_envelopes, teardown_profiling, en profiles = [item for item in envelopes[0].items if item.type == "profile"] assert len(profiles) == 1 - for profile in profiles: - transactions = profile.payload.json["transactions"] + for item in profiles: + transactions = item.payload.json["transactions"] assert len(transactions) == 1 assert str(data["active"]) == transactions[0]["active_thread_id"] + transactions = [item for item in envelopes[0].items if item.type == "transaction"] + assert len(transactions) == 1 + + for item in transactions: + transaction = item.payload.json + trace_context = transaction["contexts"]["trace"] + assert str(data["active"]) == trace_context["data"]["thread.id"] + def test_original_request_not_scrubbed(sentry_init, capture_events): sentry_init(integrations=[StarletteIntegration()]) From 72f4d991d70b95edb40fb71e506e93cf5a90e1a2 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 11:02:46 +0100 Subject: [PATCH 002/591] ci(tox): Exclude fakeredis 2.26.0 on py3.6 and 3.7 (#3695) `fakeredis` `2.26.0` [broke on Python 3.6 and 3.7](https://github.com/cunla/fakeredis-py/issues/341). A fix should be available when the next version is available. --- tox.ini | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tox.ini b/tox.ini index b53cc73d7f..02e2dee388 100644 --- a/tox.ini +++ b/tox.ini @@ -583,6 +583,7 @@ deps = # Redis redis: fakeredis!=1.7.4 redis: pytest<8.0.0 + {py3.6,py3.7}-redis: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 {py3.7,py3.8,py3.9,py3.10,py3.11}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 @@ -602,7 +603,9 @@ deps = rq-v{0.6}: redis<3.2.2 rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 rq-v{1.15,1.16}: fakeredis + {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 rq-latest: fakeredis + {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 rq-v0.6: rq~=0.6.0 rq-v0.13: rq~=0.13.0 rq-v1.0: rq~=1.0.0 From 483a0bdf324cf6dfd1fc6399a15568b9e942f8b1 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 13:36:16 +0100 Subject: [PATCH 003/591] build: Remove pytest pin in requirements-devenv.txt (#3696) The pytest pin in requirements-devenv.txt appears to be unnecessary. Our tests anyways do not seem to respect this pin; the actual pins are defined for each environment in tox.ini. ref #3035 --- requirements-devenv.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements-devenv.txt b/requirements-devenv.txt index 29d3f15ec9..c0fa5cf245 100644 --- a/requirements-devenv.txt +++ b/requirements-devenv.txt @@ -1,5 +1,5 @@ -r requirements-linting.txt -r requirements-testing.txt mockupdb # required by `pymongo` tests that are enabled by `pymongo` from linter requirements -pytest<7.0.0 # https://github.com/pytest-dev/pytest/issues/9621; see tox.ini +pytest pytest-asyncio From 6b8114c3009e40e3663c209255189f90037557f9 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 14:10:19 +0100 Subject: [PATCH 004/591] ci: Run CodeQL action on all PRs (#3698) This action only is triggered on PRs to `master`, but the action is required. This becomes a problem when a PR is opened against a branch other than `master` (e.g. as part of a PR tree). When the parent branch is merged to `master`, the PR's base automatically changes to `master`, but this action does not get triggered. Instead, it blocks on "Expected" and can only be run by adding commits to the branch. Running the action on PRs against any branch should fix this. Also, add logic to cancel in-progress workflows on pull requests (logic taken from our other actions) --- .github/workflows/codeql-analysis.yml | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index 573c49fb01..d95353c652 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -17,13 +17,15 @@ on: - master - sentry-sdk-2.0 pull_request: - # The branches below must be a subset of the branches above - branches: - - master - - sentry-sdk-2.0 schedule: - cron: '18 18 * * 3' +# Cancel in progress workflows on pull_requests. +# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value +concurrency: + group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} + cancel-in-progress: true + permissions: contents: read From 1ce7c31a41aac2b63be225858747c7ddfc846420 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 14:12:28 +0100 Subject: [PATCH 005/591] ci: Run license compliance action on all PRs (#3699) This action only is triggered on PRs to master, but the action is required. This becomes a problem when a PR is opened against a branch other than master (e.g. as part of a PR tree). When the parent branch is merged to master, the PR's base automatically changes to master, but this action does not get triggered. Instead, it blocks on "Expected" and can only be run by adding commits to the branch. Running the action on PRs against any branch should fix this. Also, add logic to cancel in-progress workflows on pull requests (logic taken from our other actions) --- .github/workflows/enforce-license-compliance.yml | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/.github/workflows/enforce-license-compliance.yml b/.github/workflows/enforce-license-compliance.yml index 01e02ccb8b..ef79ed112b 100644 --- a/.github/workflows/enforce-license-compliance.yml +++ b/.github/workflows/enforce-license-compliance.yml @@ -8,10 +8,11 @@ on: - release/* - sentry-sdk-2.0 pull_request: - branches: - - master - - main - - sentry-sdk-2.0 + +# Cancel in progress workflows on pull_requests. +# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value +concurrency: + group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} jobs: enforce-license-compliance: From 200be874daa55d5a72b0f0713381370dda9dc414 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 14:29:29 +0100 Subject: [PATCH 006/591] ci(tox): Unpin `pytest` for Python 3.8+ `common` tests (#3697) This pin appears to be unnecessary on Python 3.8+. ref #3035 --- tox.ini | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tox.ini b/tox.ini index 02e2dee388..17e36c29bb 100644 --- a/tox.ini +++ b/tox.ini @@ -294,8 +294,8 @@ deps = # See https://github.com/pytest-dev/pytest/issues/9621 # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12}-common: pytest<7.0.0 - py3.13-common: pytest + {py3.6,py3.7}-common: pytest<7.0.0 + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common: pytest # === Gevent === {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 From 7e52235ec6587d4225bf1e5bac0e6e812543d0dd Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 14:31:00 +0100 Subject: [PATCH 007/591] test(tox): Unpin `pytest` on Python 3.8+ `gevent` tests (#3700) The pin appears to be unnecessary in Python 3.8+. ref #3035 --- tox.ini | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/tox.ini b/tox.ini index 17e36c29bb..690fb36558 100644 --- a/tox.ini +++ b/tox.ini @@ -303,7 +303,8 @@ deps = # See https://github.com/pytest-dev/pytest/issues/9621 # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest<7.0.0 + {py3.6,py3.7}-gevent: pytest<7.0.0 + {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest # === Integrations === From b6482f0a474847b1e65b5ec1a9575b929b7207c6 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 15:03:21 +0100 Subject: [PATCH 008/591] test(tox): Unpin `pytest` for `celery` tests (#3701) Unpin pytest for Celery tests. This requires adding a placeholder test to workaround a bug with pytest-forked. ref #3035 --- tests/integrations/celery/test_celery.py | 8 ++++++++ tox.ini | 1 - 2 files changed, 8 insertions(+), 1 deletion(-) diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index ffd3f0db62..e51341599f 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -831,3 +831,11 @@ def test_send_task_wrapped( assert span["description"] == "very_creative_task_name" assert span["op"] == "queue.submit.celery" assert span["trace_id"] == kwargs["headers"]["sentry-trace"].split("-")[0] + + +@pytest.mark.skip(reason="placeholder so that forked test does not come last") +def test_placeholder(): + """Forked tests must not come last in the module. + See https://github.com/pytest-dev/pytest-forked/issues/67#issuecomment-1964718720. + """ + pass diff --git a/tox.ini b/tox.ini index 690fb36558..75d74dbb03 100644 --- a/tox.ini +++ b/tox.ini @@ -375,7 +375,6 @@ deps = celery-latest: Celery celery: newrelic - celery: pytest<7 {py3.7}-celery: importlib-metadata<5.0 # Chalice From 4c1367b300811d4f1693b5af206b749f2139a18f Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 28 Oct 2024 17:51:06 +0100 Subject: [PATCH 009/591] test: Disable broken RQ test in newly-released RQ 2.0 (#3708) See #3707 --- tests/integrations/rq/test_rq.py | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/tests/integrations/rq/test_rq.py b/tests/integrations/rq/test_rq.py index e445b588be..0b690ca3dc 100644 --- a/tests/integrations/rq/test_rq.py +++ b/tests/integrations/rq/test_rq.py @@ -35,6 +35,7 @@ def _patch_rq_get_server_version(monkeypatch): def crashing_job(foo): + print("RUNNING CRASHING JOB") 1 / 0 @@ -254,6 +255,11 @@ def test_traces_sampler_gets_correct_values_in_sampling_context( @pytest.mark.skipif( parse_version(rq.__version__) < (1, 5), reason="At least rq-1.5 required" ) +@pytest.mark.skipif( + parse_version(rq.__version__) >= (2,), + reason="Test broke in RQ 2.0. Investigate and fix. " + "See https://github.com/getsentry/sentry-python/issues/3707.", +) def test_job_with_retries(sentry_init, capture_events): sentry_init(integrations=[RqIntegration()]) events = capture_events() From 897333bce69d18a9d356ca7748b3079c02576f45 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 29 Oct 2024 09:45:29 +0100 Subject: [PATCH 010/591] test(rq): Remove accidentally-committed print (#3712) #3708 got auto-merged before I had the chance to remove this print statement. --- tests/integrations/rq/test_rq.py | 1 - 1 file changed, 1 deletion(-) diff --git a/tests/integrations/rq/test_rq.py b/tests/integrations/rq/test_rq.py index 0b690ca3dc..ffd6f458e1 100644 --- a/tests/integrations/rq/test_rq.py +++ b/tests/integrations/rq/test_rq.py @@ -35,7 +35,6 @@ def _patch_rq_get_server_version(monkeypatch): def crashing_job(foo): - print("RUNNING CRASHING JOB") 1 / 0 From d48dc46823d3602ca899ecb2178cfe4b8267f89c Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 29 Oct 2024 10:13:05 +0100 Subject: [PATCH 011/591] ci: Clarify that only pinned tests are required (#3713) Rename the action that checks that all our pinned-version tests for our integrations are named "All pinned XXX tests passed" rather than just "All XXX tests passed." The old name was confusing because the action only checks that the pinned tests have passed. --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-aws-lambda.yml | 2 +- .github/workflows/test-integrations-cloud-computing.yml | 2 +- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-data-processing.yml | 2 +- .github/workflows/test-integrations-databases.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-miscellaneous.yml | 2 +- .github/workflows/test-integrations-networking.yml | 2 +- .github/workflows/test-integrations-web-frameworks-1.yml | 2 +- .github/workflows/test-integrations-web-frameworks-2.yml | 2 +- scripts/split-tox-gh-actions/templates/check_required.jinja | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 723f9c8412..24ccc77a87 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -165,7 +165,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All AI tests passed + name: All pinned AI tests passed needs: test-ai-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-aws-lambda.yml b/.github/workflows/test-integrations-aws-lambda.yml index 38c838ab33..6f5ea794b8 100644 --- a/.github/workflows/test-integrations-aws-lambda.yml +++ b/.github/workflows/test-integrations-aws-lambda.yml @@ -112,7 +112,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All AWS Lambda tests passed + name: All pinned AWS Lambda tests passed needs: test-aws_lambda-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-cloud-computing.yml b/.github/workflows/test-integrations-cloud-computing.yml index a3b7fc57ab..1f6913ea4a 100644 --- a/.github/workflows/test-integrations-cloud-computing.yml +++ b/.github/workflows/test-integrations-cloud-computing.yml @@ -157,7 +157,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Cloud Computing tests passed + name: All pinned Cloud Computing tests passed needs: test-cloud_computing-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 8116b1b67c..ecffdb6f3e 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -77,7 +77,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Common tests passed + name: All pinned Common tests passed needs: test-common-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-data-processing.yml b/.github/workflows/test-integrations-data-processing.yml index acabcd1748..49d18fc24c 100644 --- a/.github/workflows/test-integrations-data-processing.yml +++ b/.github/workflows/test-integrations-data-processing.yml @@ -193,7 +193,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Data Processing tests passed + name: All pinned Data Processing tests passed needs: test-data_processing-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-databases.yml b/.github/workflows/test-integrations-databases.yml index 741e8fc43e..49d3e923ee 100644 --- a/.github/workflows/test-integrations-databases.yml +++ b/.github/workflows/test-integrations-databases.yml @@ -211,7 +211,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Databases tests passed + name: All pinned Databases tests passed needs: test-databases-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index ba4091215e..2cefb5d191 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -157,7 +157,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All GraphQL tests passed + name: All pinned GraphQL tests passed needs: test-graphql-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-miscellaneous.yml b/.github/workflows/test-integrations-miscellaneous.yml index 064d083335..0b49a27219 100644 --- a/.github/workflows/test-integrations-miscellaneous.yml +++ b/.github/workflows/test-integrations-miscellaneous.yml @@ -165,7 +165,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Miscellaneous tests passed + name: All pinned Miscellaneous tests passed needs: test-miscellaneous-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-networking.yml b/.github/workflows/test-integrations-networking.yml index 192eb1b35b..c24edff174 100644 --- a/.github/workflows/test-integrations-networking.yml +++ b/.github/workflows/test-integrations-networking.yml @@ -157,7 +157,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Networking tests passed + name: All pinned Networking tests passed needs: test-networking-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-web-frameworks-1.yml b/.github/workflows/test-integrations-web-frameworks-1.yml index f2bcb336dd..a655710843 100644 --- a/.github/workflows/test-integrations-web-frameworks-1.yml +++ b/.github/workflows/test-integrations-web-frameworks-1.yml @@ -193,7 +193,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Web Frameworks 1 tests passed + name: All pinned Web Frameworks 1 tests passed needs: test-web_frameworks_1-pinned # Always run this, even if a dependent job failed if: always() diff --git a/.github/workflows/test-integrations-web-frameworks-2.yml b/.github/workflows/test-integrations-web-frameworks-2.yml index 8f6bd543df..d3f1001e2c 100644 --- a/.github/workflows/test-integrations-web-frameworks-2.yml +++ b/.github/workflows/test-integrations-web-frameworks-2.yml @@ -205,7 +205,7 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All Web Frameworks 2 tests passed + name: All pinned Web Frameworks 2 tests passed needs: test-web_frameworks_2-pinned # Always run this, even if a dependent job failed if: always() diff --git a/scripts/split-tox-gh-actions/templates/check_required.jinja b/scripts/split-tox-gh-actions/templates/check_required.jinja index b9b0f54015..ddb47cddf1 100644 --- a/scripts/split-tox-gh-actions/templates/check_required.jinja +++ b/scripts/split-tox-gh-actions/templates/check_required.jinja @@ -1,5 +1,5 @@ check_required_tests: - name: All {{ group }} tests passed + name: All pinned {{ group }} tests passed {% if "pinned" in categories %} needs: test-{{ group | replace(" ", "_") | lower }}-pinned {% endif %} From c21962e98d8879f550725d6ececb6b6c28f9d32c Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 29 Oct 2024 13:09:04 +0100 Subject: [PATCH 012/591] test(redis): Install `pytest-asyncio` for `redis` tests (Python 3.12-13) (#3706) Although we run the `redis` tests on Python 3.12 and 3.13, we don't install `pytest-asyncio` on these versions. We likely should. --- tox.ini | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tox.ini b/tox.ini index 75d74dbb03..67d6166461 100644 --- a/tox.ini +++ b/tox.ini @@ -584,7 +584,7 @@ deps = redis: fakeredis!=1.7.4 redis: pytest<8.0.0 {py3.6,py3.7}-redis: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - {py3.7,py3.8,py3.9,py3.10,py3.11}-redis: pytest-asyncio + {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 redis-v5: redis~=5.0 From 000c8e6c4eedf046c601b81d5d8d82f92115eddd Mon Sep 17 00:00:00 2001 From: Ben Beasley Date: Tue, 29 Oct 2024 08:13:56 -0400 Subject: [PATCH 013/591] fix(starlette): Prefer python_multipart import over multipart (#3710) See also releases 0.0.13 through 0.0.16 at https://github.com/Kludex/python-multipart/releases. --------- Co-authored-by: Daniel Szoke --- sentry_sdk/integrations/starlette.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index 52c64f6843..d9db8bd6b8 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -65,7 +65,12 @@ try: # Optional dependency of Starlette to parse form data. - import multipart # type: ignore + try: + # python-multipart 0.0.13 and later + import python_multipart as multipart # type: ignore + except ImportError: + # python-multipart 0.0.12 and earlier + import multipart # type: ignore except ImportError: multipart = None From bf400904245c3809bad5f20fd637408f519e7a15 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 29 Oct 2024 13:56:50 +0100 Subject: [PATCH 014/591] test(tornado): Unpin `pytest` for `tornado-latest` tests (#3714) The Pytest version pin is only needed for `tornado-v6.0` and `tornado-v6.2`. The incompatibility with the latest Pytest versions has been fixed in newer Tornado versions. --- tox.ini | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/tox.ini b/tox.ini index 67d6166461..ef30e539b5 100644 --- a/tox.ini +++ b/tox.ini @@ -673,7 +673,9 @@ deps = strawberry-latest: strawberry-graphql[fastapi,flask] # Tornado - tornado: pytest<8.2 + # Tornado <6.4.1 is incompatible with Pytest ≥8.2 + # See https://github.com/tornadoweb/tornado/pull/3382. + tornado-{v6.0,v6.2}: pytest<8.2 tornado-v6.0: tornado~=6.0.0 tornado-v6.2: tornado~=6.2.0 tornado-latest: tornado From 02d09346e6d070e03b828807d72485b6f23b2c11 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Tue, 29 Oct 2024 13:14:06 -0400 Subject: [PATCH 015/591] fix(profiling): Use `type()` instead when extracting frames (#3716) When extract frame names, we should avoid accessing the `__class__` attribute as it can be overwritten in the class implementation. In this particular instance, the `SimpleLazyObject` class in django wraps `__class__` so when it is accessed, it can cause the underlying lazy object to be evaluation unexpectedly. To avoid this, use the `type()` builtin function which does cannot be overwritten and will return the correct class. Note that this does not work with old style classes but since dropping python 2 support, we only need to consider new style classes. --- sentry_sdk/profiler/utils.py | 2 +- tests/integrations/django/test_basic.py | 48 +++++++++++++++++++++++++ 2 files changed, 49 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/profiler/utils.py b/sentry_sdk/profiler/utils.py index e78ea54256..3554cddb5d 100644 --- a/sentry_sdk/profiler/utils.py +++ b/sentry_sdk/profiler/utils.py @@ -89,7 +89,7 @@ def get_frame_name(frame): and co_varnames[0] == "self" and "self" in frame.f_locals ): - for cls in frame.f_locals["self"].__class__.__mro__: + for cls in type(frame.f_locals["self"]).__mro__: if name in cls.__dict__: return "{}.{}".format(cls.__name__, name) except (AttributeError, ValueError): diff --git a/tests/integrations/django/test_basic.py b/tests/integrations/django/test_basic.py index c8282412ea..0e3f700105 100644 --- a/tests/integrations/django/test_basic.py +++ b/tests/integrations/django/test_basic.py @@ -1,6 +1,8 @@ +import inspect import json import os import re +import sys import pytest from functools import partial from unittest.mock import patch @@ -12,6 +14,7 @@ from django.core.management import execute_from_command_line from django.db.utils import OperationalError, ProgrammingError, DataError from django.http.request import RawPostDataException +from django.utils.functional import SimpleLazyObject try: from django.urls import reverse @@ -29,6 +32,7 @@ ) from sentry_sdk.integrations.django.signals_handlers import _get_receiver_name from sentry_sdk.integrations.executing import ExecutingIntegration +from sentry_sdk.profiler.utils import get_frame_name from sentry_sdk.tracing import Span from tests.conftest import unpack_werkzeug_response from tests.integrations.django.myapp.wsgi import application @@ -1295,3 +1299,47 @@ def test_ensures_no_spotlight_middleware_when_no_spotlight( added = frozenset(settings.MIDDLEWARE) ^ original_middleware assert "sentry_sdk.spotlight.SpotlightMiddleware" not in added + + +def test_get_frame_name_when_in_lazy_object(): + allowed_to_init = False + + class SimpleLazyObjectWrapper(SimpleLazyObject): + def unproxied_method(self): + """ + For testing purposes. We inject a method on the SimpleLazyObject + class so if python is executing this method, we should get + this class instead of the wrapped class and avoid evaluating + the wrapped object too early. + """ + return inspect.currentframe() + + class GetFrame: + def __init__(self): + assert allowed_to_init, "GetFrame not permitted to initialize yet" + + def proxied_method(self): + """ + For testing purposes. We add an proxied method on the instance + class so if python is executing this method, we should get + this class instead of the wrapper class. + """ + return inspect.currentframe() + + instance = SimpleLazyObjectWrapper(lambda: GetFrame()) + + assert get_frame_name(instance.unproxied_method()) == ( + "SimpleLazyObjectWrapper.unproxied_method" + if sys.version_info < (3, 11) + else "test_get_frame_name_when_in_lazy_object..SimpleLazyObjectWrapper.unproxied_method" + ) + + # Now that we're about to access an instance method on the wrapped class, + # we should permit initializing it + allowed_to_init = True + + assert get_frame_name(instance.proxied_method()) == ( + "GetFrame.proxied_method" + if sys.version_info < (3, 11) + else "test_get_frame_name_when_in_lazy_object..GetFrame.proxied_method" + ) From ce9986cb19ee80d92bdf68bee6243d5c049fdb54 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Thu, 31 Oct 2024 12:57:41 +0000 Subject: [PATCH 016/591] fix(http2): Check for h2 existence (#3690) The new `HTTP2Transport` needs `httpcore` _and_ `h2` and we only checked for `httpcore`. This caused runtime errors and dropping of all events during testing as the test platform had `httpcore` installed but not `h2`. This patch adds both as conditions for the new transport implementation. Ideally, when we switch out the old transport, we'd silently check for `h2` existence only and set the `http2` option accordingly. --- sentry_sdk/transport.py | 1 + 1 file changed, 1 insertion(+) diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index 1b1842d03e..8798115898 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -720,6 +720,7 @@ def _request( try: import httpcore + import h2 # type: ignore # noqa: F401 except ImportError: # Sorry, no Http2Transport for you class Http2Transport(HttpTransport): From 5c5d98a7937330bd4ab69ee8a10b0d4e438c00ea Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Thu, 31 Oct 2024 15:59:36 +0000 Subject: [PATCH 017/591] test: Fix UTC assuming test (#3722) Fixes #3720. --- tests/test_utils.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/test_utils.py b/tests/test_utils.py index 87e2659a12..6e01bb4f3a 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -69,8 +69,8 @@ def _normalize_distribution_name(name): ), # UTC time ( "2021-01-01T00:00:00.000000", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), # No TZ -- assume UTC + datetime(2021, 1, 1).astimezone(timezone.utc), + ), # No TZ -- assume local but convert to UTC ( "2021-01-01T00:00:00Z", datetime(2021, 1, 1, tzinfo=timezone.utc), From 5e2d2cf7fdf367dc3bced0d4c4efe33c1046887c Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Thu, 31 Oct 2024 16:12:07 -0400 Subject: [PATCH 018/591] fix(tracing): End http.client span on timeout (#3723) If the http request times out, the http client span never gets finished. So make sure to finish it no matter what. --- sentry_sdk/integrations/stdlib.py | 10 ++++--- tests/integrations/stdlib/test_httplib.py | 33 +++++++++++++++++++++++ 2 files changed, 39 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/stdlib.py b/sentry_sdk/integrations/stdlib.py index 287c8cb272..d388c5bca6 100644 --- a/sentry_sdk/integrations/stdlib.py +++ b/sentry_sdk/integrations/stdlib.py @@ -127,11 +127,13 @@ def getresponse(self, *args, **kwargs): if span is None: return real_getresponse(self, *args, **kwargs) - rv = real_getresponse(self, *args, **kwargs) + try: + rv = real_getresponse(self, *args, **kwargs) - span.set_http_status(int(rv.status)) - span.set_data("reason", rv.reason) - span.finish() + span.set_http_status(int(rv.status)) + span.set_data("reason", rv.reason) + finally: + span.finish() return rv diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index c327331608..200b282f53 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -1,5 +1,6 @@ import random from http.client import HTTPConnection, HTTPSConnection +from socket import SocketIO from urllib.request import urlopen from unittest import mock @@ -342,3 +343,35 @@ def test_span_origin(sentry_init, capture_events): assert event["spans"][0]["op"] == "http.client" assert event["spans"][0]["origin"] == "auto.http.stdlib.httplib" + + +def test_http_timeout(monkeypatch, sentry_init, capture_envelopes): + mock_readinto = mock.Mock(side_effect=TimeoutError) + monkeypatch.setattr(SocketIO, "readinto", mock_readinto) + + sentry_init(traces_sample_rate=1.0) + + envelopes = capture_envelopes() + + with start_transaction(op="op", name="name"): + try: + conn = HTTPSConnection("www.squirrelchasers.com") + conn.request("GET", "/top-chasers") + conn.getresponse() + except Exception: + pass + + items = [ + item + for envelope in envelopes + for item in envelope.items + if item.type == "transaction" + ] + assert len(items) == 1 + + transaction = items[0].payload.json + assert len(transaction["spans"]) == 1 + + span = transaction["spans"][0] + assert span["op"] == "http.client" + assert span["description"] == "GET https://www.squirrelchasers.com/top-chasers" From d06a1897e5106e2a0521bc51857eb30abddb0ef4 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 4 Nov 2024 11:24:01 +0100 Subject: [PATCH 019/591] docs(hub): Correct typo in a comment (#3726) --- sentry_sdk/hub.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/hub.py b/sentry_sdk/hub.py index ec30e25419..7fda9202df 100644 --- a/sentry_sdk/hub.py +++ b/sentry_sdk/hub.py @@ -101,7 +101,7 @@ def current(cls): rv = _local.get(None) if rv is None: with _suppress_hub_deprecation_warning(): - # This will raise a deprecation warning; supress it since we already warned above. + # This will raise a deprecation warning; suppress it since we already warned above. rv = Hub(GLOBAL_HUB) _local.set(rv) return rv From dd1117d63fd690d502b32c263e9e970b682fa280 Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Mon, 4 Nov 2024 06:00:41 -0600 Subject: [PATCH 020/591] Add LaunchDarkly and OpenFeature integration (#3648) Adds LaunchDarkly and OpenFeature integration and extends the `Scope` with a `flags` property. As flags are evaluated by an application they are stored within the Sentry SDK (lru cache). When an error occurs we fetch the flags stored in the SDK and serialize them on the event. --------- Co-authored-by: Anton Pirker Co-authored-by: Ivana Kellyer Co-authored-by: Andrew Liu <159852527+aliu39@users.noreply.github.com> --- .../test-integrations-miscellaneous.yml | 16 +++ mypy.ini | 2 + requirements-linting.txt | 2 + .../split-tox-gh-actions.py | 2 + sentry_sdk/_lru_cache.py | 17 +++ sentry_sdk/consts.py | 1 + sentry_sdk/flag_utils.py | 47 +++++++ sentry_sdk/integrations/launchdarkly.py | 64 ++++++++++ sentry_sdk/integrations/openfeature.py | 43 +++++++ sentry_sdk/scope.py | 16 +++ setup.py | 2 + tests/integrations/launchdarkly/__init__.py | 3 + .../launchdarkly/test_launchdarkly.py | 116 ++++++++++++++++++ tests/integrations/openfeature/__init__.py | 3 + .../openfeature/test_openfeature.py | 80 ++++++++++++ tests/test_flag_utils.py | 43 +++++++ tests/test_lru_cache.py | 23 ++++ tox.ini | 18 +++ 18 files changed, 498 insertions(+) create mode 100644 sentry_sdk/flag_utils.py create mode 100644 sentry_sdk/integrations/launchdarkly.py create mode 100644 sentry_sdk/integrations/openfeature.py create mode 100644 tests/integrations/launchdarkly/__init__.py create mode 100644 tests/integrations/launchdarkly/test_launchdarkly.py create mode 100644 tests/integrations/openfeature/__init__.py create mode 100644 tests/integrations/openfeature/test_openfeature.py create mode 100644 tests/test_flag_utils.py diff --git a/.github/workflows/test-integrations-miscellaneous.yml b/.github/workflows/test-integrations-miscellaneous.yml index 0b49a27219..88a576505e 100644 --- a/.github/workflows/test-integrations-miscellaneous.yml +++ b/.github/workflows/test-integrations-miscellaneous.yml @@ -45,10 +45,18 @@ jobs: - name: Erase coverage run: | coverage erase + - name: Test launchdarkly latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-launchdarkly-latest" - name: Test loguru latest run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-loguru-latest" + - name: Test openfeature latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature-latest" - name: Test opentelemetry latest run: | set -x # print commands that are executed @@ -117,10 +125,18 @@ jobs: - name: Erase coverage run: | coverage erase + - name: Test launchdarkly pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-launchdarkly" - name: Test loguru pinned run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-loguru" + - name: Test openfeature pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openfeature" - name: Test opentelemetry pinned run: | set -x # print commands that are executed diff --git a/mypy.ini b/mypy.ini index bacba96ceb..63fa7f334f 100644 --- a/mypy.ini +++ b/mypy.ini @@ -74,6 +74,8 @@ ignore_missing_imports = True ignore_missing_imports = True [mypy-openai.*] ignore_missing_imports = True +[mypy-openfeature.*] +ignore_missing_imports = True [mypy-huggingface_hub.*] ignore_missing_imports = True [mypy-arq.*] diff --git a/requirements-linting.txt b/requirements-linting.txt index d2a65b31db..c9d4bd7f5c 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -15,3 +15,5 @@ flake8-bugbear pep8-naming pre-commit # local linting httpcore +openfeature-sdk +launchdarkly-server-sdk diff --git a/scripts/split-tox-gh-actions/split-tox-gh-actions.py b/scripts/split-tox-gh-actions/split-tox-gh-actions.py index 7ed2505f40..c0bf2a7a09 100755 --- a/scripts/split-tox-gh-actions/split-tox-gh-actions.py +++ b/scripts/split-tox-gh-actions/split-tox-gh-actions.py @@ -125,7 +125,9 @@ "tornado", ], "Miscellaneous": [ + "launchdarkly", "loguru", + "openfeature", "opentelemetry", "potel", "pure_eval", diff --git a/sentry_sdk/_lru_cache.py b/sentry_sdk/_lru_cache.py index 37e86e5fe3..ec557b1093 100644 --- a/sentry_sdk/_lru_cache.py +++ b/sentry_sdk/_lru_cache.py @@ -62,6 +62,8 @@ """ +from copy import copy + SENTINEL = object() @@ -89,6 +91,13 @@ def __init__(self, max_size): self.hits = self.misses = 0 + def __copy__(self): + cache = LRUCache(self.max_size) + cache.full = self.full + cache.cache = copy(self.cache) + cache.root = copy(self.root) + return cache + def set(self, key, value): link = self.cache.get(key, SENTINEL) @@ -154,3 +163,11 @@ def get(self, key, default=None): self.hits += 1 return link[VALUE] + + def get_all(self): + nodes = [] + node = self.root[NEXT] + while node is not self.root: + nodes.append((node[KEY], node[VALUE])) + node = node[NEXT] + return nodes diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 6791abeb0e..fdb20caadf 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -59,6 +59,7 @@ class CompressionAlgo(Enum): "Experiments", { "max_spans": Optional[int], + "max_flags": Optional[int], "record_sql_params": Optional[bool], "continuous_profiling_auto_start": Optional[bool], "continuous_profiling_mode": Optional[ContinuousProfilerMode], diff --git a/sentry_sdk/flag_utils.py b/sentry_sdk/flag_utils.py new file mode 100644 index 0000000000..2b345a7f0b --- /dev/null +++ b/sentry_sdk/flag_utils.py @@ -0,0 +1,47 @@ +from copy import copy +from typing import TYPE_CHECKING + +import sentry_sdk +from sentry_sdk._lru_cache import LRUCache + +if TYPE_CHECKING: + from typing import TypedDict, Optional + from sentry_sdk._types import Event, ExcInfo + + FlagData = TypedDict("FlagData", {"flag": str, "result": bool}) + + +DEFAULT_FLAG_CAPACITY = 100 + + +class FlagBuffer: + + def __init__(self, capacity): + # type: (int) -> None + self.buffer = LRUCache(capacity) + self.capacity = capacity + + def clear(self): + # type: () -> None + self.buffer = LRUCache(self.capacity) + + def __copy__(self): + # type: () -> FlagBuffer + buffer = FlagBuffer(capacity=self.capacity) + buffer.buffer = copy(self.buffer) + return buffer + + def get(self): + # type: () -> list[FlagData] + return [{"flag": key, "result": value} for key, value in self.buffer.get_all()] + + def set(self, flag, result): + # type: (str, bool) -> None + self.buffer.set(flag, result) + + +def flag_error_processor(event, exc_info): + # type: (Event, ExcInfo) -> Optional[Event] + scope = sentry_sdk.get_current_scope() + event["contexts"]["flags"] = {"values": scope.flags.get()} + return event diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py new file mode 100644 index 0000000000..9e00e12ede --- /dev/null +++ b/sentry_sdk/integrations/launchdarkly.py @@ -0,0 +1,64 @@ +from typing import TYPE_CHECKING +import sentry_sdk + +from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.flag_utils import flag_error_processor + +try: + import ldclient + from ldclient.hook import Hook, Metadata + + if TYPE_CHECKING: + from ldclient import LDClient + from ldclient.hook import EvaluationSeriesContext + from ldclient.evaluation import EvaluationDetail + + from typing import Any +except ImportError: + raise DidNotEnable("LaunchDarkly is not installed") + + +class LaunchDarklyIntegration(Integration): + identifier = "launchdarkly" + + def __init__(self, ld_client=None): + # type: (LDClient | None) -> None + """ + :param client: An initialized LDClient instance. If a client is not provided, this + integration will attempt to use the shared global instance. + """ + try: + client = ld_client or ldclient.get() + except Exception as exc: + raise DidNotEnable("Error getting LaunchDarkly client. " + repr(exc)) + + if not client.is_initialized(): + raise DidNotEnable("LaunchDarkly client is not initialized.") + + # Register the flag collection hook with the LD client. + client.add_hook(LaunchDarklyHook()) + + @staticmethod + def setup_once(): + # type: () -> None + scope = sentry_sdk.get_current_scope() + scope.add_error_processor(flag_error_processor) + + +class LaunchDarklyHook(Hook): + + @property + def metadata(self): + # type: () -> Metadata + return Metadata(name="sentry-feature-flag-recorder") + + def after_evaluation(self, series_context, data, detail): + # type: (EvaluationSeriesContext, dict[Any, Any], EvaluationDetail) -> dict[Any, Any] + if isinstance(detail.value, bool): + flags = sentry_sdk.get_current_scope().flags + flags.set(series_context.key, detail.value) + return data + + def before_evaluation(self, series_context, data): + # type: (EvaluationSeriesContext, dict[Any, Any]) -> dict[Any, Any] + return data # No-op. diff --git a/sentry_sdk/integrations/openfeature.py b/sentry_sdk/integrations/openfeature.py new file mode 100644 index 0000000000..18f968a703 --- /dev/null +++ b/sentry_sdk/integrations/openfeature.py @@ -0,0 +1,43 @@ +from typing import TYPE_CHECKING +import sentry_sdk + +from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.flag_utils import flag_error_processor + +try: + from openfeature import api + from openfeature.hook import Hook + + if TYPE_CHECKING: + from openfeature.flag_evaluation import FlagEvaluationDetails + from openfeature.hook import HookContext, HookHints +except ImportError: + raise DidNotEnable("OpenFeature is not installed") + + +class OpenFeatureIntegration(Integration): + identifier = "openfeature" + + @staticmethod + def setup_once(): + # type: () -> None + scope = sentry_sdk.get_current_scope() + scope.add_error_processor(flag_error_processor) + + # Register the hook within the global openfeature hooks list. + api.add_hooks(hooks=[OpenFeatureHook()]) + + +class OpenFeatureHook(Hook): + + def after(self, hook_context, details, hints): + # type: (HookContext, FlagEvaluationDetails[bool], HookHints) -> None + if isinstance(details.value, bool): + flags = sentry_sdk.get_current_scope().flags + flags.set(details.flag_key, details.value) + + def error(self, hook_context, exception, hints): + # type: (HookContext, Exception, HookHints) -> None + if isinstance(hook_context.default_value, bool): + flags = sentry_sdk.get_current_scope().flags + flags.set(hook_context.flag_key, hook_context.default_value) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 0c0482904e..34ccc7f940 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -11,6 +11,7 @@ from sentry_sdk.attachments import Attachment from sentry_sdk.consts import DEFAULT_MAX_BREADCRUMBS, FALSE_VALUES, INSTRUMENTER +from sentry_sdk.flag_utils import FlagBuffer, DEFAULT_FLAG_CAPACITY from sentry_sdk.profiler.continuous_profiler import try_autostart_continuous_profiler from sentry_sdk.profiler.transaction_profiler import Profile from sentry_sdk.session import Session @@ -192,6 +193,7 @@ class Scope: "client", "_type", "_last_event_id", + "_flags", ) def __init__(self, ty=None, client=None): @@ -249,6 +251,8 @@ def __copy__(self): rv._last_event_id = self._last_event_id + rv._flags = copy(self._flags) + return rv @classmethod @@ -685,6 +689,7 @@ def clear(self): # self._last_event_id is only applicable to isolation scopes self._last_event_id = None # type: Optional[str] + self._flags = None # type: Optional[FlagBuffer] @_attr_setter def level(self, value): @@ -1546,6 +1551,17 @@ def __repr__(self): self._type, ) + @property + def flags(self): + # type: () -> FlagBuffer + if self._flags is None: + max_flags = ( + self.get_client().options["_experiments"].get("max_flags") + or DEFAULT_FLAG_CAPACITY + ) + self._flags = FlagBuffer(capacity=max_flags) + return self._flags + @contextmanager def new_scope(): diff --git a/setup.py b/setup.py index e9c83eb1fa..e5e0c8eaa4 100644 --- a/setup.py +++ b/setup.py @@ -63,9 +63,11 @@ def get_file_text(file_name): "huey": ["huey>=2"], "huggingface_hub": ["huggingface_hub>=0.22"], "langchain": ["langchain>=0.0.210"], + "launchdarkly": ["launchdarkly-server-sdk>=9.8.0"], "litestar": ["litestar>=2.0.0"], "loguru": ["loguru>=0.5"], "openai": ["openai>=1.0.0", "tiktoken>=0.3.0"], + "openfeature": ["openfeature-sdk>=0.7.1"], "opentelemetry": ["opentelemetry-distro>=0.35b0"], "opentelemetry-experimental": ["opentelemetry-distro"], "pure_eval": ["pure_eval", "executing", "asttokens"], diff --git a/tests/integrations/launchdarkly/__init__.py b/tests/integrations/launchdarkly/__init__.py new file mode 100644 index 0000000000..06e09884c8 --- /dev/null +++ b/tests/integrations/launchdarkly/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("ldclient") diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py new file mode 100644 index 0000000000..acbe764104 --- /dev/null +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -0,0 +1,116 @@ +import asyncio +import concurrent.futures as cf + +import ldclient + +import sentry_sdk +import pytest + +from ldclient import LDClient +from ldclient.config import Config +from ldclient.context import Context +from ldclient.integrations.test_data import TestData + +from sentry_sdk.integrations import DidNotEnable +from sentry_sdk.integrations.launchdarkly import LaunchDarklyIntegration + + +@pytest.mark.parametrize( + "use_global_client", + (False, True), +) +def test_launchdarkly_integration(sentry_init, use_global_client): + td = TestData.data_source() + config = Config("sdk-key", update_processor_class=td) + if use_global_client: + ldclient.set_config(config) + sentry_init(integrations=[LaunchDarklyIntegration()]) + client = ldclient.get() + else: + client = LDClient(config=config) + sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) + + # Set test values + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(True)) + + # Evaluate + client.variation("hello", Context.create("my-org", "organization"), False) + client.variation("world", Context.create("user1", "user"), False) + client.variation("other", Context.create("user2", "user"), False) + + assert sentry_sdk.get_current_scope().flags.get() == [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": True}, + {"flag": "other", "result": False}, + ] + + +def test_launchdarkly_integration_threaded(sentry_init): + td = TestData.data_source() + client = LDClient(config=Config("sdk-key", update_processor_class=td)) + sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) + context = Context.create("user1") + + def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + client.variation(flag_key, context, False) + return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(False)) + # Capture an eval before we split isolation scopes. + client.variation("hello", context, False) + + with cf.ThreadPoolExecutor(max_workers=2) as pool: + results = list(pool.map(task, ["world", "other"])) + + assert results[0] == ["hello", "world"] + assert results[1] == ["hello", "other"] + + +def test_launchdarkly_integration_asyncio(sentry_init): + """Assert concurrently evaluated flags do not pollute one another.""" + td = TestData.data_source() + client = LDClient(config=Config("sdk-key", update_processor_class=td)) + sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) + context = Context.create("user1") + + async def task(flag_key): + with sentry_sdk.isolation_scope(): + client.variation(flag_key, context, False) + return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + + async def runner(): + return asyncio.gather(task("world"), task("other")) + + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(False)) + client.variation("hello", context, False) + + results = asyncio.run(runner()).result() + assert results[0] == ["hello", "world"] + assert results[1] == ["hello", "other"] + + +def test_launchdarkly_integration_did_not_enable(monkeypatch): + # Client is not passed in and set_config wasn't called. + # TODO: Bad practice to access internals like this. We can skip this test, or remove this + # case entirely (force user to pass in a client instance). + ldclient._reset_client() + try: + ldclient.__lock.lock() + ldclient.__config = None + finally: + ldclient.__lock.unlock() + + with pytest.raises(DidNotEnable): + LaunchDarklyIntegration() + + # Client not initialized. + client = LDClient(config=Config("sdk-key")) + monkeypatch.setattr(client, "is_initialized", lambda: False) + with pytest.raises(DidNotEnable): + LaunchDarklyIntegration(ld_client=client) diff --git a/tests/integrations/openfeature/__init__.py b/tests/integrations/openfeature/__init__.py new file mode 100644 index 0000000000..a17549ea79 --- /dev/null +++ b/tests/integrations/openfeature/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("openfeature") diff --git a/tests/integrations/openfeature/test_openfeature.py b/tests/integrations/openfeature/test_openfeature.py new file mode 100644 index 0000000000..24e7857f9a --- /dev/null +++ b/tests/integrations/openfeature/test_openfeature.py @@ -0,0 +1,80 @@ +import asyncio +import concurrent.futures as cf +import sentry_sdk + +from openfeature import api +from openfeature.provider.in_memory_provider import InMemoryFlag, InMemoryProvider +from sentry_sdk.integrations.openfeature import OpenFeatureIntegration + + +def test_openfeature_integration(sentry_init): + sentry_init(integrations=[OpenFeatureIntegration()]) + + flags = { + "hello": InMemoryFlag("on", {"on": True, "off": False}), + "world": InMemoryFlag("off", {"on": True, "off": False}), + } + api.set_provider(InMemoryProvider(flags)) + + client = api.get_client() + client.get_boolean_value("hello", default_value=False) + client.get_boolean_value("world", default_value=False) + client.get_boolean_value("other", default_value=True) + + assert sentry_sdk.get_current_scope().flags.get() == [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + {"flag": "other", "result": True}, + ] + + +def test_openfeature_integration_threaded(sentry_init): + sentry_init(integrations=[OpenFeatureIntegration()]) + + flags = { + "hello": InMemoryFlag("on", {"on": True, "off": False}), + "world": InMemoryFlag("off", {"on": True, "off": False}), + } + api.set_provider(InMemoryProvider(flags)) + + client = api.get_client() + client.get_boolean_value("hello", default_value=False) + + def task(flag): + # Create a new isolation scope for the thread. This means the flags + with sentry_sdk.isolation_scope(): + client.get_boolean_value(flag, default_value=False) + return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + + with cf.ThreadPoolExecutor(max_workers=2) as pool: + results = list(pool.map(task, ["world", "other"])) + + assert results[0] == ["hello", "world"] + assert results[1] == ["hello", "other"] + + +def test_openfeature_integration_asyncio(sentry_init): + """Assert concurrently evaluated flags do not pollute one another.""" + + async def task(flag): + with sentry_sdk.isolation_scope(): + client.get_boolean_value(flag, default_value=False) + return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + + async def runner(): + return asyncio.gather(task("world"), task("other")) + + sentry_init(integrations=[OpenFeatureIntegration()]) + + flags = { + "hello": InMemoryFlag("on", {"on": True, "off": False}), + "world": InMemoryFlag("off", {"on": True, "off": False}), + } + api.set_provider(InMemoryProvider(flags)) + + client = api.get_client() + client.get_boolean_value("hello", default_value=False) + + results = asyncio.run(runner()).result() + assert results[0] == ["hello", "world"] + assert results[1] == ["hello", "other"] diff --git a/tests/test_flag_utils.py b/tests/test_flag_utils.py new file mode 100644 index 0000000000..3fa4f3abfe --- /dev/null +++ b/tests/test_flag_utils.py @@ -0,0 +1,43 @@ +from sentry_sdk.flag_utils import FlagBuffer + + +def test_flag_tracking(): + """Assert the ring buffer works.""" + buffer = FlagBuffer(capacity=3) + buffer.set("a", True) + flags = buffer.get() + assert len(flags) == 1 + assert flags == [{"flag": "a", "result": True}] + + buffer.set("b", True) + flags = buffer.get() + assert len(flags) == 2 + assert flags == [{"flag": "a", "result": True}, {"flag": "b", "result": True}] + + buffer.set("c", True) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "a", "result": True}, + {"flag": "b", "result": True}, + {"flag": "c", "result": True}, + ] + + buffer.set("d", False) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "b", "result": True}, + {"flag": "c", "result": True}, + {"flag": "d", "result": False}, + ] + + buffer.set("e", False) + buffer.set("f", False) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "d", "result": False}, + {"flag": "e", "result": False}, + {"flag": "f", "result": False}, + ] diff --git a/tests/test_lru_cache.py b/tests/test_lru_cache.py index 5343e76169..3e9c0ac964 100644 --- a/tests/test_lru_cache.py +++ b/tests/test_lru_cache.py @@ -35,3 +35,26 @@ def test_cache_eviction(): cache.set(4, 4) assert cache.get(3) is None assert cache.get(4) == 4 + + +def test_cache_miss(): + cache = LRUCache(1) + assert cache.get(0) is None + + +def test_cache_set_overwrite(): + cache = LRUCache(3) + cache.set(0, 0) + cache.set(0, 1) + assert cache.get(0) == 1 + + +def test_cache_get_all(): + cache = LRUCache(3) + cache.set(0, 0) + cache.set(1, 1) + cache.set(2, 2) + cache.set(3, 3) + assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] + cache.get(1) + assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] diff --git a/tox.ini b/tox.ini index ef30e539b5..f3a7ba4ea0 100644 --- a/tox.ini +++ b/tox.ini @@ -184,6 +184,14 @@ envlist = {py3.9,py3.11,py3.12}-openai-latest {py3.9,py3.11,py3.12}-openai-notiktoken + # OpenFeature + {py3.8,py3.12,py3.13}-openfeature-v0.7 + {py3.8,py3.12,py3.13}-openfeature-latest + + # LaunchDarkly + {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 + {py3.8,py3.12,py3.13}-launchdarkly-latest + # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -539,6 +547,14 @@ deps = openai-latest: tiktoken~=0.6.0 openai-notiktoken: openai + # OpenFeature + openfeature-v0.7: openfeature-sdk~=0.7.1 + openfeature-latest: openfeature-sdk + + # LaunchDarkly + launchdarkly-v9.8.0: launchdarkly-server-sdk~=9.8.0 + launchdarkly-latest: launchdarkly-server-sdk + # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -727,9 +743,11 @@ setenv = huey: TESTPATH=tests/integrations/huey huggingface_hub: TESTPATH=tests/integrations/huggingface_hub langchain: TESTPATH=tests/integrations/langchain + launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru openai: TESTPATH=tests/integrations/openai + openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry potel: TESTPATH=tests/integrations/opentelemetry pure_eval: TESTPATH=tests/integrations/pure_eval From 0a8ef922b8b5c933a5c0478622e2db0f1768244c Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 4 Nov 2024 13:16:51 +0000 Subject: [PATCH 021/591] release: 2.18.0 --- CHANGELOG.md | 29 +++++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 32 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2df6014abc..0bc4d1beb0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,34 @@ # Changelog +## 2.18.0 + +### Various fixes & improvements + +- Add LaunchDarkly and OpenFeature integration (#3648) by @cmanallen +- docs(hub): Correct typo in a comment (#3726) by @szokeasaurusrex +- fix(tracing): End http.client span on timeout (#3723) by @Zylphrex +- test: Fix UTC assuming test (#3722) by @BYK +- fix(http2): Check for h2 existence (#3690) by @BYK +- fix(profiling): Use `type()` instead when extracting frames (#3716) by @Zylphrex +- test(tornado): Unpin `pytest` for `tornado-latest` tests (#3714) by @szokeasaurusrex +- fix(starlette): Prefer python_multipart import over multipart (#3710) by @musicinmybrain +- test(redis): Install `pytest-asyncio` for `redis` tests (Python 3.12-13) (#3706) by @szokeasaurusrex +- ci: Clarify that only pinned tests are required (#3713) by @szokeasaurusrex +- test(rq): Remove accidentally-committed print (#3712) by @szokeasaurusrex +- test: Disable broken RQ test in newly-released RQ 2.0 (#3708) by @szokeasaurusrex +- test(tox): Unpin `pytest` for `celery` tests (#3701) by @szokeasaurusrex +- test(tox): Unpin `pytest` on Python 3.8+ `gevent` tests (#3700) by @szokeasaurusrex +- ci(tox): Unpin `pytest` for Python 3.8+ `common` tests (#3697) by @szokeasaurusrex +- ci: Run license compliance action on all PRs (#3699) by @szokeasaurusrex +- ci: Run CodeQL action on all PRs (#3698) by @szokeasaurusrex +- build: Remove pytest pin in requirements-devenv.txt (#3696) by @szokeasaurusrex +- ci(tox): Exclude fakeredis 2.26.0 on py3.6 and 3.7 (#3695) by @szokeasaurusrex +- fix(profiling): Update active thread for asgi (#3669) by @Zylphrex +- tests: Test with Falcon 4.0 (#3684) by @sentrivana +- fix(HTTP2Transport): Only enable HTTP2 when DSN is HTTPS (#3678) by @BYK +- fix(strawberry): prepare for upstream extension removal (#3649) by @DoctorJohn +- docs(sdk): Enhance README with improved clarity and developer-friendly examples (#3667) by @UTSAVS26 + ## 2.17.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 0489358dd9..6d33e5809a 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.17.0" +release = "2.18.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index fdb20caadf..ae32294d05 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -575,4 +575,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.17.0" +VERSION = "2.18.0" diff --git a/setup.py b/setup.py index e5e0c8eaa4..7ac4b56fde 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.17.0", + version="2.18.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 09946cb6246e700c4cfbdb880dda5751472249aa Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 4 Nov 2024 14:34:24 +0100 Subject: [PATCH 022/591] Update CHANGELOG.md --- CHANGELOG.md | 46 +++++++++++++++++++++++----------------------- 1 file changed, 23 insertions(+), 23 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 0bc4d1beb0..c47d0e0458 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,29 +5,29 @@ ### Various fixes & improvements - Add LaunchDarkly and OpenFeature integration (#3648) by @cmanallen -- docs(hub): Correct typo in a comment (#3726) by @szokeasaurusrex -- fix(tracing): End http.client span on timeout (#3723) by @Zylphrex -- test: Fix UTC assuming test (#3722) by @BYK -- fix(http2): Check for h2 existence (#3690) by @BYK -- fix(profiling): Use `type()` instead when extracting frames (#3716) by @Zylphrex -- test(tornado): Unpin `pytest` for `tornado-latest` tests (#3714) by @szokeasaurusrex -- fix(starlette): Prefer python_multipart import over multipart (#3710) by @musicinmybrain -- test(redis): Install `pytest-asyncio` for `redis` tests (Python 3.12-13) (#3706) by @szokeasaurusrex -- ci: Clarify that only pinned tests are required (#3713) by @szokeasaurusrex -- test(rq): Remove accidentally-committed print (#3712) by @szokeasaurusrex -- test: Disable broken RQ test in newly-released RQ 2.0 (#3708) by @szokeasaurusrex -- test(tox): Unpin `pytest` for `celery` tests (#3701) by @szokeasaurusrex -- test(tox): Unpin `pytest` on Python 3.8+ `gevent` tests (#3700) by @szokeasaurusrex -- ci(tox): Unpin `pytest` for Python 3.8+ `common` tests (#3697) by @szokeasaurusrex -- ci: Run license compliance action on all PRs (#3699) by @szokeasaurusrex -- ci: Run CodeQL action on all PRs (#3698) by @szokeasaurusrex -- build: Remove pytest pin in requirements-devenv.txt (#3696) by @szokeasaurusrex -- ci(tox): Exclude fakeredis 2.26.0 on py3.6 and 3.7 (#3695) by @szokeasaurusrex -- fix(profiling): Update active thread for asgi (#3669) by @Zylphrex -- tests: Test with Falcon 4.0 (#3684) by @sentrivana -- fix(HTTP2Transport): Only enable HTTP2 when DSN is HTTPS (#3678) by @BYK -- fix(strawberry): prepare for upstream extension removal (#3649) by @DoctorJohn -- docs(sdk): Enhance README with improved clarity and developer-friendly examples (#3667) by @UTSAVS26 +- Correct typo in a comment (#3726) by @szokeasaurusrex +- End `http.client` span on timeout (#3723) by @Zylphrex +- Check for `h2` existence in HTTP/2 transport (#3690) by @BYK +- Use `type()` instead when extracting frames (#3716) by @Zylphrex +- Prefer `python_multipart` import over `multipart` (#3710) by @musicinmybrain +- Update active thread for asgi (#3669) by @Zylphrex +- Only enable HTTP2 when DSN is HTTPS (#3678) by @BYK +- Prepare for upstream Strawberry extension removal (#3649) by @DoctorJohn +- Enhance README with improved clarity and developer-friendly examples (#3667) by @UTSAVS26 +- Run license compliance action on all PRs (#3699) by @szokeasaurusrex +- Run CodeQL action on all PRs (#3698) by @szokeasaurusrex +- Fix UTC assuming test (#3722) by @BYK +- Exclude fakeredis 2.26.0 on py3.6 and 3.7 (#3695) by @szokeasaurusrex +- Unpin `pytest` for `tornado-latest` tests (#3714) by @szokeasaurusrex +- Install `pytest-asyncio` for `redis` tests (Python 3.12-13) (#3706) by @szokeasaurusrex +- Clarify that only pinned tests are required (#3713) by @szokeasaurusrex +- Remove accidentally-committed print (#3712) by @szokeasaurusrex +- Disable broken RQ test in newly-released RQ 2.0 (#3708) by @szokeasaurusrex +- Unpin `pytest` for `celery` tests (#3701) by @szokeasaurusrex +- Unpin `pytest` on Python 3.8+ `gevent` tests (#3700) by @szokeasaurusrex +- Unpin `pytest` for Python 3.8+ `common` tests (#3697) by @szokeasaurusrex +- Remove `pytest` pin in `requirements-devenv.txt` (#3696) by @szokeasaurusrex +- Test with Falcon 4.0 (#3684) by @sentrivana ## 2.17.0 From e28dcf6bc0c3c83219e2336c57de380c3d76a934 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 5 Nov 2024 12:57:52 +0000 Subject: [PATCH 023/591] build(deps): bump actions/checkout from 4.2.1 to 4.2.2 (#3691) * build(deps): bump actions/checkout from 4.2.1 to 4.2.2 Bumps [actions/checkout](https://github.com/actions/checkout) from 4.2.1 to 4.2.2. - [Release notes](https://github.com/actions/checkout/releases) - [Changelog](https://github.com/actions/checkout/blob/main/CHANGELOG.md) - [Commits](https://github.com/actions/checkout/compare/v4.2.1...v4.2.2) --- updated-dependencies: - dependency-name: actions/checkout dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] --------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/ci.yml | 8 ++++---- .github/workflows/codeql-analysis.yml | 2 +- .github/workflows/release.yml | 2 +- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws-lambda.yml | 4 ++-- .github/workflows/test-integrations-cloud-computing.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-data-processing.yml | 4 ++-- .github/workflows/test-integrations-databases.yml | 4 ++-- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-miscellaneous.yml | 4 ++-- .github/workflows/test-integrations-networking.yml | 4 ++-- .github/workflows/test-integrations-web-frameworks-1.yml | 4 ++-- .github/workflows/test-integrations-web-frameworks-2.yml | 4 ++-- .../templates/check_permissions.jinja | 2 +- scripts/split-tox-gh-actions/templates/test_group.jinja | 2 +- 16 files changed, 29 insertions(+), 29 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 7e06911346..ed035b4ab0 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -24,7 +24,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -39,7 +39,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -54,7 +54,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -85,7 +85,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: 3.12 diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index d95353c652..e362d1e620 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -48,7 +48,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v4.2.1 + uses: actions/checkout@v4.2.2 # Initializes the CodeQL tools for scanning. - name: Initialize CodeQL diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index a2819a7591..268f62c4cc 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -18,7 +18,7 @@ jobs: runs-on: ubuntu-latest name: "Release a new version" steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 with: token: ${{ secrets.GH_RELEASE_PAT }} fetch-depth: 0 diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 24ccc77a87..dd230a6461 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -106,7 +106,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-aws-lambda.yml b/.github/workflows/test-integrations-aws-lambda.yml index 6f5ea794b8..c9837c08d0 100644 --- a/.github/workflows/test-integrations-aws-lambda.yml +++ b/.github/workflows/test-integrations-aws-lambda.yml @@ -32,7 +32,7 @@ jobs: name: permissions check runs-on: ubuntu-20.04 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 with: persist-credentials: false - name: Check permissions on PR @@ -67,7 +67,7 @@ jobs: os: [ubuntu-20.04] needs: check-permissions steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 with: ref: ${{ github.event.pull_request.head.sha || github.ref }} - uses: actions/setup-python@v5 diff --git a/.github/workflows/test-integrations-cloud-computing.yml b/.github/workflows/test-integrations-cloud-computing.yml index 1f6913ea4a..3217811539 100644 --- a/.github/workflows/test-integrations-cloud-computing.yml +++ b/.github/workflows/test-integrations-cloud-computing.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -102,7 +102,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index ecffdb6f3e..912eb3b18c 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-data-processing.yml b/.github/workflows/test-integrations-data-processing.yml index 49d18fc24c..128463a66a 100644 --- a/.github/workflows/test-integrations-data-processing.yml +++ b/.github/workflows/test-integrations-data-processing.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -120,7 +120,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-databases.yml b/.github/workflows/test-integrations-databases.yml index 49d3e923ee..2cdcd9d3b9 100644 --- a/.github/workflows/test-integrations-databases.yml +++ b/.github/workflows/test-integrations-databases.yml @@ -52,7 +52,7 @@ jobs: SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -147,7 +147,7 @@ jobs: SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 2cefb5d191..522dc2acc1 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -102,7 +102,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-miscellaneous.yml b/.github/workflows/test-integrations-miscellaneous.yml index 88a576505e..03d6559108 100644 --- a/.github/workflows/test-integrations-miscellaneous.yml +++ b/.github/workflows/test-integrations-miscellaneous.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -114,7 +114,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-networking.yml b/.github/workflows/test-integrations-networking.yml index c24edff174..31342151e9 100644 --- a/.github/workflows/test-integrations-networking.yml +++ b/.github/workflows/test-integrations-networking.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -102,7 +102,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-web-frameworks-1.yml b/.github/workflows/test-integrations-web-frameworks-1.yml index a655710843..706feb385f 100644 --- a/.github/workflows/test-integrations-web-frameworks-1.yml +++ b/.github/workflows/test-integrations-web-frameworks-1.yml @@ -52,7 +52,7 @@ jobs: SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -138,7 +138,7 @@ jobs: SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-web-frameworks-2.yml b/.github/workflows/test-integrations-web-frameworks-2.yml index d3f1001e2c..f700952e00 100644 --- a/.github/workflows/test-integrations-web-frameworks-2.yml +++ b/.github/workflows/test-integrations-web-frameworks-2.yml @@ -34,7 +34,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -126,7 +126,7 @@ jobs: # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} diff --git a/scripts/split-tox-gh-actions/templates/check_permissions.jinja b/scripts/split-tox-gh-actions/templates/check_permissions.jinja index e6d83b538a..390f447856 100644 --- a/scripts/split-tox-gh-actions/templates/check_permissions.jinja +++ b/scripts/split-tox-gh-actions/templates/check_permissions.jinja @@ -2,7 +2,7 @@ name: permissions check runs-on: ubuntu-20.04 steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 with: persist-credentials: false diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split-tox-gh-actions/templates/test_group.jinja index 5ee809aa96..9055070c72 100644 --- a/scripts/split-tox-gh-actions/templates/test_group.jinja +++ b/scripts/split-tox-gh-actions/templates/test_group.jinja @@ -39,7 +39,7 @@ {% endif %} steps: - - uses: actions/checkout@v4.2.1 + - uses: actions/checkout@v4.2.2 {% if needs_github_secrets %} {% raw %} with: From 24e5359580374ba474cbb2fb2837ed4c8a29cae6 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Tue, 5 Nov 2024 14:38:46 +0000 Subject: [PATCH 024/591] feat(spotlight): Add info logs when Sentry is enabled (#3735) This came as user feedback (getsentry/spotlight#543). Intentionally not making this part of Sentry logging as I think if one is enabling Spotlight, they should be seeing this in their logs, regardless of their SENTRY_DEBUG setting, which tends to be noisy. --- sentry_sdk/spotlight.py | 14 +++++++++----- 1 file changed, 9 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index b1ebf847ab..e7e90f9822 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -1,4 +1,5 @@ import io +import logging import os import urllib.parse import urllib.request @@ -108,11 +109,10 @@ def setup_spotlight(options): url = options.get("spotlight") - if isinstance(url, str): - pass - elif url is True: + if url is True: url = DEFAULT_SPOTLIGHT_URL - else: + + if not isinstance(url, str): return None if ( @@ -126,5 +126,9 @@ def setup_spotlight(options): settings.MIDDLEWARE = type(middleware)( chain(middleware, (DJANGO_SPOTLIGHT_MIDDLEWARE_PATH,)) ) + logging.info("Enabled Spotlight integration for Django") + + client = SpotlightClient(url) + logging.info("Enabled Spotlight at %s", url) - return SpotlightClient(url) + return client From c2dfbcc3c3de1c32de516ec4268a602cb42e0694 Mon Sep 17 00:00:00 2001 From: saber solooki Date: Wed, 6 Nov 2024 18:10:35 +0100 Subject: [PATCH 025/591] Fix(Arq): fix integration with Worker settings as a dict (#3742) --- sentry_sdk/integrations/arq.py | 11 +++ tests/integrations/arq/test_arq.py | 113 +++++++++++++++++++++++++---- 2 files changed, 110 insertions(+), 14 deletions(-) diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index 4640204725..d568714fe2 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -198,6 +198,17 @@ def _sentry_create_worker(*args, **kwargs): # type: (*Any, **Any) -> Worker settings_cls = args[0] + if isinstance(settings_cls, dict): + if "functions" in settings_cls: + settings_cls["functions"] = [ + _get_arq_function(func) for func in settings_cls["functions"] + ] + if "cron_jobs" in settings_cls: + settings_cls["cron_jobs"] = [ + _get_arq_cron_job(cron_job) + for cron_job in settings_cls["cron_jobs"] + ] + if hasattr(settings_cls, "functions"): settings_cls.functions = [ _get_arq_function(func) for func in settings_cls.functions diff --git a/tests/integrations/arq/test_arq.py b/tests/integrations/arq/test_arq.py index cd4cad67b8..e74395e26c 100644 --- a/tests/integrations/arq/test_arq.py +++ b/tests/integrations/arq/test_arq.py @@ -83,14 +83,65 @@ class WorkerSettings: return inner +@pytest.fixture +def init_arq_with_dict_settings(sentry_init): + def inner( + cls_functions=None, + cls_cron_jobs=None, + kw_functions=None, + kw_cron_jobs=None, + allow_abort_jobs_=False, + ): + cls_functions = cls_functions or [] + cls_cron_jobs = cls_cron_jobs or [] + + kwargs = {} + if kw_functions is not None: + kwargs["functions"] = kw_functions + if kw_cron_jobs is not None: + kwargs["cron_jobs"] = kw_cron_jobs + + sentry_init( + integrations=[ArqIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + server = FakeRedis() + pool = ArqRedis(pool_or_conn=server.connection_pool) + + worker_settings = { + "functions": cls_functions, + "cron_jobs": cls_cron_jobs, + "redis_pool": pool, + "allow_abort_jobs": allow_abort_jobs_, + } + + if not worker_settings["functions"]: + del worker_settings["functions"] + if not worker_settings["cron_jobs"]: + del worker_settings["cron_jobs"] + + worker = arq.worker.create_worker(worker_settings, **kwargs) + + return pool, worker + + return inner + + @pytest.mark.asyncio -async def test_job_result(init_arq): +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) +async def test_job_result(init_arq_settings, request): async def increase(ctx, num): return num + 1 + init_fixture_method = request.getfixturevalue(init_arq_settings) + increase.__qualname__ = increase.__name__ - pool, worker = init_arq([increase]) + pool, worker = init_fixture_method([increase]) job = await pool.enqueue_job("increase", 3) @@ -105,14 +156,19 @@ async def increase(ctx, num): @pytest.mark.asyncio -async def test_job_retry(capture_events, init_arq): +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) +async def test_job_retry(capture_events, init_arq_settings, request): async def retry_job(ctx): if ctx["job_try"] < 2: raise arq.worker.Retry + init_fixture_method = request.getfixturevalue(init_arq_settings) + retry_job.__qualname__ = retry_job.__name__ - pool, worker = init_arq([retry_job]) + pool, worker = init_fixture_method([retry_job]) job = await pool.enqueue_job("retry_job") @@ -139,11 +195,18 @@ async def retry_job(ctx): "source", [("cls_functions", "cls_cron_jobs"), ("kw_functions", "kw_cron_jobs")] ) @pytest.mark.parametrize("job_fails", [True, False], ids=["error", "success"]) +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) @pytest.mark.asyncio -async def test_job_transaction(capture_events, init_arq, source, job_fails): +async def test_job_transaction( + capture_events, init_arq_settings, source, job_fails, request +): async def division(_, a, b=0): return a / b + init_fixture_method = request.getfixturevalue(init_arq_settings) + division.__qualname__ = division.__name__ cron_func = async_partial(division, a=1, b=int(not job_fails)) @@ -152,7 +215,9 @@ async def division(_, a, b=0): cron_job = cron(cron_func, minute=0, run_at_startup=True) functions_key, cron_jobs_key = source - pool, worker = init_arq(**{functions_key: [division], cron_jobs_key: [cron_job]}) + pool, worker = init_fixture_method( + **{functions_key: [division], cron_jobs_key: [cron_job]} + ) events = capture_events() @@ -213,12 +278,17 @@ async def division(_, a, b=0): @pytest.mark.parametrize("source", ["cls_functions", "kw_functions"]) +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) @pytest.mark.asyncio -async def test_enqueue_job(capture_events, init_arq, source): +async def test_enqueue_job(capture_events, init_arq_settings, source, request): async def dummy_job(_): pass - pool, _ = init_arq(**{source: [dummy_job]}) + init_fixture_method = request.getfixturevalue(init_arq_settings) + + pool, _ = init_fixture_method(**{source: [dummy_job]}) events = capture_events() @@ -236,13 +306,18 @@ async def dummy_job(_): @pytest.mark.asyncio -async def test_execute_job_without_integration(init_arq): +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) +async def test_execute_job_without_integration(init_arq_settings, request): async def dummy_job(_ctx): pass + init_fixture_method = request.getfixturevalue(init_arq_settings) + dummy_job.__qualname__ = dummy_job.__name__ - pool, worker = init_arq([dummy_job]) + pool, worker = init_fixture_method([dummy_job]) # remove the integration to trigger the edge case get_client().integrations.pop("arq") @@ -254,12 +329,17 @@ async def dummy_job(_ctx): @pytest.mark.parametrize("source", ["cls_functions", "kw_functions"]) +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) @pytest.mark.asyncio -async def test_span_origin_producer(capture_events, init_arq, source): +async def test_span_origin_producer(capture_events, init_arq_settings, source, request): async def dummy_job(_): pass - pool, _ = init_arq(**{source: [dummy_job]}) + init_fixture_method = request.getfixturevalue(init_arq_settings) + + pool, _ = init_fixture_method(**{source: [dummy_job]}) events = capture_events() @@ -272,13 +352,18 @@ async def dummy_job(_): @pytest.mark.asyncio -async def test_span_origin_consumer(capture_events, init_arq): +@pytest.mark.parametrize( + "init_arq_settings", ["init_arq", "init_arq_with_dict_settings"] +) +async def test_span_origin_consumer(capture_events, init_arq_settings, request): async def job(ctx): pass + init_fixture_method = request.getfixturevalue(init_arq_settings) + job.__qualname__ = job.__name__ - pool, worker = init_arq([job]) + pool, worker = init_fixture_method([job]) job = await pool.enqueue_job("retry_job") From 200d0cdde8eed2caa89b91db8b17baabe983d2de Mon Sep 17 00:00:00 2001 From: Guilherme Martins Crocetti <24530683+gmcrocetti@users.noreply.github.com> Date: Thu, 7 Nov 2024 11:19:03 -0300 Subject: [PATCH 026/591] Handle parameter `stack_info` for the `LoggingIntegration` Add capability for the logging integration to use the parameter 'stack_info' (added in Python 3.2). When set to True the stack trace will be retrieved and properly handled. Fixes #2804 --- sentry_sdk/integrations/logging.py | 2 +- tests/integrations/logging/test_logging.py | 11 +++++++++-- 2 files changed, 10 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 5d23440ad1..b792510d6c 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -202,7 +202,7 @@ def _emit(self, record): client_options=client_options, mechanism={"type": "logging", "handled": True}, ) - elif record.exc_info and record.exc_info[0] is None: + elif (record.exc_info and record.exc_info[0] is None) or record.stack_info: event = {} hint = {} with capture_internal_exceptions(): diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 02eb26a04d..8c325bc86c 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -77,11 +77,18 @@ def test_logging_extra_data_integer_keys(sentry_init, capture_events): assert event["extra"] == {"1": 1} -def test_logging_stack(sentry_init, capture_events): +@pytest.mark.parametrize( + "enable_stack_trace_kwarg", + ( + pytest.param({"exc_info": True}, id="exc_info"), + pytest.param({"stack_info": True}, id="stack_info"), + ), +) +def test_logging_stack_trace(sentry_init, capture_events, enable_stack_trace_kwarg): sentry_init(integrations=[LoggingIntegration()], default_integrations=False) events = capture_events() - logger.error("first", exc_info=True) + logger.error("first", **enable_stack_trace_kwarg) logger.error("second") ( From d42422674379afd90ac5039e4fbac13281178ff2 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 12 Nov 2024 09:16:11 +0100 Subject: [PATCH 027/591] ref(init): Deprecate `sentry_sdk.init` context manager (#3729) It is possible to use the return value of `sentry_sdk.init` as a context manager; however, this functionality has not been maintained for a long time, and it does not seem to be documented anywhere. So, we are deprecating this functionality, and we will remove it in the next major release. Closes #3282 --- sentry_sdk/_init_implementation.py | 21 +++++++++++++++++++++ tests/test_api.py | 17 +++++++++++++++++ 2 files changed, 38 insertions(+) diff --git a/sentry_sdk/_init_implementation.py b/sentry_sdk/_init_implementation.py index 256a69ee83..eb02b3d11e 100644 --- a/sentry_sdk/_init_implementation.py +++ b/sentry_sdk/_init_implementation.py @@ -1,3 +1,5 @@ +import warnings + from typing import TYPE_CHECKING import sentry_sdk @@ -9,16 +11,35 @@ class _InitGuard: + _CONTEXT_MANAGER_DEPRECATION_WARNING_MESSAGE = ( + "Using the return value of sentry_sdk.init as a context manager " + "and manually calling the __enter__ and __exit__ methods on the " + "return value are deprecated. We are no longer maintaining this " + "functionality, and we will remove it in the next major release." + ) + def __init__(self, client): # type: (sentry_sdk.Client) -> None self._client = client def __enter__(self): # type: () -> _InitGuard + warnings.warn( + self._CONTEXT_MANAGER_DEPRECATION_WARNING_MESSAGE, + stacklevel=2, + category=DeprecationWarning, + ) + return self def __exit__(self, exc_type, exc_value, tb): # type: (Any, Any, Any) -> None + warnings.warn( + self._CONTEXT_MANAGER_DEPRECATION_WARNING_MESSAGE, + stacklevel=2, + category=DeprecationWarning, + ) + c = self._client if c is not None: c.close() diff --git a/tests/test_api.py b/tests/test_api.py index ae194af7fd..3b2a9c8fb7 100644 --- a/tests/test_api.py +++ b/tests/test_api.py @@ -1,6 +1,7 @@ import pytest from unittest import mock +import sentry_sdk from sentry_sdk import ( capture_exception, continue_trace, @@ -195,3 +196,19 @@ def test_push_scope_deprecation(): with pytest.warns(DeprecationWarning): with push_scope(): ... + + +def test_init_context_manager_deprecation(): + with pytest.warns(DeprecationWarning): + with sentry_sdk.init(): + ... + + +def test_init_enter_deprecation(): + with pytest.warns(DeprecationWarning): + sentry_sdk.init().__enter__() + + +def test_init_exit_deprecation(): + with pytest.warns(DeprecationWarning): + sentry_sdk.init().__exit__(None, None, None) From 417be9ffe5e2c72e459646dc7ec14399f78c015e Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Tue, 12 Nov 2024 13:28:51 +0000 Subject: [PATCH 028/591] feat(spotlight): Inject Spotlight button on Django (#3751) This patch expands the `SpotlightMiddleware` for Django and injects the Spotlight button to all HTML responses when Spotlight is enabled and running. It requires Spotlight 2.6.0 to work this way. Ref: getsentry/spotlight#543 --- sentry_sdk/spotlight.py | 159 ++++++++++++++++++++++++++++++++-------- 1 file changed, 130 insertions(+), 29 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index e7e90f9822..806ba5a09e 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -5,8 +5,9 @@ import urllib.request import urllib.error import urllib3 +import sys -from itertools import chain +from itertools import chain, product from typing import TYPE_CHECKING @@ -15,11 +16,19 @@ from typing import Callable from typing import Dict from typing import Optional + from typing import Self -from sentry_sdk.utils import logger, env_to_bool, capture_internal_exceptions +from sentry_sdk.utils import ( + logger as sentry_logger, + env_to_bool, + capture_internal_exceptions, +) from sentry_sdk.envelope import Envelope +logger = logging.getLogger("spotlight") + + DEFAULT_SPOTLIGHT_URL = "http://localhost:8969/stream" DJANGO_SPOTLIGHT_MIDDLEWARE_PATH = "sentry_sdk.spotlight.SpotlightMiddleware" @@ -34,7 +43,7 @@ def __init__(self, url): def capture_envelope(self, envelope): # type: (Envelope) -> None if self.tries > 3: - logger.warning( + sentry_logger.warning( "Too many errors sending to Spotlight, stop sending events there." ) return @@ -52,50 +61,137 @@ def capture_envelope(self, envelope): req.close() except Exception as e: self.tries += 1 - logger.warning(str(e)) + sentry_logger.warning(str(e)) try: - from django.http import HttpResponseServerError + from django.utils.deprecation import MiddlewareMixin + from django.http import HttpResponseServerError, HttpResponse, HttpRequest from django.conf import settings - class SpotlightMiddleware: - def __init__(self, get_response): - # type: (Any, Callable[..., Any]) -> None - self.get_response = get_response - - def __call__(self, request): - # type: (Any, Any) -> Any - return self.get_response(request) + SPOTLIGHT_JS_ENTRY_PATH = "/assets/main.js" + SPOTLIGHT_JS_SNIPPET_PATTERN = ( + '' + ) + SPOTLIGHT_ERROR_PAGE_SNIPPET = ( + '\n' + '\n' + ) + CHARSET_PREFIX = "charset=" + BODY_TAG_NAME = "body" + BODY_CLOSE_TAG_POSSIBILITIES = tuple( + "".format("".join(chars)) + for chars in product(*zip(BODY_TAG_NAME.upper(), BODY_TAG_NAME.lower())) + ) + + class SpotlightMiddleware(MiddlewareMixin): # type: ignore[misc] + _spotlight_script = None # type: Optional[str] - def process_exception(self, _request, exception): - # type: (Any, Any, Exception) -> Optional[HttpResponseServerError] - if not settings.DEBUG: - return None + def __init__(self, get_response): + # type: (Self, Callable[..., HttpResponse]) -> None + super().__init__(get_response) import sentry_sdk.api - spotlight_client = sentry_sdk.api.get_client().spotlight + self.sentry_sdk = sentry_sdk.api + + spotlight_client = self.sentry_sdk.get_client().spotlight if spotlight_client is None: + sentry_logger.warning( + "Cannot find Spotlight client from SpotlightMiddleware, disabling the middleware." + ) return None - # Spotlight URL has a trailing `/stream` part at the end so split it off - spotlight_url = spotlight_client.url.rsplit("/", 1)[0] + self._spotlight_url = urllib.parse.urljoin(spotlight_client.url, "../") + + @property + def spotlight_script(self): + # type: (Self) -> Optional[str] + if self._spotlight_script is None: + try: + spotlight_js_url = urllib.parse.urljoin( + self._spotlight_url, SPOTLIGHT_JS_ENTRY_PATH + ) + req = urllib.request.Request( + spotlight_js_url, + method="HEAD", + ) + urllib.request.urlopen(req) + self._spotlight_script = SPOTLIGHT_JS_SNIPPET_PATTERN.format( + spotlight_js_url + ) + except urllib.error.URLError as err: + sentry_logger.debug( + "Cannot get Spotlight JS to inject at %s. SpotlightMiddleware will not be very useful.", + spotlight_js_url, + exc_info=err, + ) + + return self._spotlight_script + + def process_response(self, _request, response): + # type: (Self, HttpRequest, HttpResponse) -> Optional[HttpResponse] + content_type_header = tuple( + p.strip() + for p in response.headers.get("Content-Type", "").lower().split(";") + ) + content_type = content_type_header[0] + if len(content_type_header) > 1 and content_type_header[1].startswith( + CHARSET_PREFIX + ): + encoding = content_type_header[1][len(CHARSET_PREFIX) :] + else: + encoding = "utf-8" + + if ( + self.spotlight_script is not None + and not response.streaming + and content_type == "text/html" + ): + content_length = len(response.content) + injection = self.spotlight_script.encode(encoding) + injection_site = next( + ( + idx + for idx in ( + response.content.rfind(body_variant.encode(encoding)) + for body_variant in BODY_CLOSE_TAG_POSSIBILITIES + ) + if idx > -1 + ), + content_length, + ) + + # This approach works even when we don't have a `` tag + response.content = ( + response.content[:injection_site] + + injection + + response.content[injection_site:] + ) + + if response.has_header("Content-Length"): + response.headers["Content-Length"] = content_length + len(injection) + + return response + + def process_exception(self, _request, exception): + # type: (Self, HttpRequest, Exception) -> Optional[HttpResponseServerError] + if not settings.DEBUG: + return None try: - spotlight = urllib.request.urlopen(spotlight_url).read().decode("utf-8") + spotlight = ( + urllib.request.urlopen(self._spotlight_url).read().decode("utf-8") + ) except urllib.error.URLError: return None else: - event_id = sentry_sdk.api.capture_exception(exception) + event_id = self.sentry_sdk.capture_exception(exception) return HttpResponseServerError( spotlight.replace( "", - ( - f'' - ''.format( - event_id=event_id - ) + SPOTLIGHT_ERROR_PAGE_SNIPPET.format( + spotlight_url=self._spotlight_url, event_id=event_id ), ) ) @@ -106,6 +202,10 @@ def process_exception(self, _request, exception): def setup_spotlight(options): # type: (Dict[str, Any]) -> Optional[SpotlightClient] + _handler = logging.StreamHandler(sys.stderr) + _handler.setFormatter(logging.Formatter(" [spotlight] %(levelname)s: %(message)s")) + logger.addHandler(_handler) + logger.setLevel(logging.INFO) url = options.get("spotlight") @@ -119,6 +219,7 @@ def setup_spotlight(options): settings is not None and settings.DEBUG and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_ON_ERROR", "1")) + and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_MIDDLEWARE", "1")) ): with capture_internal_exceptions(): middleware = settings.MIDDLEWARE @@ -126,9 +227,9 @@ def setup_spotlight(options): settings.MIDDLEWARE = type(middleware)( chain(middleware, (DJANGO_SPOTLIGHT_MIDDLEWARE_PATH,)) ) - logging.info("Enabled Spotlight integration for Django") + logger.info("Enabled Spotlight integration for Django") client = SpotlightClient(url) - logging.info("Enabled Spotlight at %s", url) + logger.info("Enabled Spotlight using sidecar at %s", url) return client From c2361a32d58eb38465e41c967788cae991a4e510 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 13 Nov 2024 13:50:01 +0100 Subject: [PATCH 029/591] Fix aws lambda tests (by reducing event size) (#3770) Our AWS Lambda tests rely on outputting our events as JSON to stdout and parsing this output. AWS Lambda limits the amount of stdout it returns. So by reducing the size of the events we can fix the tests, that where broken by printing to much data to stdout so the output is truncated and can not be parsed into actual JSON structures again. --- tests/integrations/aws_lambda/test_aws.py | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/tests/integrations/aws_lambda/test_aws.py b/tests/integrations/aws_lambda/test_aws.py index 75dc930da5..e229812336 100644 --- a/tests/integrations/aws_lambda/test_aws.py +++ b/tests/integrations/aws_lambda/test_aws.py @@ -98,7 +98,7 @@ def truncate_data(data): elif key == "cloudwatch logs": for cloudwatch_key in data["extra"]["cloudwatch logs"].keys(): if cloudwatch_key in ["url", "log_group", "log_stream"]: - cleaned_data["extra"].setdefault("cloudwatch logs", {})[cloudwatch_key] = data["extra"]["cloudwatch logs"][cloudwatch_key] + cleaned_data["extra"].setdefault("cloudwatch logs", {})[cloudwatch_key] = data["extra"]["cloudwatch logs"][cloudwatch_key].split("=")[0] if data.get("level") is not None: cleaned_data["level"] = data.get("level") @@ -228,7 +228,7 @@ def test_handler(event, context): assert event["extra"]["lambda"]["function_name"].startswith("test_") logs_url = event["extra"]["cloudwatch logs"]["url"] - assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region=") + assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region") assert not re.search("(=;|=$)", logs_url) assert event["extra"]["cloudwatch logs"]["log_group"].startswith( "/aws/lambda/test_" @@ -370,7 +370,7 @@ def test_handler(event, context): assert event["extra"]["lambda"]["function_name"].startswith("test_") logs_url = event["extra"]["cloudwatch logs"]["url"] - assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region=") + assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region") assert not re.search("(=;|=$)", logs_url) assert event["extra"]["cloudwatch logs"]["log_group"].startswith( "/aws/lambda/test_" @@ -462,11 +462,11 @@ def test_handler(event, context): "X-Forwarded-Proto": "https" }, "httpMethod": "GET", - "path": "/path1", + "path": "/1", "queryStringParameters": { - "done": "false" + "done": "f" }, - "dog": "Maisey" + "d": "D1" }, { "headers": { @@ -474,11 +474,11 @@ def test_handler(event, context): "X-Forwarded-Proto": "http" }, "httpMethod": "POST", - "path": "/path2", + "path": "/2", "queryStringParameters": { - "done": "true" + "done": "t" }, - "dog": "Charlie" + "d": "D2" } ] """, @@ -538,9 +538,9 @@ def test_handler(event, context): request_data = { "headers": {"Host": "x1.io", "X-Forwarded-Proto": "https"}, "method": "GET", - "url": "https://x1.io/path1", + "url": "https://x1.io/1", "query_string": { - "done": "false", + "done": "f", }, } else: From 4bec4a4729b64525ef55947fd4042e0d62ef72cc Mon Sep 17 00:00:00 2001 From: matt-codecov <137832199+matt-codecov@users.noreply.github.com> Date: Wed, 13 Nov 2024 05:30:58 -0800 Subject: [PATCH 030/591] feat: introduce rust_tracing integration (#3717) Introduce a new integration that allows traces to descend into code in Rust native extensions by hooking into Rust's popular `tracing` framework. it relies on the Rust native extension using [`pyo3-python-tracing-subscriber`](https://crates.io/crates/pyo3-python-tracing-subscriber), a crate i recently published under Sentry, to expose a way for the Python SDK to hook into `tracing`. in this screenshot, the transaction was started in Python but the rest of the span tree reflects the structure and performance of a naive fibonacci generator in Rust: https://github.com/user-attachments/assets/ae2caff6-1842-45d0-a604-2f3b6305f330 --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/rust_tracing.py | 274 +++++++++++ tests/integrations/rust_tracing/__init__.py | 0 .../rust_tracing/test_rust_tracing.py | 450 ++++++++++++++++++ 3 files changed, 724 insertions(+) create mode 100644 sentry_sdk/integrations/rust_tracing.py create mode 100644 tests/integrations/rust_tracing/__init__.py create mode 100644 tests/integrations/rust_tracing/test_rust_tracing.py diff --git a/sentry_sdk/integrations/rust_tracing.py b/sentry_sdk/integrations/rust_tracing.py new file mode 100644 index 0000000000..121bf082b8 --- /dev/null +++ b/sentry_sdk/integrations/rust_tracing.py @@ -0,0 +1,274 @@ +""" +This integration ingests tracing data from native extensions written in Rust. + +Using it requires additional setup on the Rust side to accept a +`RustTracingLayer` Python object and register it with the `tracing-subscriber` +using an adapter from the `pyo3-python-tracing-subscriber` crate. For example: +```rust +#[pyfunction] +pub fn initialize_tracing(py_impl: Bound<'_, PyAny>) { + tracing_subscriber::registry() + .with(pyo3_python_tracing_subscriber::PythonCallbackLayerBridge::new(py_impl)) + .init(); +} +``` + +Usage in Python would then look like: +``` +sentry_sdk.init( + dsn=sentry_dsn, + integrations=[ + RustTracingIntegration( + "demo_rust_extension", + demo_rust_extension.initialize_tracing, + event_type_mapping=event_type_mapping, + ) + ], +) +``` + +Each native extension requires its own integration. +""" + +import json +from enum import Enum, auto +from typing import Any, Callable, Dict, Tuple, Optional + +import sentry_sdk +from sentry_sdk.integrations import Integration +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.tracing import Span as SentrySpan +from sentry_sdk.utils import SENSITIVE_DATA_SUBSTITUTE + +TraceState = Optional[Tuple[Optional[SentrySpan], SentrySpan]] + + +class RustTracingLevel(Enum): + Trace: str = "TRACE" + Debug: str = "DEBUG" + Info: str = "INFO" + Warn: str = "WARN" + Error: str = "ERROR" + + +class EventTypeMapping(Enum): + Ignore = auto() + Exc = auto() + Breadcrumb = auto() + Event = auto() + + +def tracing_level_to_sentry_level(level): + # type: (str) -> sentry_sdk._types.LogLevelStr + level = RustTracingLevel(level) + if level in (RustTracingLevel.Trace, RustTracingLevel.Debug): + return "debug" + elif level == RustTracingLevel.Info: + return "info" + elif level == RustTracingLevel.Warn: + return "warning" + elif level == RustTracingLevel.Error: + return "error" + else: + # Better this than crashing + return "info" + + +def extract_contexts(event: Dict[str, Any]) -> Dict[str, Any]: + metadata = event.get("metadata", {}) + contexts = {} + + location = {} + for field in ["module_path", "file", "line"]: + if field in metadata: + location[field] = metadata[field] + if len(location) > 0: + contexts["rust_tracing_location"] = location + + fields = {} + for field in metadata.get("fields", []): + fields[field] = event.get(field) + if len(fields) > 0: + contexts["rust_tracing_fields"] = fields + + return contexts + + +def process_event(event: Dict[str, Any]) -> None: + metadata = event.get("metadata", {}) + + logger = metadata.get("target") + level = tracing_level_to_sentry_level(metadata.get("level")) + message = event.get("message") # type: sentry_sdk._types.Any + contexts = extract_contexts(event) + + sentry_event = { + "logger": logger, + "level": level, + "message": message, + "contexts": contexts, + } # type: sentry_sdk._types.Event + + sentry_sdk.capture_event(sentry_event) + + +def process_exception(event: Dict[str, Any]) -> None: + process_event(event) + + +def process_breadcrumb(event: Dict[str, Any]) -> None: + level = tracing_level_to_sentry_level(event.get("metadata", {}).get("level")) + message = event.get("message") + + sentry_sdk.add_breadcrumb(level=level, message=message) + + +def default_span_filter(metadata: Dict[str, Any]) -> bool: + return RustTracingLevel(metadata.get("level")) in ( + RustTracingLevel.Error, + RustTracingLevel.Warn, + RustTracingLevel.Info, + ) + + +def default_event_type_mapping(metadata: Dict[str, Any]) -> EventTypeMapping: + level = RustTracingLevel(metadata.get("level")) + if level == RustTracingLevel.Error: + return EventTypeMapping.Exc + elif level in (RustTracingLevel.Warn, RustTracingLevel.Info): + return EventTypeMapping.Breadcrumb + elif level in (RustTracingLevel.Debug, RustTracingLevel.Trace): + return EventTypeMapping.Ignore + else: + return EventTypeMapping.Ignore + + +class RustTracingLayer: + def __init__( + self, + origin: str, + event_type_mapping: Callable[ + [Dict[str, Any]], EventTypeMapping + ] = default_event_type_mapping, + span_filter: Callable[[Dict[str, Any]], bool] = default_span_filter, + send_sensitive_data: Optional[bool] = None, + ): + self.origin = origin + self.event_type_mapping = event_type_mapping + self.span_filter = span_filter + self.send_sensitive_data = send_sensitive_data + + def on_event(self, event: str, _span_state: TraceState) -> None: + deserialized_event = json.loads(event) + metadata = deserialized_event.get("metadata", {}) + + event_type = self.event_type_mapping(metadata) + if event_type == EventTypeMapping.Ignore: + return + elif event_type == EventTypeMapping.Exc: + process_exception(deserialized_event) + elif event_type == EventTypeMapping.Breadcrumb: + process_breadcrumb(deserialized_event) + elif event_type == EventTypeMapping.Event: + process_event(deserialized_event) + + def on_new_span(self, attrs: str, span_id: str) -> TraceState: + attrs = json.loads(attrs) + metadata = attrs.get("metadata", {}) + + if not self.span_filter(metadata): + return None + + module_path = metadata.get("module_path") + name = metadata.get("name") + message = attrs.get("message") + + if message is not None: + sentry_span_name = message + elif module_path is not None and name is not None: + sentry_span_name = f"{module_path}::{name}" # noqa: E231 + elif name is not None: + sentry_span_name = name + else: + sentry_span_name = "" + + kwargs = { + "op": "function", + "name": sentry_span_name, + "origin": self.origin, + } + + scope = sentry_sdk.get_current_scope() + parent_sentry_span = scope.span + if parent_sentry_span: + sentry_span = parent_sentry_span.start_child(**kwargs) + else: + sentry_span = scope.start_span(**kwargs) + + fields = metadata.get("fields", []) + for field in fields: + sentry_span.set_data(field, attrs.get(field)) + + scope.span = sentry_span + return (parent_sentry_span, sentry_span) + + def on_close(self, span_id: str, span_state: TraceState) -> None: + if span_state is None: + return + + parent_sentry_span, sentry_span = span_state + sentry_span.finish() + sentry_sdk.get_current_scope().span = parent_sentry_span + + def on_record(self, span_id: str, values: str, span_state: TraceState) -> None: + if span_state is None: + return + _parent_sentry_span, sentry_span = span_state + + send_sensitive_data = ( + should_send_default_pii() + if self.send_sensitive_data is None + else self.send_sensitive_data + ) + + deserialized_values = json.loads(values) + for key, value in deserialized_values.items(): + if send_sensitive_data: + sentry_span.set_data(key, value) + else: + sentry_span.set_data(key, SENSITIVE_DATA_SUBSTITUTE) + + +class RustTracingIntegration(Integration): + """ + Ingests tracing data from a Rust native extension's `tracing` instrumentation. + + If a project uses more than one Rust native extension, each one will need + its own instance of `RustTracingIntegration` with an initializer function + specific to that extension. + + Since all of the setup for this integration requires instance-specific state + which is not available in `setup_once()`, setup instead happens in `__init__()`. + """ + + def __init__( + self, + identifier: str, + initializer: Callable[[RustTracingLayer], None], + event_type_mapping: Callable[ + [Dict[str, Any]], EventTypeMapping + ] = default_event_type_mapping, + span_filter: Callable[[Dict[str, Any]], bool] = default_span_filter, + send_sensitive_data: Optional[bool] = None, + ): + self.identifier = identifier + origin = f"auto.function.rust_tracing.{identifier}" + self.tracing_layer = RustTracingLayer( + origin, event_type_mapping, span_filter, send_sensitive_data + ) + + initializer(self.tracing_layer) + + @staticmethod + def setup_once() -> None: + pass diff --git a/tests/integrations/rust_tracing/__init__.py b/tests/integrations/rust_tracing/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/rust_tracing/test_rust_tracing.py b/tests/integrations/rust_tracing/test_rust_tracing.py new file mode 100644 index 0000000000..b1fad1a7f7 --- /dev/null +++ b/tests/integrations/rust_tracing/test_rust_tracing.py @@ -0,0 +1,450 @@ +import pytest + +from string import Template +from typing import Dict + +import sentry_sdk +from sentry_sdk.integrations.rust_tracing import ( + RustTracingIntegration, + RustTracingLayer, + RustTracingLevel, + EventTypeMapping, +) +from sentry_sdk import start_transaction, capture_message + + +def _test_event_type_mapping(metadata: Dict[str, object]) -> EventTypeMapping: + level = RustTracingLevel(metadata.get("level")) + if level == RustTracingLevel.Error: + return EventTypeMapping.Exc + elif level in (RustTracingLevel.Warn, RustTracingLevel.Info): + return EventTypeMapping.Breadcrumb + elif level == RustTracingLevel.Debug: + return EventTypeMapping.Event + elif level == RustTracingLevel.Trace: + return EventTypeMapping.Ignore + else: + return EventTypeMapping.Ignore + + +class FakeRustTracing: + # Parameters: `level`, `index` + span_template = Template( + """{"index":$index,"is_root":false,"metadata":{"fields":["index","use_memoized","version"],"file":"src/lib.rs","is_event":false,"is_span":true,"level":"$level","line":40,"module_path":"_bindings","name":"fibonacci","target":"_bindings"},"parent":null,"use_memoized":true}""" + ) + + # Parameters: `level`, `index` + event_template = Template( + """{"message":"Getting the ${index}th fibonacci number","metadata":{"fields":["message"],"file":"src/lib.rs","is_event":true,"is_span":false,"level":"$level","line":23,"module_path":"_bindings","name":"event src/lib.rs:23","target":"_bindings"}}""" + ) + + def __init__(self): + self.spans = {} + + def set_layer_impl(self, layer: RustTracingLayer): + self.layer = layer + + def new_span(self, level: RustTracingLevel, span_id: int, index_arg: int = 10): + span_attrs = self.span_template.substitute(level=level.value, index=index_arg) + state = self.layer.on_new_span(span_attrs, str(span_id)) + self.spans[span_id] = state + + def close_span(self, span_id: int): + state = self.spans.pop(span_id) + self.layer.on_close(str(span_id), state) + + def event(self, level: RustTracingLevel, span_id: int, index_arg: int = 10): + event = self.event_template.substitute(level=level.value, index=index_arg) + state = self.spans[span_id] + self.layer.on_event(event, state) + + def record(self, span_id: int): + state = self.spans[span_id] + self.layer.on_record(str(span_id), """{"version": "memoized"}""", state) + + +def test_on_new_span_on_close(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_new_span_on_close", rust_tracing.set_layer_impl + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + sentry_first_rust_span = sentry_sdk.get_current_span() + _, rust_first_rust_span = rust_tracing.spans[3] + + assert sentry_first_rust_span == rust_first_rust_span + + rust_tracing.close_span(3) + assert sentry_sdk.get_current_span() != sentry_first_rust_span + + (event,) = events + assert len(event["spans"]) == 1 + + # Ensure the span metadata is wired up + span = event["spans"][0] + assert span["op"] == "function" + assert span["origin"] == "auto.function.rust_tracing.test_on_new_span_on_close" + assert span["description"] == "_bindings::fibonacci" + + # Ensure the span was opened/closed appropriately + assert span["start_timestamp"] is not None + assert span["timestamp"] is not None + + # Ensure the extra data from Rust is hooked up + data = span["data"] + assert data["use_memoized"] + assert data["index"] == 10 + assert data["version"] is None + + +def test_nested_on_new_span_on_close(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_nested_on_new_span_on_close", rust_tracing.set_layer_impl + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + with start_transaction(): + original_sentry_span = sentry_sdk.get_current_span() + + rust_tracing.new_span(RustTracingLevel.Info, 3, index_arg=10) + sentry_first_rust_span = sentry_sdk.get_current_span() + _, rust_first_rust_span = rust_tracing.spans[3] + + # Use a different `index_arg` value for the inner span to help + # distinguish the two at the end of the test + rust_tracing.new_span(RustTracingLevel.Info, 5, index_arg=9) + sentry_second_rust_span = sentry_sdk.get_current_span() + rust_parent_span, rust_second_rust_span = rust_tracing.spans[5] + + assert rust_second_rust_span == sentry_second_rust_span + assert rust_parent_span == sentry_first_rust_span + assert rust_parent_span == rust_first_rust_span + assert rust_parent_span != rust_second_rust_span + + rust_tracing.close_span(5) + + # Ensure the current sentry span was moved back to the parent + sentry_span_after_close = sentry_sdk.get_current_span() + assert sentry_span_after_close == sentry_first_rust_span + + rust_tracing.close_span(3) + + assert sentry_sdk.get_current_span() == original_sentry_span + + (event,) = events + assert len(event["spans"]) == 2 + + # Ensure the span metadata is wired up for all spans + first_span, second_span = event["spans"] + assert first_span["op"] == "function" + assert ( + first_span["origin"] + == "auto.function.rust_tracing.test_nested_on_new_span_on_close" + ) + assert first_span["description"] == "_bindings::fibonacci" + assert second_span["op"] == "function" + assert ( + second_span["origin"] + == "auto.function.rust_tracing.test_nested_on_new_span_on_close" + ) + assert second_span["description"] == "_bindings::fibonacci" + + # Ensure the spans were opened/closed appropriately + assert first_span["start_timestamp"] is not None + assert first_span["timestamp"] is not None + assert second_span["start_timestamp"] is not None + assert second_span["timestamp"] is not None + + # Ensure the extra data from Rust is hooked up in both spans + first_span_data = first_span["data"] + assert first_span_data["use_memoized"] + assert first_span_data["index"] == 10 + assert first_span_data["version"] is None + + second_span_data = second_span["data"] + assert second_span_data["use_memoized"] + assert second_span_data["index"] == 9 + assert second_span_data["version"] is None + + +def test_on_new_span_without_transaction(sentry_init): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_new_span_without_transaction", rust_tracing.set_layer_impl + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + assert sentry_sdk.get_current_span() is None + + # Should still create a span hierarchy, it just will not be under a txn + rust_tracing.new_span(RustTracingLevel.Info, 3) + current_span = sentry_sdk.get_current_span() + assert current_span is not None + assert current_span.containing_transaction is None + + +def test_on_event_exception(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_event_exception", + rust_tracing.set_layer_impl, + event_type_mapping=_test_event_type_mapping, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + sentry_sdk.get_isolation_scope().clear_breadcrumbs() + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + # Mapped to Exception + rust_tracing.event(RustTracingLevel.Error, 3) + + rust_tracing.close_span(3) + + assert len(events) == 2 + exc, _tx = events + assert exc["level"] == "error" + assert exc["logger"] == "_bindings" + assert exc["message"] == "Getting the 10th fibonacci number" + assert exc["breadcrumbs"]["values"] == [] + + location_context = exc["contexts"]["rust_tracing_location"] + assert location_context["module_path"] == "_bindings" + assert location_context["file"] == "src/lib.rs" + assert location_context["line"] == 23 + + field_context = exc["contexts"]["rust_tracing_fields"] + assert field_context["message"] == "Getting the 10th fibonacci number" + + +def test_on_event_breadcrumb(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_event_breadcrumb", + rust_tracing.set_layer_impl, + event_type_mapping=_test_event_type_mapping, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + sentry_sdk.get_isolation_scope().clear_breadcrumbs() + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + # Mapped to Breadcrumb + rust_tracing.event(RustTracingLevel.Info, 3) + + rust_tracing.close_span(3) + capture_message("test message") + + assert len(events) == 2 + message, _tx = events + + breadcrumbs = message["breadcrumbs"]["values"] + assert len(breadcrumbs) == 1 + assert breadcrumbs[0]["level"] == "info" + assert breadcrumbs[0]["message"] == "Getting the 10th fibonacci number" + assert breadcrumbs[0]["type"] == "default" + + +def test_on_event_event(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_event_event", + rust_tracing.set_layer_impl, + event_type_mapping=_test_event_type_mapping, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + sentry_sdk.get_isolation_scope().clear_breadcrumbs() + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + # Mapped to Event + rust_tracing.event(RustTracingLevel.Debug, 3) + + rust_tracing.close_span(3) + + assert len(events) == 2 + event, _tx = events + + assert event["logger"] == "_bindings" + assert event["level"] == "debug" + assert event["message"] == "Getting the 10th fibonacci number" + assert event["breadcrumbs"]["values"] == [] + + location_context = event["contexts"]["rust_tracing_location"] + assert location_context["module_path"] == "_bindings" + assert location_context["file"] == "src/lib.rs" + assert location_context["line"] == 23 + + field_context = event["contexts"]["rust_tracing_fields"] + assert field_context["message"] == "Getting the 10th fibonacci number" + + +def test_on_event_ignored(sentry_init, capture_events): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_on_event_ignored", + rust_tracing.set_layer_impl, + event_type_mapping=_test_event_type_mapping, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + sentry_sdk.get_isolation_scope().clear_breadcrumbs() + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + # Ignored + rust_tracing.event(RustTracingLevel.Trace, 3) + + rust_tracing.close_span(3) + + assert len(events) == 1 + (tx,) = events + assert tx["type"] == "transaction" + assert "message" not in tx + + +def test_span_filter(sentry_init, capture_events): + def span_filter(metadata: Dict[str, object]) -> bool: + return RustTracingLevel(metadata.get("level")) in ( + RustTracingLevel.Error, + RustTracingLevel.Warn, + RustTracingLevel.Info, + RustTracingLevel.Debug, + ) + + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_span_filter", rust_tracing.set_layer_impl, span_filter=span_filter + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + events = capture_events() + with start_transaction(): + original_sentry_span = sentry_sdk.get_current_span() + + # Span is not ignored + rust_tracing.new_span(RustTracingLevel.Info, 3, index_arg=10) + info_span = sentry_sdk.get_current_span() + + # Span is ignored, current span should remain the same + rust_tracing.new_span(RustTracingLevel.Trace, 5, index_arg=9) + assert sentry_sdk.get_current_span() == info_span + + # Closing the filtered span should leave the current span alone + rust_tracing.close_span(5) + assert sentry_sdk.get_current_span() == info_span + + rust_tracing.close_span(3) + assert sentry_sdk.get_current_span() == original_sentry_span + + (event,) = events + assert len(event["spans"]) == 1 + # The ignored span has index == 9 + assert event["spans"][0]["data"]["index"] == 10 + + +def test_record(sentry_init): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_record", + initializer=rust_tracing.set_layer_impl, + send_sensitive_data=True, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + span_before_record = sentry_sdk.get_current_span().to_json() + assert span_before_record["data"]["version"] is None + + rust_tracing.record(3) + + span_after_record = sentry_sdk.get_current_span().to_json() + assert span_after_record["data"]["version"] == "memoized" + + +def test_record_in_ignored_span(sentry_init): + def span_filter(metadata: Dict[str, object]) -> bool: + # Just ignore Trace + return RustTracingLevel(metadata.get("level")) != RustTracingLevel.Trace + + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_record_in_ignored_span", + rust_tracing.set_layer_impl, + span_filter=span_filter, + ) + sentry_init(integrations=[integration], traces_sample_rate=1.0) + + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + span_before_record = sentry_sdk.get_current_span().to_json() + assert span_before_record["data"]["version"] is None + + rust_tracing.new_span(RustTracingLevel.Trace, 5) + rust_tracing.record(5) + + # `on_record()` should not do anything to the current Sentry span if the associated Rust span was ignored + span_after_record = sentry_sdk.get_current_span().to_json() + assert span_after_record["data"]["version"] is None + + +@pytest.mark.parametrize( + "send_default_pii, send_sensitive_data, sensitive_data_expected", + [ + (True, True, True), + (True, False, False), + (True, None, True), + (False, True, True), + (False, False, False), + (False, None, False), + ], +) +def test_sensitive_data( + sentry_init, send_default_pii, send_sensitive_data, sensitive_data_expected +): + rust_tracing = FakeRustTracing() + integration = RustTracingIntegration( + "test_record", + initializer=rust_tracing.set_layer_impl, + send_sensitive_data=send_sensitive_data, + ) + + sentry_init( + integrations=[integration], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + with start_transaction(): + rust_tracing.new_span(RustTracingLevel.Info, 3) + + span_before_record = sentry_sdk.get_current_span().to_json() + assert span_before_record["data"]["version"] is None + + rust_tracing.record(3) + + span_after_record = sentry_sdk.get_current_span().to_json() + + if sensitive_data_expected: + assert span_after_record["data"]["version"] == "memoized" + else: + assert span_after_record["data"]["version"] == "[Filtered]" From da0b086333e03292da97993cf3e718fa1e9937a5 Mon Sep 17 00:00:00 2001 From: matt-codecov <137832199+matt-codecov@users.noreply.github.com> Date: Thu, 14 Nov 2024 23:55:56 -0800 Subject: [PATCH 031/591] fix: include_tracing_fields arg to control unvetted data in rust_tracing integration (#3780) Rename `send_sensitive_data` flag to `include_tracing_fields`. the data in question is generally data the user expects `tracing` to record or data they explicitly passed into a log statement to be recorded, so if we call it "sensitive" they may think we are referring to something else also, apply the same condition to both `on_record()` and `on_new_span()`. both callbacks set the same fields, so they should either both be redacted or both be allowed. previously only `on_record()` had the condition applied. Co-authored-by: Anton Pirker --- sentry_sdk/integrations/rust_tracing.py | 34 ++++++++----- .../rust_tracing/test_rust_tracing.py | 49 ++++++++++++++----- 2 files changed, 59 insertions(+), 24 deletions(-) diff --git a/sentry_sdk/integrations/rust_tracing.py b/sentry_sdk/integrations/rust_tracing.py index 121bf082b8..ae52c850c3 100644 --- a/sentry_sdk/integrations/rust_tracing.py +++ b/sentry_sdk/integrations/rust_tracing.py @@ -151,12 +151,25 @@ def __init__( [Dict[str, Any]], EventTypeMapping ] = default_event_type_mapping, span_filter: Callable[[Dict[str, Any]], bool] = default_span_filter, - send_sensitive_data: Optional[bool] = None, + include_tracing_fields: Optional[bool] = None, ): self.origin = origin self.event_type_mapping = event_type_mapping self.span_filter = span_filter - self.send_sensitive_data = send_sensitive_data + self.include_tracing_fields = include_tracing_fields + + def _include_tracing_fields(self) -> bool: + """ + By default, the values of tracing fields are not included in case they + contain PII. A user may override that by passing `True` for the + `include_tracing_fields` keyword argument of this integration or by + setting `send_default_pii` to `True` in their Sentry client options. + """ + return ( + should_send_default_pii() + if self.include_tracing_fields is None + else self.include_tracing_fields + ) def on_event(self, event: str, _span_state: TraceState) -> None: deserialized_event = json.loads(event) @@ -207,7 +220,10 @@ def on_new_span(self, attrs: str, span_id: str) -> TraceState: fields = metadata.get("fields", []) for field in fields: - sentry_span.set_data(field, attrs.get(field)) + if self._include_tracing_fields(): + sentry_span.set_data(field, attrs.get(field)) + else: + sentry_span.set_data(field, SENSITIVE_DATA_SUBSTITUTE) scope.span = sentry_span return (parent_sentry_span, sentry_span) @@ -225,15 +241,9 @@ def on_record(self, span_id: str, values: str, span_state: TraceState) -> None: return _parent_sentry_span, sentry_span = span_state - send_sensitive_data = ( - should_send_default_pii() - if self.send_sensitive_data is None - else self.send_sensitive_data - ) - deserialized_values = json.loads(values) for key, value in deserialized_values.items(): - if send_sensitive_data: + if self._include_tracing_fields(): sentry_span.set_data(key, value) else: sentry_span.set_data(key, SENSITIVE_DATA_SUBSTITUTE) @@ -259,12 +269,12 @@ def __init__( [Dict[str, Any]], EventTypeMapping ] = default_event_type_mapping, span_filter: Callable[[Dict[str, Any]], bool] = default_span_filter, - send_sensitive_data: Optional[bool] = None, + include_tracing_fields: Optional[bool] = None, ): self.identifier = identifier origin = f"auto.function.rust_tracing.{identifier}" self.tracing_layer = RustTracingLayer( - origin, event_type_mapping, span_filter, send_sensitive_data + origin, event_type_mapping, span_filter, include_tracing_fields ) initializer(self.tracing_layer) diff --git a/tests/integrations/rust_tracing/test_rust_tracing.py b/tests/integrations/rust_tracing/test_rust_tracing.py index b1fad1a7f7..893fc86966 100644 --- a/tests/integrations/rust_tracing/test_rust_tracing.py +++ b/tests/integrations/rust_tracing/test_rust_tracing.py @@ -1,3 +1,4 @@ +from unittest import mock import pytest from string import Template @@ -66,7 +67,9 @@ def record(self, span_id: int): def test_on_new_span_on_close(sentry_init, capture_events): rust_tracing = FakeRustTracing() integration = RustTracingIntegration( - "test_on_new_span_on_close", rust_tracing.set_layer_impl + "test_on_new_span_on_close", + initializer=rust_tracing.set_layer_impl, + include_tracing_fields=True, ) sentry_init(integrations=[integration], traces_sample_rate=1.0) @@ -105,7 +108,9 @@ def test_on_new_span_on_close(sentry_init, capture_events): def test_nested_on_new_span_on_close(sentry_init, capture_events): rust_tracing = FakeRustTracing() integration = RustTracingIntegration( - "test_nested_on_new_span_on_close", rust_tracing.set_layer_impl + "test_nested_on_new_span_on_close", + initializer=rust_tracing.set_layer_impl, + include_tracing_fields=True, ) sentry_init(integrations=[integration], traces_sample_rate=1.0) @@ -331,7 +336,10 @@ def span_filter(metadata: Dict[str, object]) -> bool: rust_tracing = FakeRustTracing() integration = RustTracingIntegration( - "test_span_filter", rust_tracing.set_layer_impl, span_filter=span_filter + "test_span_filter", + initializer=rust_tracing.set_layer_impl, + span_filter=span_filter, + include_tracing_fields=True, ) sentry_init(integrations=[integration], traces_sample_rate=1.0) @@ -365,7 +373,7 @@ def test_record(sentry_init): integration = RustTracingIntegration( "test_record", initializer=rust_tracing.set_layer_impl, - send_sensitive_data=True, + include_tracing_fields=True, ) sentry_init(integrations=[integration], traces_sample_rate=1.0) @@ -391,6 +399,7 @@ def span_filter(metadata: Dict[str, object]) -> bool: "test_record_in_ignored_span", rust_tracing.set_layer_impl, span_filter=span_filter, + include_tracing_fields=True, ) sentry_init(integrations=[integration], traces_sample_rate=1.0) @@ -409,7 +418,7 @@ def span_filter(metadata: Dict[str, object]) -> bool: @pytest.mark.parametrize( - "send_default_pii, send_sensitive_data, sensitive_data_expected", + "send_default_pii, include_tracing_fields, tracing_fields_expected", [ (True, True, True), (True, False, False), @@ -419,14 +428,14 @@ def span_filter(metadata: Dict[str, object]) -> bool: (False, None, False), ], ) -def test_sensitive_data( - sentry_init, send_default_pii, send_sensitive_data, sensitive_data_expected +def test_include_tracing_fields( + sentry_init, send_default_pii, include_tracing_fields, tracing_fields_expected ): rust_tracing = FakeRustTracing() integration = RustTracingIntegration( "test_record", initializer=rust_tracing.set_layer_impl, - send_sensitive_data=send_sensitive_data, + include_tracing_fields=include_tracing_fields, ) sentry_init( @@ -438,13 +447,29 @@ def test_sensitive_data( rust_tracing.new_span(RustTracingLevel.Info, 3) span_before_record = sentry_sdk.get_current_span().to_json() - assert span_before_record["data"]["version"] is None + if tracing_fields_expected: + assert span_before_record["data"]["version"] is None + else: + assert span_before_record["data"]["version"] == "[Filtered]" rust_tracing.record(3) span_after_record = sentry_sdk.get_current_span().to_json() - if sensitive_data_expected: - assert span_after_record["data"]["version"] == "memoized" + if tracing_fields_expected: + assert span_after_record["data"] == { + "thread.id": mock.ANY, + "thread.name": mock.ANY, + "use_memoized": True, + "version": "memoized", + "index": 10, + } + else: - assert span_after_record["data"]["version"] == "[Filtered]" + assert span_after_record["data"] == { + "thread.id": mock.ANY, + "thread.name": mock.ANY, + "use_memoized": "[Filtered]", + "version": "[Filtered]", + "index": "[Filtered]", + } From a82651928148a9fc1a9b903ecd0cc6e1f6d551d9 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 18 Nov 2024 09:30:01 +0100 Subject: [PATCH 032/591] tests: Test with pyspark prerelease (#3760) --- tox.ini | 2 ++ 1 file changed, 2 insertions(+) diff --git a/tox.ini b/tox.ini index f3a7ba4ea0..6acff6b8e8 100644 --- a/tox.ini +++ b/tox.ini @@ -647,6 +647,8 @@ deps = spark-v3.1: pyspark~=3.1.0 spark-v3.3: pyspark~=3.3.0 spark-v3.5: pyspark~=3.5.0 + # TODO: update to ~=4.0.0 once stable is out + spark-v4.0: pyspark==4.0.0.dev2 spark-latest: pyspark # Starlette From ec2d929e9f2b4cdcbbb13a3685c9d420ce47289b Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 18 Nov 2024 10:00:47 +0100 Subject: [PATCH 033/591] Make sentry-sdk[pure-eval] installable with pip==24.0 (#3757) --- setup.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/setup.py b/setup.py index 7ac4b56fde..29a40c6663 100644 --- a/setup.py +++ b/setup.py @@ -70,7 +70,7 @@ def get_file_text(file_name): "openfeature": ["openfeature-sdk>=0.7.1"], "opentelemetry": ["opentelemetry-distro>=0.35b0"], "opentelemetry-experimental": ["opentelemetry-distro"], - "pure_eval": ["pure_eval", "executing", "asttokens"], + "pure-eval": ["pure_eval", "executing", "asttokens"], "pymongo": ["pymongo>=3.1"], "pyspark": ["pyspark>=2.4.4"], "quart": ["quart>=0.16.1", "blinker>=1.1"], From 955108e5642d74d9d95535c2a1f263fcbbc62c92 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Tue, 19 Nov 2024 08:55:21 +0000 Subject: [PATCH 034/591] feat(spotlight): Auto enable cache_spans for Spotlight on DEBUG (#3791) This patch enables `cache_spans` in Django integration automatically when Spotlight is enabled and `DEBUG` is set in Django settings. --- sentry_sdk/integrations/django/caching.py | 22 ++++++++++++++++------ 1 file changed, 16 insertions(+), 6 deletions(-) diff --git a/sentry_sdk/integrations/django/caching.py b/sentry_sdk/integrations/django/caching.py index 4bd7cb7236..39d1679183 100644 --- a/sentry_sdk/integrations/django/caching.py +++ b/sentry_sdk/integrations/django/caching.py @@ -132,10 +132,22 @@ def _get_address_port(settings): return address, int(port) if port is not None else None -def patch_caching(): - # type: () -> None +def should_enable_cache_spans(): + # type: () -> bool from sentry_sdk.integrations.django import DjangoIntegration + client = sentry_sdk.get_client() + integration = client.get_integration(DjangoIntegration) + from django.conf import settings + + return integration is not None and ( + (client.spotlight is not None and settings.DEBUG is True) + or integration.cache_spans is True + ) + + +def patch_caching(): + # type: () -> None if not hasattr(CacheHandler, "_sentry_patched"): if DJANGO_VERSION < (3, 2): original_get_item = CacheHandler.__getitem__ @@ -145,8 +157,7 @@ def sentry_get_item(self, alias): # type: (CacheHandler, str) -> Any cache = original_get_item(self, alias) - integration = sentry_sdk.get_client().get_integration(DjangoIntegration) - if integration is not None and integration.cache_spans: + if should_enable_cache_spans(): from django.conf import settings address, port = _get_address_port( @@ -168,8 +179,7 @@ def sentry_create_connection(self, alias): # type: (CacheHandler, str) -> Any cache = original_create_connection(self, alias) - integration = sentry_sdk.get_client().get_integration(DjangoIntegration) - if integration is not None and integration.cache_spans: + if should_enable_cache_spans(): address, port = _get_address_port(self.settings[alias or "default"]) _patch_cache(cache, address, port) From 1bd744dbb854508fc287862f4d17cc99501e3150 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 19 Nov 2024 10:49:25 +0100 Subject: [PATCH 035/591] build(deps): bump codecov/codecov-action from 4.6.0 to 5.0.2 (#3792) * build(deps): bump codecov/codecov-action from 4.6.0 to 5.0.2 Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 4.6.0 to 5.0.2. - [Release notes](https://github.com/codecov/codecov-action/releases) - [Changelog](https://github.com/codecov/codecov-action/blob/main/CHANGELOG.md) - [Commits](https://github.com/codecov/codecov-action/compare/v4.6.0...v5.0.2) --- updated-dependencies: - dependency-name: codecov/codecov-action dependency-type: direct:production update-type: version-update:semver-major ... Signed-off-by: dependabot[bot] --------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws-lambda.yml | 2 +- .github/workflows/test-integrations-cloud-computing.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-data-processing.yml | 4 ++-- .github/workflows/test-integrations-databases.yml | 4 ++-- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-miscellaneous.yml | 4 ++-- .github/workflows/test-integrations-networking.yml | 4 ++-- .github/workflows/test-integrations-web-frameworks-1.yml | 4 ++-- .github/workflows/test-integrations-web-frameworks-2.yml | 4 ++-- scripts/split-tox-gh-actions/templates/test_group.jinja | 2 +- 12 files changed, 21 insertions(+), 21 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index dd230a6461..c7cf4a1d85 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -78,7 +78,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -150,7 +150,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws-lambda.yml b/.github/workflows/test-integrations-aws-lambda.yml index c9837c08d0..d85d1d4a8e 100644 --- a/.github/workflows/test-integrations-aws-lambda.yml +++ b/.github/workflows/test-integrations-aws-lambda.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud-computing.yml b/.github/workflows/test-integrations-cloud-computing.yml index 3217811539..9013a02af3 100644 --- a/.github/workflows/test-integrations-cloud-computing.yml +++ b/.github/workflows/test-integrations-cloud-computing.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 912eb3b18c..6983a079ef 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -62,7 +62,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-data-processing.yml b/.github/workflows/test-integrations-data-processing.yml index 128463a66a..6ad3d707fe 100644 --- a/.github/workflows/test-integrations-data-processing.yml +++ b/.github/workflows/test-integrations-data-processing.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-databases.yml b/.github/workflows/test-integrations-databases.yml index 2cdcd9d3b9..045f942b9c 100644 --- a/.github/workflows/test-integrations-databases.yml +++ b/.github/workflows/test-integrations-databases.yml @@ -101,7 +101,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -196,7 +196,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 522dc2acc1..57d14cff10 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-miscellaneous.yml b/.github/workflows/test-integrations-miscellaneous.yml index 03d6559108..ebb486b6b6 100644 --- a/.github/workflows/test-integrations-miscellaneous.yml +++ b/.github/workflows/test-integrations-miscellaneous.yml @@ -86,7 +86,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-networking.yml b/.github/workflows/test-integrations-networking.yml index 31342151e9..2c9a788954 100644 --- a/.github/workflows/test-integrations-networking.yml +++ b/.github/workflows/test-integrations-networking.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-frameworks-1.yml b/.github/workflows/test-integrations-web-frameworks-1.yml index 706feb385f..d4a9aff6f1 100644 --- a/.github/workflows/test-integrations-web-frameworks-1.yml +++ b/.github/workflows/test-integrations-web-frameworks-1.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-frameworks-2.yml b/.github/workflows/test-integrations-web-frameworks-2.yml index f700952e00..f0cdcc4510 100644 --- a/.github/workflows/test-integrations-web-frameworks-2.yml +++ b/.github/workflows/test-integrations-web-frameworks-2.yml @@ -98,7 +98,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -190,7 +190,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split-tox-gh-actions/templates/test_group.jinja index 9055070c72..4560a7d42d 100644 --- a/scripts/split-tox-gh-actions/templates/test_group.jinja +++ b/scripts/split-tox-gh-actions/templates/test_group.jinja @@ -92,7 +92,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v4.6.0 + uses: codecov/codecov-action@v5.0.2 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From d894fc232055ea06ac2ba1431519849e97973423 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 19 Nov 2024 15:29:12 +0100 Subject: [PATCH 036/591] Shorten CI workflow names (#3805) Getting around a GH UI issue where you can't see the whole name of the test that failed --- ...-aws-lambda.yml => test-integrations-aws.yml} | 12 ++++++------ ...computing.yml => test-integrations-cloud.yml} | 16 ++++++++-------- ...s-databases.yml => test-integrations-dbs.yml} | 16 ++++++++-------- ...cellaneous.yml => test-integrations-misc.yml} | 16 ++++++++-------- ...working.yml => test-integrations-network.yml} | 16 ++++++++-------- ...rocessing.yml => test-integrations-tasks.yml} | 16 ++++++++-------- ...meworks-1.yml => test-integrations-web-1.yml} | 16 ++++++++-------- ...meworks-2.yml => test-integrations-web-2.yml} | 16 ++++++++-------- .../split-tox-gh-actions/split-tox-gh-actions.py | 16 ++++++++-------- 9 files changed, 70 insertions(+), 70 deletions(-) rename .github/workflows/{test-integrations-aws-lambda.yml => test-integrations-aws.yml} (94%) rename .github/workflows/{test-integrations-cloud-computing.yml => test-integrations-cloud.yml} (93%) rename .github/workflows/{test-integrations-databases.yml => test-integrations-dbs.yml} (96%) rename .github/workflows/{test-integrations-miscellaneous.yml => test-integrations-misc.yml} (95%) rename .github/workflows/{test-integrations-networking.yml => test-integrations-network.yml} (94%) rename .github/workflows/{test-integrations-data-processing.yml => test-integrations-tasks.yml} (95%) rename .github/workflows/{test-integrations-web-frameworks-1.yml => test-integrations-web-1.yml} (94%) rename .github/workflows/{test-integrations-web-frameworks-2.yml => test-integrations-web-2.yml} (95%) diff --git a/.github/workflows/test-integrations-aws-lambda.yml b/.github/workflows/test-integrations-aws.yml similarity index 94% rename from .github/workflows/test-integrations-aws-lambda.yml rename to .github/workflows/test-integrations-aws.yml index d85d1d4a8e..67c0ec31c7 100644 --- a/.github/workflows/test-integrations-aws-lambda.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test AWS Lambda +name: Test AWS on: push: branches: @@ -52,8 +52,8 @@ jobs: - name: Check permissions on repo branch if: github.event_name == 'push' run: true - test-aws_lambda-pinned: - name: AWS Lambda (pinned) + test-aws-pinned: + name: AWS (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -112,13 +112,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned AWS Lambda tests passed - needs: test-aws_lambda-pinned + name: All pinned AWS tests passed + needs: test-aws-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-aws_lambda-pinned.result, 'failure') || contains(needs.test-aws_lambda-pinned.result, 'skipped') + if: contains(needs.test-aws-pinned.result, 'failure') || contains(needs.test-aws-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-cloud-computing.yml b/.github/workflows/test-integrations-cloud.yml similarity index 93% rename from .github/workflows/test-integrations-cloud-computing.yml rename to .github/workflows/test-integrations-cloud.yml index 9013a02af3..62d67200a5 100644 --- a/.github/workflows/test-integrations-cloud-computing.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Cloud Computing +name: Test Cloud on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-cloud_computing-latest: - name: Cloud Computing (latest) + test-cloud-latest: + name: Cloud (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -88,8 +88,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-cloud_computing-pinned: - name: Cloud Computing (pinned) + test-cloud-pinned: + name: Cloud (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -157,13 +157,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Cloud Computing tests passed - needs: test-cloud_computing-pinned + name: All pinned Cloud tests passed + needs: test-cloud-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-cloud_computing-pinned.result, 'failure') || contains(needs.test-cloud_computing-pinned.result, 'skipped') + if: contains(needs.test-cloud-pinned.result, 'failure') || contains(needs.test-cloud-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-databases.yml b/.github/workflows/test-integrations-dbs.yml similarity index 96% rename from .github/workflows/test-integrations-databases.yml rename to .github/workflows/test-integrations-dbs.yml index 045f942b9c..1612dfb432 100644 --- a/.github/workflows/test-integrations-databases.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Databases +name: Test DBs on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-databases-latest: - name: Databases (latest) + test-dbs-latest: + name: DBs (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -115,8 +115,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-databases-pinned: - name: Databases (pinned) + test-dbs-pinned: + name: DBs (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -211,13 +211,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Databases tests passed - needs: test-databases-pinned + name: All pinned DBs tests passed + needs: test-dbs-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-databases-pinned.result, 'failure') || contains(needs.test-databases-pinned.result, 'skipped') + if: contains(needs.test-dbs-pinned.result, 'failure') || contains(needs.test-dbs-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-miscellaneous.yml b/.github/workflows/test-integrations-misc.yml similarity index 95% rename from .github/workflows/test-integrations-miscellaneous.yml rename to .github/workflows/test-integrations-misc.yml index ebb486b6b6..5f2baa5759 100644 --- a/.github/workflows/test-integrations-miscellaneous.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Miscellaneous +name: Test Misc on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-miscellaneous-latest: - name: Miscellaneous (latest) + test-misc-latest: + name: Misc (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -100,8 +100,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-miscellaneous-pinned: - name: Miscellaneous (pinned) + test-misc-pinned: + name: Misc (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -181,13 +181,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Miscellaneous tests passed - needs: test-miscellaneous-pinned + name: All pinned Misc tests passed + needs: test-misc-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-miscellaneous-pinned.result, 'failure') || contains(needs.test-miscellaneous-pinned.result, 'skipped') + if: contains(needs.test-misc-pinned.result, 'failure') || contains(needs.test-misc-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-networking.yml b/.github/workflows/test-integrations-network.yml similarity index 94% rename from .github/workflows/test-integrations-networking.yml rename to .github/workflows/test-integrations-network.yml index 2c9a788954..7c1c343aac 100644 --- a/.github/workflows/test-integrations-networking.yml +++ b/.github/workflows/test-integrations-network.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Networking +name: Test Network on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-networking-latest: - name: Networking (latest) + test-network-latest: + name: Network (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -88,8 +88,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-networking-pinned: - name: Networking (pinned) + test-network-pinned: + name: Network (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -157,13 +157,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Networking tests passed - needs: test-networking-pinned + name: All pinned Network tests passed + needs: test-network-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-networking-pinned.result, 'failure') || contains(needs.test-networking-pinned.result, 'skipped') + if: contains(needs.test-network-pinned.result, 'failure') || contains(needs.test-network-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-data-processing.yml b/.github/workflows/test-integrations-tasks.yml similarity index 95% rename from .github/workflows/test-integrations-data-processing.yml rename to .github/workflows/test-integrations-tasks.yml index 6ad3d707fe..1c4259ac05 100644 --- a/.github/workflows/test-integrations-data-processing.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Data Processing +name: Test Tasks on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-data_processing-latest: - name: Data Processing (latest) + test-tasks-latest: + name: Tasks (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -106,8 +106,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-data_processing-pinned: - name: Data Processing (pinned) + test-tasks-pinned: + name: Tasks (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -193,13 +193,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Data Processing tests passed - needs: test-data_processing-pinned + name: All pinned Tasks tests passed + needs: test-tasks-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-data_processing-pinned.result, 'failure') || contains(needs.test-data_processing-pinned.result, 'skipped') + if: contains(needs.test-tasks-pinned.result, 'failure') || contains(needs.test-tasks-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-frameworks-1.yml b/.github/workflows/test-integrations-web-1.yml similarity index 94% rename from .github/workflows/test-integrations-web-frameworks-1.yml rename to .github/workflows/test-integrations-web-1.yml index d4a9aff6f1..6a6a01e8ff 100644 --- a/.github/workflows/test-integrations-web-frameworks-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Web Frameworks 1 +name: Test Web 1 on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_frameworks_1-latest: - name: Web Frameworks 1 (latest) + test-web_1-latest: + name: Web 1 (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -106,8 +106,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-web_frameworks_1-pinned: - name: Web Frameworks 1 (pinned) + test-web_1-pinned: + name: Web 1 (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -193,13 +193,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Web Frameworks 1 tests passed - needs: test-web_frameworks_1-pinned + name: All pinned Web 1 tests passed + needs: test-web_1-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-web_frameworks_1-pinned.result, 'failure') || contains(needs.test-web_frameworks_1-pinned.result, 'skipped') + if: contains(needs.test-web_1-pinned.result, 'failure') || contains(needs.test-web_1-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-frameworks-2.yml b/.github/workflows/test-integrations-web-2.yml similarity index 95% rename from .github/workflows/test-integrations-web-frameworks-2.yml rename to .github/workflows/test-integrations-web-2.yml index f0cdcc4510..11cfc20612 100644 --- a/.github/workflows/test-integrations-web-frameworks-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -1,6 +1,6 @@ # Do not edit this file. This file is generated automatically by executing # python scripts/split-tox-gh-actions/split-tox-gh-actions.py -name: Test Web Frameworks 2 +name: Test Web 2 on: push: branches: @@ -20,8 +20,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_frameworks_2-latest: - name: Web Frameworks 2 (latest) + test-web_2-latest: + name: Web 2 (latest) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -112,8 +112,8 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: .junitxml verbose: true - test-web_frameworks_2-pinned: - name: Web Frameworks 2 (pinned) + test-web_2-pinned: + name: Web 2 (pinned) timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -205,13 +205,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Web Frameworks 2 tests passed - needs: test-web_frameworks_2-pinned + name: All pinned Web 2 tests passed + needs: test-web_2-pinned # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-20.04 steps: - name: Check for failures - if: contains(needs.test-web_frameworks_2-pinned.result, 'failure') || contains(needs.test-web_frameworks_2-pinned.result, 'skipped') + if: contains(needs.test-web_2-pinned.result, 'failure') || contains(needs.test-web_2-pinned.result, 'skipped') run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/scripts/split-tox-gh-actions/split-tox-gh-actions.py b/scripts/split-tox-gh-actions/split-tox-gh-actions.py index c0bf2a7a09..c4b8f3e5e5 100755 --- a/scripts/split-tox-gh-actions/split-tox-gh-actions.py +++ b/scripts/split-tox-gh-actions/split-tox-gh-actions.py @@ -65,18 +65,18 @@ "openai", "huggingface_hub", ], - "AWS Lambda": [ + "AWS": [ # this is separate from Cloud Computing because only this one test suite # needs to run with access to GitHub secrets "aws_lambda", ], - "Cloud Computing": [ + "Cloud": [ "boto3", "chalice", "cloud_resource_context", "gcp", ], - "Data Processing": [ + "Tasks": [ "arq", "beam", "celery", @@ -86,7 +86,7 @@ "rq", "spark", ], - "Databases": [ + "DBs": [ "asyncpg", "clickhouse_driver", "pymongo", @@ -100,19 +100,19 @@ "graphene", "strawberry", ], - "Networking": [ + "Network": [ "gevent", "grpc", "httpx", "requests", ], - "Web Frameworks 1": [ + "Web 1": [ "django", "flask", "starlette", "fastapi", ], - "Web Frameworks 2": [ + "Web 2": [ "aiohttp", "asgi", "bottle", @@ -124,7 +124,7 @@ "starlite", "tornado", ], - "Miscellaneous": [ + "Misc": [ "launchdarkly", "loguru", "openfeature", From 01146bd3adeb220bcf6cdd7ca634d2d2bc83b18f Mon Sep 17 00:00:00 2001 From: sourceful-rob <84452928+sourceful-rob@users.noreply.github.com> Date: Tue, 19 Nov 2024 15:27:13 +0000 Subject: [PATCH 037/591] fix(openai): Use name instead of description (#3807) Update the arguments in the start_span function. Specifically, changing the deprecated "description" to "name". This was causing a deprecation warning when running tests. --- sentry_sdk/integrations/openai.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index e6ac36f3cb..61d335b170 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -137,7 +137,7 @@ def _new_chat_completion_common(f, *args, **kwargs): span = sentry_sdk.start_span( op=consts.OP.OPENAI_CHAT_COMPLETIONS_CREATE, - description="Chat Completion", + name="Chat Completion", origin=OpenAIIntegration.origin, ) span.__enter__() From 3e2885322a633398d62e8f1dae6315eefec35a34 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?L=C3=A9o=20Figea?= <59359380+malkovro@users.noreply.github.com> Date: Tue, 19 Nov 2024 16:51:29 +0100 Subject: [PATCH 038/591] fix(integrations): Check retries_left before capturing exception (#3803) Since rq/rq#1964 the job status is set to Failed before the handler decides whether to capture or not the exception while handle_job_failure has not yet been called so the job is not yet re-scheduled leading to all exceptions getting captured in RQ version >= 2.0. Related to #1076 Fixes #3707 --- sentry_sdk/integrations/rq.py | 10 +++++++--- tests/integrations/rq/test_rq.py | 5 ----- 2 files changed, 7 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/integrations/rq.py b/sentry_sdk/integrations/rq.py index c0df1c5e53..462f3ad30a 100644 --- a/sentry_sdk/integrations/rq.py +++ b/sentry_sdk/integrations/rq.py @@ -90,9 +90,13 @@ def sentry_patched_perform_job(self, job, *args, **kwargs): def sentry_patched_handle_exception(self, job, *exc_info, **kwargs): # type: (Worker, Any, *Any, **Any) -> Any - # Note, the order of the `or` here is important, - # because calling `job.is_failed` will change `_status`. - if job._status == JobStatus.FAILED or job.is_failed: + retry = ( + hasattr(job, "retries_left") + and job.retries_left + and job.retries_left > 0 + ) + failed = job._status == JobStatus.FAILED or job.is_failed + if failed and not retry: _capture_exception(exc_info) return old_handle_exception(self, job, *exc_info, **kwargs) diff --git a/tests/integrations/rq/test_rq.py b/tests/integrations/rq/test_rq.py index ffd6f458e1..e445b588be 100644 --- a/tests/integrations/rq/test_rq.py +++ b/tests/integrations/rq/test_rq.py @@ -254,11 +254,6 @@ def test_traces_sampler_gets_correct_values_in_sampling_context( @pytest.mark.skipif( parse_version(rq.__version__) < (1, 5), reason="At least rq-1.5 required" ) -@pytest.mark.skipif( - parse_version(rq.__version__) >= (2,), - reason="Test broke in RQ 2.0. Investigate and fix. " - "See https://github.com/getsentry/sentry-python/issues/3707.", -) def test_job_with_retries(sentry_init, capture_events): sentry_init(integrations=[RqIntegration()]) events = capture_events() From aa6e8fd05ca5812213c96cdaf125ab3ae23726f8 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 21 Nov 2024 11:32:32 +0100 Subject: [PATCH 039/591] fix(falcon): Don't exhaust request body stream (#3768) Only read the cached `request._media`, since reading `request.media` will exhaust the `request.bounded_stream` if it has not been read before. Note that this means that we will now only send the JSON request body to Sentry if the Falcon request handler reads the JSON data. Fixes #3761 Co-authored-by: Anton Pirker --- sentry_sdk/integrations/falcon.py | 44 ++++++++++++----------- tests/integrations/falcon/test_falcon.py | 45 ++++++++++++++++++++++++ 2 files changed, 68 insertions(+), 21 deletions(-) diff --git a/sentry_sdk/integrations/falcon.py b/sentry_sdk/integrations/falcon.py index 00ac106e15..ce771d16e7 100644 --- a/sentry_sdk/integrations/falcon.py +++ b/sentry_sdk/integrations/falcon.py @@ -43,6 +43,12 @@ FALCON3 = False +_FALCON_UNSET = None # type: Optional[object] +if FALCON3: # falcon.request._UNSET is only available in Falcon 3.0+ + with capture_internal_exceptions(): + from falcon.request import _UNSET as _FALCON_UNSET # type: ignore[import-not-found, no-redef] + + class FalconRequestExtractor(RequestExtractor): def env(self): # type: () -> Dict[str, Any] @@ -73,27 +79,23 @@ def raw_data(self): else: return None - if FALCON3: - - def json(self): - # type: () -> Optional[Dict[str, Any]] - try: - return self.request.media - except falcon.errors.HTTPBadRequest: - return None - - else: - - def json(self): - # type: () -> Optional[Dict[str, Any]] - try: - return self.request.media - except falcon.errors.HTTPBadRequest: - # NOTE(jmagnusson): We return `falcon.Request._media` here because - # falcon 1.4 doesn't do proper type checking in - # `falcon.Request.media`. This has been fixed in 2.0. - # Relevant code: https://github.com/falconry/falcon/blob/1.4.1/falcon/request.py#L953 - return self.request._media + def json(self): + # type: () -> Optional[Dict[str, Any]] + # fallback to cached_media = None if self.request._media is not available + cached_media = None + with capture_internal_exceptions(): + # self.request._media is the cached self.request.media + # value. It is only available if self.request.media + # has already been accessed. Therefore, reading + # self.request._media will not exhaust the raw request + # stream (self.request.bounded_stream) because it has + # already been read if self.request._media is set. + cached_media = self.request._media + + if cached_media is not _FALCON_UNSET: + return cached_media + + return None class SentryFalconMiddleware: diff --git a/tests/integrations/falcon/test_falcon.py b/tests/integrations/falcon/test_falcon.py index 0607d3fdeb..51a1d94334 100644 --- a/tests/integrations/falcon/test_falcon.py +++ b/tests/integrations/falcon/test_falcon.py @@ -460,3 +460,48 @@ def test_span_origin(sentry_init, capture_events, make_client): (_, event) = events assert event["contexts"]["trace"]["origin"] == "auto.http.falcon" + + +def test_falcon_request_media(sentry_init): + # test_passed stores whether the test has passed. + test_passed = False + + # test_failure_reason stores the reason why the test failed + # if test_passed is False. The value is meaningless when + # test_passed is True. + test_failure_reason = "test endpoint did not get called" + + class SentryCaptureMiddleware: + def process_request(self, _req, _resp): + # This capture message forces Falcon event processors to run + # before the request handler runs + sentry_sdk.capture_message("Processing request") + + class RequestMediaResource: + def on_post(self, req, _): + nonlocal test_passed, test_failure_reason + raw_data = req.bounded_stream.read() + + # If the raw_data is empty, the request body stream + # has been exhausted by the SDK. Test should fail in + # this case. + test_passed = raw_data != b"" + test_failure_reason = "request body has been read" + + sentry_init(integrations=[FalconIntegration()]) + + try: + app_class = falcon.App # Falcon ≥3.0 + except AttributeError: + app_class = falcon.API # Falcon <3.0 + + app = app_class(middleware=[SentryCaptureMiddleware()]) + app.add_route("/read_body", RequestMediaResource()) + + client = falcon.testing.TestClient(app) + + client.simulate_post("/read_body", json={"foo": "bar"}) + + # Check that simulate_post actually calls the resource, and + # that the SDK does not exhaust the request body stream. + assert test_passed, test_failure_reason From e9ec6c1812b3c4c0bebdfb736869c1f6a226dc71 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 21 Nov 2024 12:46:47 +0100 Subject: [PATCH 040/591] test(gcp): Only run GCP tests when they should (#3721) GCP tests have been running in our common test suite, including on Python versions other than 3.7 (the only version which supports the GCP integration), even though we have a separate `py3.7-gcp` tox environment for these tests. The tests take a long time, so only executing in the appropriate `tox` environment should speed up CI time. Co-authored-by: Anton Pirker --- tests/integrations/gcp/__init__.py | 6 ++++++ 1 file changed, 6 insertions(+) create mode 100644 tests/integrations/gcp/__init__.py diff --git a/tests/integrations/gcp/__init__.py b/tests/integrations/gcp/__init__.py new file mode 100644 index 0000000000..eaf1ba89bb --- /dev/null +++ b/tests/integrations/gcp/__init__.py @@ -0,0 +1,6 @@ +import pytest +import os + + +if "gcp" not in os.environ.get("TOX_ENV_NAME", ""): + pytest.skip("GCP tests only run in GCP environment", allow_module_level=True) From bd50c386527f0d014e2e3c5dea274f6836e713e6 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 21 Nov 2024 13:00:16 +0100 Subject: [PATCH 041/591] fix(httpx): Prevent Sentry baggage duplication (#3728) Sentry baggage will get added to an HTTPX request multiple times if the same request is repeated. To prevent this from occurring, we can strip any existing Sentry baggage before adding Sentry baggage to the request. Fixes #3709 --------- Co-authored-by: Ivana Kellyer Co-authored-by: Anton Pirker --- sentry_sdk/integrations/httpx.py | 29 +++++++++++++++++++++++------ sentry_sdk/tracing_utils.py | 15 +++++++++++++++ tests/test_tracing_utils.py | 23 ++++++++++++++++++++++- 3 files changed, 60 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/httpx.py b/sentry_sdk/integrations/httpx.py index 6f80b93f4d..2ddd44489f 100644 --- a/sentry_sdk/integrations/httpx.py +++ b/sentry_sdk/integrations/httpx.py @@ -2,7 +2,7 @@ from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.tracing import BAGGAGE_HEADER_NAME -from sentry_sdk.tracing_utils import should_propagate_trace +from sentry_sdk.tracing_utils import Baggage, should_propagate_trace from sentry_sdk.utils import ( SENSITIVE_DATA_SUBSTITUTE, capture_internal_exceptions, @@ -14,6 +14,7 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: + from collections.abc import MutableMapping from typing import Any @@ -76,11 +77,9 @@ def send(self, request, **kwargs): key=key, value=value, url=request.url ) ) - if key == BAGGAGE_HEADER_NAME and request.headers.get( - BAGGAGE_HEADER_NAME - ): - # do not overwrite any existing baggage, just append to it - request.headers[key] += "," + value + + if key == BAGGAGE_HEADER_NAME: + _add_sentry_baggage_to_headers(request.headers, value) else: request.headers[key] = value @@ -148,3 +147,21 @@ async def send(self, request, **kwargs): return rv AsyncClient.send = send + + +def _add_sentry_baggage_to_headers(headers, sentry_baggage): + # type: (MutableMapping[str, str], str) -> None + """Add the Sentry baggage to the headers. + + This function directly mutates the provided headers. The provided sentry_baggage + is appended to the existing baggage. If the baggage already contains Sentry items, + they are stripped out first. + """ + existing_baggage = headers.get(BAGGAGE_HEADER_NAME, "") + stripped_existing_baggage = Baggage.strip_sentry_baggage(existing_baggage) + + separator = "," if len(stripped_existing_baggage) > 0 else "" + + headers[BAGGAGE_HEADER_NAME] = ( + stripped_existing_baggage + separator + sentry_baggage + ) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 150e73661e..0459563776 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -609,6 +609,21 @@ def serialize(self, include_third_party=False): return ",".join(items) + @staticmethod + def strip_sentry_baggage(header): + # type: (str) -> str + """Remove Sentry baggage from the given header. + + Given a Baggage header, return a new Baggage header with all Sentry baggage items removed. + """ + return ",".join( + ( + item + for item in header.split(",") + if not Baggage.SENTRY_PREFIX_REGEX.match(item.strip()) + ) + ) + def should_propagate_trace(client, url): # type: (sentry_sdk.client.BaseClient, str) -> bool diff --git a/tests/test_tracing_utils.py b/tests/test_tracing_utils.py index 239e631156..5c1f70516d 100644 --- a/tests/test_tracing_utils.py +++ b/tests/test_tracing_utils.py @@ -1,7 +1,7 @@ from dataclasses import asdict, dataclass from typing import Optional, List -from sentry_sdk.tracing_utils import _should_be_included +from sentry_sdk.tracing_utils import _should_be_included, Baggage import pytest @@ -94,3 +94,24 @@ def test_should_be_included(test_case, expected): kwargs = asdict(test_case) kwargs.pop("id") assert _should_be_included(**kwargs) == expected + + +@pytest.mark.parametrize( + ("header", "expected"), + ( + ("", ""), + ("foo=bar", "foo=bar"), + (" foo=bar, baz = qux ", " foo=bar, baz = qux "), + ("sentry-trace_id=123", ""), + (" sentry-trace_id = 123 ", ""), + ("sentry-trace_id=123,sentry-public_key=456", ""), + ("foo=bar,sentry-trace_id=123", "foo=bar"), + ("foo=bar,sentry-trace_id=123,baz=qux", "foo=bar,baz=qux"), + ( + "foo=bar,sentry-trace_id=123,baz=qux,sentry-public_key=456", + "foo=bar,baz=qux", + ), + ), +) +def test_strip_sentry_baggage(header, expected): + assert Baggage.strip_sentry_baggage(header) == expected From 295dd8d50fc161c79db7249d228f87d79bb5bd38 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Janek=20Nouvertn=C3=A9?= Date: Thu, 21 Nov 2024 13:02:49 +0100 Subject: [PATCH 042/591] Auto enable Litestar integration (#3540) Auto enable the Litestar integration added in #3358. --------- Co-authored-by: Ivana Kellyer Co-authored-by: Anton Pirker --- sentry_sdk/integrations/__init__.py | 1 + 1 file changed, 1 insertion(+) diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 32528246af..12336a939b 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -95,6 +95,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "sentry_sdk.integrations.huey.HueyIntegration", "sentry_sdk.integrations.huggingface_hub.HuggingfaceHubIntegration", "sentry_sdk.integrations.langchain.LangchainIntegration", + "sentry_sdk.integrations.litestar.LitestarIntegration", "sentry_sdk.integrations.loguru.LoguruIntegration", "sentry_sdk.integrations.openai.OpenAIIntegration", "sentry_sdk.integrations.pymongo.PyMongoIntegration", From 8fe5bb4b1946874f61bfc09dcce327e20bb24519 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Thu, 21 Nov 2024 15:20:56 +0000 Subject: [PATCH 043/591] feat: Send PII to Spotlight when no DSN is set (#3804) * feat: Send PII to Spotlight when no DSN is set Quick fix for getsentry/spotlight#543 until we implement a global scrubber that only scrubs events sent to the clound thorugh the DSN. * add tests fix bugs * Make scrubber initialization more explicit * Refactored to not change the default value of send_default_pii * Add test to show that there is now no way to opt out of sending PII to spotlight. * Revert "Refactored to not change the default value of send_default_pii" This reverts commit 15cf625859852b0a51c70f8126ad92af6d947d48. * Revert "Add test to show that there is now no way to opt out of sending PII to spotlight." This reverts commit de7f39818af78a1012a8fcea6bbd80f20c6b0eb3. --------- Co-authored-by: Anton Pirker --- sentry_sdk/client.py | 12 ++++++++++-- sentry_sdk/consts.py | 3 ++- tests/test_scope.py | 18 ++++++++++++++++++ 3 files changed, 30 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index b1e7868031..db2cc19110 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -128,7 +128,11 @@ def _get_options(*args, **kwargs): rv["traces_sample_rate"] = 1.0 if rv["event_scrubber"] is None: - rv["event_scrubber"] = EventScrubber(send_default_pii=rv["send_default_pii"]) + rv["event_scrubber"] = EventScrubber( + send_default_pii=( + False if rv["send_default_pii"] is None else rv["send_default_pii"] + ) + ) if rv["socket_options"] and not isinstance(rv["socket_options"], list): logger.warning( @@ -451,7 +455,11 @@ def should_send_default_pii(self): Returns whether the client should send default PII (Personally Identifiable Information) data to Sentry. """ - return self.options.get("send_default_pii", False) + result = self.options.get("send_default_pii") + if result is None: + result = not self.options["dsn"] and self.spotlight is not None + + return result @property def dsn(self): diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index ae32294d05..bb2a73337e 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -489,6 +489,7 @@ class OP: # This type exists to trick mypy and PyCharm into thinking `init` and `Client` # take these arguments (even though they take opaque **kwargs) class ClientConstructor: + def __init__( self, dsn=None, # type: Optional[str] @@ -506,7 +507,7 @@ def __init__( transport=None, # type: Optional[Union[sentry_sdk.transport.Transport, Type[sentry_sdk.transport.Transport], Callable[[Event], None]]] transport_queue_size=DEFAULT_QUEUE_SIZE, # type: int sample_rate=1.0, # type: float - send_default_pii=False, # type: bool + send_default_pii=None, # type: Optional[bool] http_proxy=None, # type: Optional[str] https_proxy=None, # type: Optional[str] ignore_errors=[], # type: Sequence[Union[type, str]] # noqa: B006 diff --git a/tests/test_scope.py b/tests/test_scope.py index 0dfa155d11..374a354446 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -811,6 +811,24 @@ def test_should_send_default_pii_false(sentry_init): assert should_send_default_pii() is False +def test_should_send_default_pii_default_false(sentry_init): + sentry_init() + + assert should_send_default_pii() is False + + +def test_should_send_default_pii_false_with_dsn_and_spotlight(sentry_init): + sentry_init(dsn="http://key@localhost/1", spotlight=True) + + assert should_send_default_pii() is False + + +def test_should_send_default_pii_true_without_dsn_and_spotlight(sentry_init): + sentry_init(spotlight=True) + + assert should_send_default_pii() is True + + def test_set_tags(): scope = Scope() scope.set_tags({"tag1": "value1", "tag2": "value2"}) From c83e7428f44263e6d62ab88cb61034e7f438b2b4 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 21 Nov 2024 15:22:15 +0000 Subject: [PATCH 044/591] release: 2.19.0 --- CHANGELOG.md | 26 ++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 29 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index c47d0e0458..dab245e15a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,31 @@ # Changelog +## 2.19.0 + +### Various fixes & improvements + +- feat: Send PII to Spotlight when no DSN is set (#3804) by @BYK +- Auto enable Litestar integration (#3540) by @provinzkraut +- fix(httpx): Prevent Sentry baggage duplication (#3728) by @szokeasaurusrex +- test(gcp): Only run GCP tests when they should (#3721) by @szokeasaurusrex +- fix(falcon): Don't exhaust request body stream (#3768) by @szokeasaurusrex +- fix(integrations): Check retries_left before capturing exception (#3803) by @malkovro +- fix(openai): Use name instead of description (#3807) by @sourceful-rob +- Shorten CI workflow names (#3805) by @sentrivana +- build(deps): bump codecov/codecov-action from 4.6.0 to 5.0.2 (#3792) by @dependabot +- feat(spotlight): Auto enable cache_spans for Spotlight on DEBUG (#3791) by @BYK +- Make sentry-sdk[pure-eval] installable with pip==24.0 (#3757) by @sentrivana +- tests: Test with pyspark prerelease (#3760) by @sentrivana +- fix: include_tracing_fields arg to control unvetted data in rust_tracing integration (#3780) by @matt-codecov +- feat: introduce rust_tracing integration (#3717) by @matt-codecov +- Fix aws lambda tests (by reducing event size) (#3770) by @antonpirker +- feat(spotlight): Inject Spotlight button on Django (#3751) by @BYK +- ref(init): Deprecate `sentry_sdk.init` context manager (#3729) by @szokeasaurusrex +- Handle parameter `stack_info` for the `LoggingIntegration` (#3745) by @gmcrocetti +- Fix(Arq): fix integration with Worker settings as a dict (#3742) by @saber-solooki +- feat(spotlight): Add info logs when Sentry is enabled (#3735) by @BYK +- build(deps): bump actions/checkout from 4.2.1 to 4.2.2 (#3691) by @dependabot + ## 2.18.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 6d33e5809a..55d5295381 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.18.0" +release = "2.19.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index bb2a73337e..488743b579 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -576,4 +576,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.18.0" +VERSION = "2.19.0" diff --git a/setup.py b/setup.py index 29a40c6663..fda3daa229 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.18.0", + version="2.19.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 039c220bcb5208b278bc1cd0b08611bdac26b895 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 21 Nov 2024 16:31:18 +0100 Subject: [PATCH 045/591] Updated changelog --- CHANGELOG.md | 30 +++++++++++++++--------------- 1 file changed, 15 insertions(+), 15 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index dab245e15a..dbb35eb1eb 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,26 +4,26 @@ ### Various fixes & improvements -- feat: Send PII to Spotlight when no DSN is set (#3804) by @BYK +- New: introduce `rust_tracing` integration. See https://docs.sentry.io/platforms/python/integrations/rust_tracing/ (#3717) by @matt-codecov - Auto enable Litestar integration (#3540) by @provinzkraut +- Deprecate `sentry_sdk.init` context manager (#3729) by @szokeasaurusrex +- feat(spotlight): Send PII to Spotlight when no DSN is set (#3804) by @BYK +- feat(spotlight): Add info logs when Sentry is enabled (#3735) by @BYK +- feat(spotlight): Inject Spotlight button on Django (#3751) by @BYK +- feat(spotlight): Auto enable cache_spans for Spotlight on DEBUG (#3791) by @BYK +- fix(logging): Handle parameter `stack_info` for the `LoggingIntegration` (#3745) by @gmcrocetti +- fix(pure-eval): Make sentry-sdk[pure-eval] installable with pip==24.0 (#3757) by @sentrivana +- fix(rust_tracing): include_tracing_fields arg to control unvetted data in rust_tracing integration (#3780) by @matt-codecov +- fix(aws) Fix aws lambda tests (by reducing event size) (#3770) by @antonpirker +- fix(arq): fix integration with Worker settings as a dict (#3742) by @saber-solooki - fix(httpx): Prevent Sentry baggage duplication (#3728) by @szokeasaurusrex -- test(gcp): Only run GCP tests when they should (#3721) by @szokeasaurusrex - fix(falcon): Don't exhaust request body stream (#3768) by @szokeasaurusrex -- fix(integrations): Check retries_left before capturing exception (#3803) by @malkovro +- fix(integrations): Check `retries_left` before capturing exception (#3803) by @malkovro - fix(openai): Use name instead of description (#3807) by @sourceful-rob -- Shorten CI workflow names (#3805) by @sentrivana +- test(gcp): Only run GCP tests when they should (#3721) by @szokeasaurusrex +- chore: Shorten CI workflow names (#3805) by @sentrivana +- chore: Test with pyspark prerelease (#3760) by @sentrivana - build(deps): bump codecov/codecov-action from 4.6.0 to 5.0.2 (#3792) by @dependabot -- feat(spotlight): Auto enable cache_spans for Spotlight on DEBUG (#3791) by @BYK -- Make sentry-sdk[pure-eval] installable with pip==24.0 (#3757) by @sentrivana -- tests: Test with pyspark prerelease (#3760) by @sentrivana -- fix: include_tracing_fields arg to control unvetted data in rust_tracing integration (#3780) by @matt-codecov -- feat: introduce rust_tracing integration (#3717) by @matt-codecov -- Fix aws lambda tests (by reducing event size) (#3770) by @antonpirker -- feat(spotlight): Inject Spotlight button on Django (#3751) by @BYK -- ref(init): Deprecate `sentry_sdk.init` context manager (#3729) by @szokeasaurusrex -- Handle parameter `stack_info` for the `LoggingIntegration` (#3745) by @gmcrocetti -- Fix(Arq): fix integration with Worker settings as a dict (#3742) by @saber-solooki -- feat(spotlight): Add info logs when Sentry is enabled (#3735) by @BYK - build(deps): bump actions/checkout from 4.2.1 to 4.2.2 (#3691) by @dependabot ## 2.18.0 From da206237473aeb38d911d9cd86f40bd928a2a350 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 25 Nov 2024 10:04:43 +0100 Subject: [PATCH 046/591] Fix spans for streaming responses in WSGI based frameworks (#3798) Fixes spans in streaming responses when using WSGI based frameworks. Only close the transaction once the response was consumed. This way all the spans created during creation of the response will be recorded with the transaction: - The transaction stays open until all the streaming blocks are sent to the client. (because of this I had to update the tests, to make sure the tests, consume the response, because the Werkzeug test client (used by Flask and Django and our Strawberry tests) will not close the WSGI response) - A maximum runtime of 5 minutes for transactions is enforced. (like Javascript does it) - When using a generator to generate the streaming response, it uses the correct scopes to have correct parent-child relationship of spans created in the generator. People having Sentry in a streaming application will: - See an increase in their transaction duration to up to 5 minutes - Get the correct span tree for streaming responses generated by a generator Fixes #3736 --- sentry_sdk/integrations/wsgi.py | 135 ++++++++++++------ sentry_sdk/tracing_utils.py | 18 +++ tests/integrations/django/test_basic.py | 46 +++--- tests/integrations/flask/test_flask.py | 22 ++- .../strawberry/test_strawberry.py | 43 ++++-- tests/integrations/wsgi/test_wsgi.py | 79 ++++++++++ 6 files changed, 270 insertions(+), 73 deletions(-) diff --git a/sentry_sdk/integrations/wsgi.py b/sentry_sdk/integrations/wsgi.py index 50deae10c5..751735f462 100644 --- a/sentry_sdk/integrations/wsgi.py +++ b/sentry_sdk/integrations/wsgi.py @@ -1,19 +1,19 @@ import sys from functools import partial +from threading import Timer import sentry_sdk from sentry_sdk._werkzeug import get_host, _get_headers from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP -from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.scope import should_send_default_pii, use_isolation_scope, use_scope from sentry_sdk.integrations._wsgi_common import ( DEFAULT_HTTP_METHODS_TO_CAPTURE, _filter_headers, - nullcontext, ) from sentry_sdk.sessions import track_session -from sentry_sdk.scope import use_isolation_scope from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing_utils import finish_running_transaction from sentry_sdk.utils import ( ContextVar, capture_internal_exceptions, @@ -46,6 +46,9 @@ def __call__(self, status, response_headers, exc_info=None): # type: ignore pass +MAX_TRANSACTION_DURATION_SECONDS = 5 * 60 + + _wsgi_middleware_applied = ContextVar("sentry_wsgi_middleware_applied") @@ -98,6 +101,7 @@ def __call__(self, environ, start_response): _wsgi_middleware_applied.set(True) try: with sentry_sdk.isolation_scope() as scope: + current_scope = sentry_sdk.get_current_scope() with track_session(scope, session_mode="request"): with capture_internal_exceptions(): scope.clear_breadcrumbs() @@ -109,6 +113,7 @@ def __call__(self, environ, start_response): ) method = environ.get("REQUEST_METHOD", "").upper() + transaction = None if method in self.http_methods_to_capture: transaction = continue_trace( @@ -119,27 +124,43 @@ def __call__(self, environ, start_response): origin=self.span_origin, ) - with ( + timer = None + if transaction is not None: sentry_sdk.start_transaction( transaction, custom_sampling_context={"wsgi_environ": environ}, + ).__enter__() + timer = Timer( + MAX_TRANSACTION_DURATION_SECONDS, + _finish_long_running_transaction, + args=(current_scope, scope), ) - if transaction is not None - else nullcontext() - ): - try: - response = self.app( - environ, - partial( - _sentry_start_response, start_response, transaction - ), - ) - except BaseException: - reraise(*_capture_exception()) + timer.start() + + try: + response = self.app( + environ, + partial( + _sentry_start_response, + start_response, + transaction, + ), + ) + except BaseException: + exc_info = sys.exc_info() + _capture_exception(exc_info) + finish_running_transaction(current_scope, exc_info, timer) + reraise(*exc_info) + finally: _wsgi_middleware_applied.set(False) - return _ScopedResponse(scope, response) + return _ScopedResponse( + response=response, + current_scope=current_scope, + isolation_scope=scope, + timer=timer, + ) def _sentry_start_response( # type: ignore @@ -201,13 +222,13 @@ def get_client_ip(environ): return environ.get("REMOTE_ADDR") -def _capture_exception(): - # type: () -> ExcInfo +def _capture_exception(exc_info=None): + # type: (Optional[ExcInfo]) -> ExcInfo """ Captures the current exception and sends it to Sentry. Returns the ExcInfo tuple to it can be reraised afterwards. """ - exc_info = sys.exc_info() + exc_info = exc_info or sys.exc_info() e = exc_info[1] # SystemExit(0) is the only uncaught exception that is expected behavior @@ -225,7 +246,7 @@ def _capture_exception(): class _ScopedResponse: """ - Users a separate scope for each response chunk. + Use separate scopes for each response chunk. This will make WSGI apps more tolerant against: - WSGI servers streaming responses from a different thread/from @@ -234,37 +255,54 @@ class _ScopedResponse: - WSGI servers streaming responses interleaved from the same thread """ - __slots__ = ("_response", "_scope") + __slots__ = ("_response", "_current_scope", "_isolation_scope", "_timer") - def __init__(self, scope, response): - # type: (sentry_sdk.scope.Scope, Iterator[bytes]) -> None - self._scope = scope + def __init__( + self, + response, # type: Iterator[bytes] + current_scope, # type: sentry_sdk.scope.Scope + isolation_scope, # type: sentry_sdk.scope.Scope + timer=None, # type: Optional[Timer] + ): + # type: (...) -> None self._response = response + self._current_scope = current_scope + self._isolation_scope = isolation_scope + self._timer = timer def __iter__(self): # type: () -> Iterator[bytes] iterator = iter(self._response) - while True: - with use_isolation_scope(self._scope): - try: - chunk = next(iterator) - except StopIteration: - break - except BaseException: - reraise(*_capture_exception()) + try: + while True: + with use_isolation_scope(self._isolation_scope): + with use_scope(self._current_scope): + try: + chunk = next(iterator) + except StopIteration: + break + except BaseException: + reraise(*_capture_exception()) + + yield chunk - yield chunk + finally: + with use_isolation_scope(self._isolation_scope): + with use_scope(self._current_scope): + finish_running_transaction(timer=self._timer) def close(self): # type: () -> None - with use_isolation_scope(self._scope): - try: - self._response.close() # type: ignore - except AttributeError: - pass - except BaseException: - reraise(*_capture_exception()) + with use_isolation_scope(self._isolation_scope): + with use_scope(self._current_scope): + try: + finish_running_transaction(timer=self._timer) + self._response.close() # type: ignore + except AttributeError: + pass + except BaseException: + reraise(*_capture_exception()) def _make_wsgi_event_processor(environ, use_x_forwarded_for): @@ -308,3 +346,18 @@ def event_processor(event, hint): return event return event_processor + + +def _finish_long_running_transaction(current_scope, isolation_scope): + # type: (sentry_sdk.scope.Scope, sentry_sdk.scope.Scope) -> None + """ + Make sure we don't keep transactions open for too long. + Triggered after MAX_TRANSACTION_DURATION_SECONDS have passed. + """ + try: + with use_isolation_scope(isolation_scope): + with use_scope(current_scope): + finish_running_transaction() + except AttributeError: + # transaction is not there anymore + pass diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 0459563776..969e0812e4 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -36,6 +36,9 @@ from types import FrameType + from sentry_sdk._types import ExcInfo + from threading import Timer + SENTRY_TRACE_REGEX = re.compile( "^[ \t]*" # whitespace @@ -739,3 +742,18 @@ def get_current_span(scope=None): if TYPE_CHECKING: from sentry_sdk.tracing import Span + + +def finish_running_transaction(scope=None, exc_info=None, timer=None): + # type: (Optional[sentry_sdk.Scope], Optional[ExcInfo], Optional[Timer]) -> None + if timer is not None: + timer.cancel() + + current_scope = scope or sentry_sdk.get_current_scope() + if current_scope.transaction is not None and hasattr( + current_scope.transaction, "_context_manager_state" + ): + if exc_info is not None: + current_scope.transaction.__exit__(*exc_info) + else: + current_scope.transaction.__exit__(None, None, None) diff --git a/tests/integrations/django/test_basic.py b/tests/integrations/django/test_basic.py index 0e3f700105..243431fdf5 100644 --- a/tests/integrations/django/test_basic.py +++ b/tests/integrations/django/test_basic.py @@ -51,7 +51,7 @@ def test_view_exceptions(sentry_init, client, capture_exceptions, capture_events sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - client.get(reverse("view_exc")) + unpack_werkzeug_response(client.get(reverse("view_exc"))) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -72,7 +72,9 @@ def test_ensures_x_forwarded_header_is_honored_in_sdk_when_enabled_in_django( sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) + unpack_werkzeug_response( + client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) + ) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -91,7 +93,9 @@ def test_ensures_x_forwarded_header_is_not_honored_when_unenabled_in_django( sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) + unpack_werkzeug_response( + client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) + ) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -103,7 +107,7 @@ def test_ensures_x_forwarded_header_is_not_honored_when_unenabled_in_django( def test_middleware_exceptions(sentry_init, client, capture_exceptions): sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() - client.get(reverse("middleware_exc")) + unpack_werkzeug_response(client.get(reverse("middleware_exc"))) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -157,7 +161,7 @@ def test_has_trace_if_performance_enabled(sentry_init, client, capture_events): traces_sample_rate=1.0, ) events = capture_events() - client.head(reverse("view_exc_with_msg")) + unpack_werkzeug_response(client.head(reverse("view_exc_with_msg"))) (msg_event, error_event, transaction_event) = events @@ -213,8 +217,10 @@ def test_trace_from_headers_if_performance_enabled(sentry_init, client, capture_ trace_id = "582b43a4192642f0b136d5159a501701" sentry_trace_header = "{}-{}-{}".format(trace_id, "6e8f22c393e68f19", 1) - client.head( - reverse("view_exc_with_msg"), headers={"sentry-trace": sentry_trace_header} + unpack_werkzeug_response( + client.head( + reverse("view_exc_with_msg"), headers={"sentry-trace": sentry_trace_header} + ) ) (msg_event, error_event, transaction_event) = events @@ -928,7 +934,7 @@ def test_render_spans(sentry_init, client, capture_events, render_span_tree): for url, expected_line in views_tests: events = capture_events() - client.get(url) + unpack_werkzeug_response(client.get(url)) transaction = events[0] assert expected_line in render_span_tree(transaction) @@ -967,7 +973,7 @@ def test_middleware_spans(sentry_init, client, capture_events, render_span_tree) ) events = capture_events() - client.get(reverse("message")) + unpack_werkzeug_response(client.get(reverse("message"))) message, transaction = events @@ -984,7 +990,7 @@ def test_middleware_spans_disabled(sentry_init, client, capture_events): ) events = capture_events() - client.get(reverse("message")) + unpack_werkzeug_response(client.get(reverse("message"))) message, transaction = events @@ -1008,7 +1014,7 @@ def test_signals_spans(sentry_init, client, capture_events, render_span_tree): ) events = capture_events() - client.get(reverse("message")) + unpack_werkzeug_response(client.get(reverse("message"))) message, transaction = events @@ -1031,7 +1037,7 @@ def test_signals_spans_disabled(sentry_init, client, capture_events): ) events = capture_events() - client.get(reverse("message")) + unpack_werkzeug_response(client.get(reverse("message"))) message, transaction = events @@ -1061,7 +1067,7 @@ def test_signals_spans_filtering(sentry_init, client, capture_events, render_spa ) events = capture_events() - client.get(reverse("send_myapp_custom_signal")) + unpack_werkzeug_response(client.get(reverse("send_myapp_custom_signal"))) (transaction,) = events @@ -1186,7 +1192,7 @@ def test_span_origin(sentry_init, client, capture_events): ) events = capture_events() - client.get(reverse("view_with_signal")) + unpack_werkzeug_response(client.get(reverse("view_with_signal"))) (transaction,) = events @@ -1211,9 +1217,9 @@ def test_transaction_http_method_default(sentry_init, client, capture_events): ) events = capture_events() - client.get("/nomessage") - client.options("/nomessage") - client.head("/nomessage") + unpack_werkzeug_response(client.get("/nomessage")) + unpack_werkzeug_response(client.options("/nomessage")) + unpack_werkzeug_response(client.head("/nomessage")) (event,) = events @@ -1235,9 +1241,9 @@ def test_transaction_http_method_custom(sentry_init, client, capture_events): ) events = capture_events() - client.get("/nomessage") - client.options("/nomessage") - client.head("/nomessage") + unpack_werkzeug_response(client.get("/nomessage")) + unpack_werkzeug_response(client.options("/nomessage")) + unpack_werkzeug_response(client.head("/nomessage")) assert len(events) == 2 diff --git a/tests/integrations/flask/test_flask.py b/tests/integrations/flask/test_flask.py index 6febb12b8b..e2c37aa5f7 100644 --- a/tests/integrations/flask/test_flask.py +++ b/tests/integrations/flask/test_flask.py @@ -394,6 +394,8 @@ def index(): client = app.test_client() response = client.post("/", data=data) assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() event, transaction_event = events @@ -746,6 +748,8 @@ def hi_tx(): with app.test_client() as client: response = client.get("/message_tx") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() message_event, transaction_event = events @@ -938,7 +942,9 @@ def test_response_status_code_not_found_in_transaction_context( envelopes = capture_envelopes() client = app.test_client() - client.get("/not-existing-route") + response = client.get("/not-existing-route") + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() sentry_sdk.get_client().flush() @@ -983,14 +989,21 @@ def test_transaction_http_method_default( events = capture_events() client = app.test_client() + response = client.get("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() response = client.options("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() response = client.head("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() (event,) = events @@ -1020,14 +1033,21 @@ def test_transaction_http_method_custom( events = capture_events() client = app.test_client() + response = client.get("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() response = client.options("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() response = client.head("/nomessage") assert response.status_code == 200 + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + response.close() assert len(events) == 2 diff --git a/tests/integrations/strawberry/test_strawberry.py b/tests/integrations/strawberry/test_strawberry.py index 7b40b238d2..0aab78f443 100644 --- a/tests/integrations/strawberry/test_strawberry.py +++ b/tests/integrations/strawberry/test_strawberry.py @@ -198,7 +198,10 @@ def test_capture_request_if_available_and_send_pii_is_on( client = client_factory(schema) query = "query ErrorQuery { error }" - client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post( + "/graphql", json={"query": query, "operationName": "ErrorQuery"} + ).close() assert len(events) == 1 @@ -253,7 +256,10 @@ def test_do_not_capture_request_if_send_pii_is_off( client = client_factory(schema) query = "query ErrorQuery { error }" - client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post( + "/graphql", json={"query": query, "operationName": "ErrorQuery"} + ).close() assert len(events) == 1 @@ -293,7 +299,8 @@ def test_breadcrumb_no_operation_name( client = client_factory(schema) query = "{ error }" - client.post("/graphql", json={"query": query}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={"query": query}).close() assert len(events) == 1 @@ -332,7 +339,10 @@ def test_capture_transaction_on_error( client = client_factory(schema) query = "query ErrorQuery { error }" - client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post( + "/graphql", json={"query": query, "operationName": "ErrorQuery"} + ).close() assert len(events) == 2 (_, transaction_event) = events @@ -409,7 +419,10 @@ def test_capture_transaction_on_success( client = client_factory(schema) query = "query GreetingQuery { hello }" - client.post("/graphql", json={"query": query, "operationName": "GreetingQuery"}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post( + "/graphql", json={"query": query, "operationName": "GreetingQuery"} + ).close() assert len(events) == 1 (transaction_event,) = events @@ -486,7 +499,8 @@ def test_transaction_no_operation_name( client = client_factory(schema) query = "{ hello }" - client.post("/graphql", json={"query": query}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={"query": query}).close() assert len(events) == 1 (transaction_event,) = events @@ -566,7 +580,8 @@ def test_transaction_mutation( client = client_factory(schema) query = 'mutation Change { change(attribute: "something") }' - client.post("/graphql", json={"query": query}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={"query": query}).close() assert len(events) == 1 (transaction_event,) = events @@ -641,7 +656,8 @@ def test_handle_none_query_gracefully( client_factory = request.getfixturevalue(client_factory) client = client_factory(schema) - client.post("/graphql", json={}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={}).close() assert len(events) == 0, "expected no events to be sent to Sentry" @@ -673,7 +689,8 @@ def test_span_origin( client = client_factory(schema) query = 'mutation Change { change(attribute: "something") }' - client.post("/graphql", json={"query": query}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={"query": query}).close() (event,) = events @@ -715,7 +732,10 @@ def test_span_origin2( client = client_factory(schema) query = "query GreetingQuery { hello }" - client.post("/graphql", json={"query": query, "operationName": "GreetingQuery"}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post( + "/graphql", json={"query": query, "operationName": "GreetingQuery"} + ).close() (event,) = events @@ -757,7 +777,8 @@ def test_span_origin3( client = client_factory(schema) query = "subscription { messageAdded { content } }" - client.post("/graphql", json={"query": query}) + # Close the response to ensure the WSGI cycle is complete and the transaction is finished + client.post("/graphql", json={"query": query}).close() (event,) = events diff --git a/tests/integrations/wsgi/test_wsgi.py b/tests/integrations/wsgi/test_wsgi.py index 656fc1757f..a4f5ca0623 100644 --- a/tests/integrations/wsgi/test_wsgi.py +++ b/tests/integrations/wsgi/test_wsgi.py @@ -1,7 +1,9 @@ +import time from collections import Counter from unittest import mock import pytest +from sentry_sdk.utils import datetime_from_isoformat from werkzeug.test import Client import sentry_sdk @@ -495,3 +497,80 @@ def dogpark(environ, start_response): (event,) = events assert event["contexts"]["trace"]["origin"] == "auto.dogpark.deluxe" + + +def test_long_running_transaction_finished(sentry_init, capture_events): + # we allow transactions to be 0.5 seconds as a maximum + new_max_duration = 0.5 + + with mock.patch.object( + sentry_sdk.integrations.wsgi, + "MAX_TRANSACTION_DURATION_SECONDS", + new_max_duration, + ): + + def generate_content(): + # This response will take 1.5 seconds to generate + for _ in range(15): + time.sleep(0.1) + yield "ok" + + def long_running_app(environ, start_response): + start_response("200 OK", []) + return generate_content() + + sentry_init(send_default_pii=True, traces_sample_rate=1.0) + app = SentryWsgiMiddleware(long_running_app) + + events = capture_events() + + client = Client(app) + response = client.get("/") + _ = response.get_data() + + (transaction,) = events + + transaction_duration = ( + datetime_from_isoformat(transaction["timestamp"]) + - datetime_from_isoformat(transaction["start_timestamp"]) + ).total_seconds() + assert ( + transaction_duration <= new_max_duration * 1.02 + ) # we allow 2% margin for processing the request + + +def test_long_running_transaction_timer_canceled(sentry_init, capture_events): + # we allow transactions to be 0.5 seconds as a maximum + new_max_duration = 0.5 + + with mock.patch.object( + sentry_sdk.integrations.wsgi, + "MAX_TRANSACTION_DURATION_SECONDS", + new_max_duration, + ): + with mock.patch( + "sentry_sdk.integrations.wsgi._finish_long_running_transaction" + ) as mock_finish: + + def generate_content(): + # This response will take 0.3 seconds to generate + for _ in range(3): + time.sleep(0.1) + yield "ok" + + def long_running_app(environ, start_response): + start_response("200 OK", []) + return generate_content() + + sentry_init(send_default_pii=True, traces_sample_rate=1.0) + app = SentryWsgiMiddleware(long_running_app) + + events = capture_events() + + client = Client(app) + response = client.get("/") + _ = response.get_data() + + (transaction,) = events + + mock_finish.assert_not_called() From 70224463e28eb26eb9c0af59233324ed79505cc2 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 27 Nov 2024 14:35:18 +0100 Subject: [PATCH 047/591] Use new clickhouse gh action (#3826) The docker image name of the official Clickhouse docker image changed, so I updated our GH action that starts that docker container and reference the new version here. --- .github/workflows/test-integrations-dbs.yml | 4 ++-- scripts/split-tox-gh-actions/templates/test_group.jinja | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 1612dfb432..a3ba66bc96 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -57,7 +57,7 @@ jobs: with: python-version: ${{ matrix.python-version }} allow-prereleases: true - - uses: getsentry/action-clickhouse-in-ci@v1 + - uses: getsentry/action-clickhouse-in-ci@v1.1 - name: Setup Test Env run: | pip install "coverage[toml]" tox @@ -152,7 +152,7 @@ jobs: with: python-version: ${{ matrix.python-version }} allow-prereleases: true - - uses: getsentry/action-clickhouse-in-ci@v1 + - uses: getsentry/action-clickhouse-in-ci@v1.1 - name: Setup Test Env run: | pip install "coverage[toml]" tox diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split-tox-gh-actions/templates/test_group.jinja index 4560a7d42d..b2de0d5393 100644 --- a/scripts/split-tox-gh-actions/templates/test_group.jinja +++ b/scripts/split-tox-gh-actions/templates/test_group.jinja @@ -51,7 +51,7 @@ python-version: {% raw %}${{ matrix.python-version }}{% endraw %} allow-prereleases: true {% if needs_clickhouse %} - - uses: getsentry/action-clickhouse-in-ci@v1 + - uses: getsentry/action-clickhouse-in-ci@v1.1 {% endif %} {% if needs_redis %} From 65b1791f5e4ec4f42a4e09caadaf7104e2875b22 Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Thu, 28 Nov 2024 03:59:18 -0800 Subject: [PATCH 048/591] ref(flags): rename launch darkly hook to match JS SDK (#3743) --- sentry_sdk/integrations/launchdarkly.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index 9e00e12ede..a9eef9e1a9 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -50,7 +50,7 @@ class LaunchDarklyHook(Hook): @property def metadata(self): # type: () -> Metadata - return Metadata(name="sentry-feature-flag-recorder") + return Metadata(name="sentry-flag-auditor") def after_evaluation(self, series_context, data, detail): # type: (EvaluationSeriesContext, dict[Any, Any], EvaluationDetail) -> dict[Any, Any] From e7130e88f6de728a66afc0209aa8f66190bd2f75 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 2 Dec 2024 11:18:10 +0100 Subject: [PATCH 049/591] Fix CI (#3834) The latest release of httpx seems to have broken the test clients of some older versions of Litestar, Starlite, Anthropic, Langchain, OpenAI, Starlette. Pinning httpx for old versions. Also tweaking what versions to test against. --- .github/workflows/test-integrations-ai.yml | 2 +- tox.ini | 58 ++++++++++++++-------- 2 files changed, 38 insertions(+), 22 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index c7cf4a1d85..7e48f62d06 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -99,7 +99,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.9","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/tox.ini b/tox.ini index 6acff6b8e8..0ecd2b697b 100644 --- a/tox.ini +++ b/tox.ini @@ -33,7 +33,7 @@ envlist = {py3.8,py3.12,py3.13}-aiohttp-latest # Anthropic - {py3.7,py3.11,py3.12}-anthropic-v{0.16,0.25} + {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} {py3.7,py3.11,py3.12}-anthropic-latest # Ariadne @@ -164,15 +164,14 @@ envlist = # Langchain {py3.9,py3.11,py3.12}-langchain-v0.1 + {py3.9,py3.11,py3.12}-langchain-v0.3 {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken # Litestar - # litestar 2.0.0 is the earliest version that supports Python < 3.12 {py3.8,py3.11}-litestar-v{2.0} - # litestar 2.3.0 is the earliest version that supports Python 3.12 - {py3.12}-litestar-v{2.3} - {py3.8,py3.11,py3.12}-litestar-v{2.5} + {py3.8,py3.11,py3.12}-litestar-v{2.6} + {py3.8,py3.11,py3.12}-litestar-v{2.12} {py3.8,py3.11,py3.12}-litestar-latest # Loguru @@ -180,7 +179,9 @@ envlist = {py3.6,py3.12,py3.13}-loguru-latest # OpenAI - {py3.9,py3.11,py3.12}-openai-v1 + {py3.9,py3.11,py3.12}-openai-v1.0 + {py3.9,py3.11,py3.12}-openai-v1.22 + {py3.9,py3.11,py3.12}-openai-v1.55 {py3.9,py3.11,py3.12}-openai-latest {py3.9,py3.11,py3.12}-openai-notiktoken @@ -256,8 +257,8 @@ envlist = # Starlette {py3.7,py3.10}-starlette-v{0.19} - {py3.7,py3.11}-starlette-v{0.20,0.24,0.28} - {py3.8,py3.11,py3.12}-starlette-v{0.32,0.36} + {py3.7,py3.11}-starlette-v{0.24,0.28} + {py3.8,py3.11,py3.12}-starlette-v{0.32,0.36,0.40} {py3.8,py3.12,py3.13}-starlette-latest # Starlite @@ -326,8 +327,10 @@ deps = # Anthropic anthropic: pytest-asyncio - anthropic-v0.25: anthropic~=0.25.0 + anthropic-v{0.16,0.28}: httpx<0.28.0 anthropic-v0.16: anthropic~=0.16.0 + anthropic-v0.28: anthropic~=0.28.0 + anthropic-v0.40: anthropic~=0.40.0 anthropic-latest: anthropic # Ariadne @@ -404,6 +407,7 @@ deps = django: psycopg2-binary django-v{1.11,2.0,2.1,2.2,3.0,3.1,3.2}: djangorestframework>=3.0.0,<4.0.0 django-v{2.0,2.2,3.0,3.2,4.0,4.1,4.2,5.0,5.1}: channels[daphne] + django-v{2.2,3.0}: six django-v{1.11,2.0,2.2,3.0,3.2}: Werkzeug<2.1.0 django-v{1.11,2.0,2.2,3.0}: pytest-django<4.0 django-v{3.2,4.0,4.1,4.2,5.0,5.1}: pytest-django @@ -517,22 +521,25 @@ deps = langchain-v0.1: openai~=1.0.0 langchain-v0.1: langchain~=0.1.11 langchain-v0.1: tiktoken~=0.6.0 - langchain-latest: langchain - langchain-latest: langchain-openai - langchain-latest: openai>=1.6.1 + langchain-v0.1: httpx<0.28.0 + langchain-v0.3: langchain~=0.3.0 + langchain-v0.3: langchain-community + langchain-v0.3: tiktoken + langchain-v0.3: openai + langchain-{latest,notiktoken}: langchain + langchain-{latest,notiktoken}: langchain-openai + langchain-{latest,notiktoken}: openai>=1.6.1 langchain-latest: tiktoken~=0.6.0 - langchain-notiktoken: langchain - langchain-notiktoken: langchain-openai - langchain-notiktoken: openai>=1.6.1 # Litestar litestar: pytest-asyncio litestar: python-multipart litestar: requests litestar: cryptography + litestar-v{2.0,2.6}: httpx<0.28 litestar-v2.0: litestar~=2.0.0 - litestar-v2.3: litestar~=2.3.0 - litestar-v2.5: litestar~=2.5.0 + litestar-v2.6: litestar~=2.6.0 + litestar-v2.12: litestar~=2.12.0 litestar-latest: litestar # Loguru @@ -541,8 +548,14 @@ deps = # OpenAI openai: pytest-asyncio - openai-v1: openai~=1.0.0 - openai-v1: tiktoken~=0.6.0 + openai-v1.0: openai~=1.0.0 + openai-v1.0: tiktoken + openai-v1.0: httpx<0.28.0 + openai-v1.22: openai~=1.22.0 + openai-v1.22: tiktoken + openai-v1.22: httpx<0.28.0 + openai-v1.55: openai~=1.55.0 + openai-v1.55: tiktoken openai-latest: openai openai-latest: tiktoken~=0.6.0 openai-notiktoken: openai @@ -655,16 +668,18 @@ deps = starlette: pytest-asyncio starlette: python-multipart starlette: requests - starlette: httpx # (this is a dependency of httpx) starlette: anyio<4.0.0 starlette: jinja2 + starlette-v{0.19,0.24,0.28,0.32,0.36}: httpx<0.28.0 + starlette-v0.40: httpx + starlette-latest: httpx starlette-v0.19: starlette~=0.19.0 - starlette-v0.20: starlette~=0.20.0 starlette-v0.24: starlette~=0.24.0 starlette-v0.28: starlette~=0.28.0 starlette-v0.32: starlette~=0.32.0 starlette-v0.36: starlette~=0.36.0 + starlette-v0.40: starlette~=0.40.0 starlette-latest: starlette # Starlite @@ -673,6 +688,7 @@ deps = starlite: requests starlite: cryptography starlite: pydantic<2.0.0 + starlite: httpx<0.28 starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 From c4274a30d495888ce00fecef21f4a25805d84fad Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 2 Dec 2024 13:42:14 +0000 Subject: [PATCH 050/591] build(deps): bump codecov/codecov-action from 5.0.2 to 5.0.7 (#3821) * build(deps): bump codecov/codecov-action from 5.0.2 to 5.0.7 Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.0.2 to 5.0.7. - [Release notes](https://github.com/codecov/codecov-action/releases) - [Changelog](https://github.com/codecov/codecov-action/blob/main/CHANGELOG.md) - [Commits](https://github.com/codecov/codecov-action/compare/v5.0.2...v5.0.7) --- updated-dependencies: - dependency-name: codecov/codecov-action dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] * template --------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-misc.yml | 4 ++-- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 4 ++-- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split-tox-gh-actions/templates/test_group.jinja | 2 +- 12 files changed, 21 insertions(+), 21 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 7e48f62d06..5d1b05add8 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -78,7 +78,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -150,7 +150,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index 67c0ec31c7..d2ce22f326 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 62d67200a5..8fdd4a0649 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 6983a079ef..8294b9480e 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -62,7 +62,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index a3ba66bc96..0d9a7bbd7d 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -101,7 +101,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -196,7 +196,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 57d14cff10..30480efe2e 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 5f2baa5759..fb76a854fb 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -86,7 +86,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 7c1c343aac..0a51866164 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 1c4259ac05..695c338721 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 6a6a01e8ff..6e172182b3 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 11cfc20612..f9f2651cb8 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -98,7 +98,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -190,7 +190,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split-tox-gh-actions/templates/test_group.jinja index b2de0d5393..522be6dc5c 100644 --- a/scripts/split-tox-gh-actions/templates/test_group.jinja +++ b/scripts/split-tox-gh-actions/templates/test_group.jinja @@ -92,7 +92,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.0.2 + uses: codecov/codecov-action@v5.0.7 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From 6bd7e08694829aade11fc60ee628f04ceeabc7dc Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Mon, 2 Dec 2024 15:51:29 +0100 Subject: [PATCH 051/591] Fix asyncio testing setup (#3832) * Fix asyncio testing setup * default `asyncio_default_fixture_loop_scope` to `function` to get rid of deprecation messages * Change `test_asyncio.py` event loop scopes to `module` to avoid that event loop bleeding into all other tests in the same `session`. * Remove explicit `event_loop`s since `pytest-asyncio` takes care of those * Bump asyncio tests to 3.8 min --- pytest.ini | 1 + tests/integrations/asyncio/test_asyncio.py | 57 +++++++++------------- tests/integrations/grpc/test_grpc_aio.py | 16 ++---- 3 files changed, 28 insertions(+), 46 deletions(-) diff --git a/pytest.ini b/pytest.ini index c03752b039..7edd6127b9 100644 --- a/pytest.ini +++ b/pytest.ini @@ -1,6 +1,7 @@ [pytest] addopts = -vvv -rfEs -s --durations=5 --cov=./sentry_sdk --cov-branch --cov-report= --tb=short --junitxml=.junitxml asyncio_mode = strict +asyncio_default_fixture_loop_scope = function markers = tests_internal_exceptions: Handle internal exceptions just as the SDK does, to test it. (Otherwise internal exceptions are recorded and reraised.) diff --git a/tests/integrations/asyncio/test_asyncio.py b/tests/integrations/asyncio/test_asyncio.py index c9e572ca73..fb75bfc69b 100644 --- a/tests/integrations/asyncio/test_asyncio.py +++ b/tests/integrations/asyncio/test_asyncio.py @@ -15,8 +15,8 @@ pass # All tests will be skipped with incompatible versions -minimum_python_37 = pytest.mark.skipif( - sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" +minimum_python_38 = pytest.mark.skipif( + sys.version_info < (3, 8), reason="Asyncio tests need Python >= 3.8" ) @@ -38,14 +38,6 @@ async def boom(): 1 / 0 -@pytest.fixture(scope="session") -def event_loop(request): - """Create an instance of the default event loop for each test case.""" - loop = asyncio.get_event_loop_policy().new_event_loop() - yield loop - loop.close() - - def get_sentry_task_factory(mock_get_running_loop): """ Patches (mocked) asyncio and gets the sentry_task_factory. @@ -57,12 +49,11 @@ def get_sentry_task_factory(mock_get_running_loop): return patched_factory -@minimum_python_37 -@pytest.mark.asyncio +@minimum_python_38 +@pytest.mark.asyncio(loop_scope="module") async def test_create_task( sentry_init, capture_events, - event_loop, ): sentry_init( traces_sample_rate=1.0, @@ -76,10 +67,10 @@ async def test_create_task( with sentry_sdk.start_transaction(name="test_transaction_for_create_task"): with sentry_sdk.start_span(op="root", name="not so important"): - tasks = [event_loop.create_task(foo()), event_loop.create_task(bar())] + tasks = [asyncio.create_task(foo()), asyncio.create_task(bar())] await asyncio.wait(tasks, return_when=asyncio.FIRST_EXCEPTION) - sentry_sdk.flush() + sentry_sdk.flush() (transaction_event,) = events @@ -101,8 +92,8 @@ async def test_create_task( ) -@minimum_python_37 -@pytest.mark.asyncio +@minimum_python_38 +@pytest.mark.asyncio(loop_scope="module") async def test_gather( sentry_init, capture_events, @@ -121,7 +112,7 @@ async def test_gather( with sentry_sdk.start_span(op="root", name="not so important"): await asyncio.gather(foo(), bar(), return_exceptions=True) - sentry_sdk.flush() + sentry_sdk.flush() (transaction_event,) = events @@ -143,12 +134,11 @@ async def test_gather( ) -@minimum_python_37 -@pytest.mark.asyncio +@minimum_python_38 +@pytest.mark.asyncio(loop_scope="module") async def test_exception( sentry_init, capture_events, - event_loop, ): sentry_init( traces_sample_rate=1.0, @@ -162,10 +152,10 @@ async def test_exception( with sentry_sdk.start_transaction(name="test_exception"): with sentry_sdk.start_span(op="root", name="not so important"): - tasks = [event_loop.create_task(boom()), event_loop.create_task(bar())] + tasks = [asyncio.create_task(boom()), asyncio.create_task(bar())] await asyncio.wait(tasks, return_when=asyncio.FIRST_EXCEPTION) - sentry_sdk.flush() + sentry_sdk.flush() (error_event, _) = events @@ -177,8 +167,8 @@ async def test_exception( assert error_event["exception"]["values"][0]["mechanism"]["type"] == "asyncio" -@minimum_python_37 -@pytest.mark.asyncio +@minimum_python_38 +@pytest.mark.asyncio(loop_scope="module") async def test_task_result(sentry_init): sentry_init( integrations=[ @@ -194,7 +184,7 @@ async def add(a, b): @minimum_python_311 -@pytest.mark.asyncio +@pytest.mark.asyncio(loop_scope="module") async def test_task_with_context(sentry_init): """ Integration test to ensure working context parameter in Python 3.11+ @@ -223,7 +213,7 @@ async def retrieve_value(): assert retrieve_task.result() == "changed value" -@minimum_python_37 +@minimum_python_38 @patch("asyncio.get_running_loop") def test_patch_asyncio(mock_get_running_loop): """ @@ -242,7 +232,7 @@ def test_patch_asyncio(mock_get_running_loop): assert callable(sentry_task_factory) -@minimum_python_37 +@minimum_python_38 @patch("asyncio.get_running_loop") @patch("sentry_sdk.integrations.asyncio.Task") def test_sentry_task_factory_no_factory(MockTask, mock_get_running_loop): # noqa: N803 @@ -271,7 +261,7 @@ def test_sentry_task_factory_no_factory(MockTask, mock_get_running_loop): # noq assert task_kwargs["loop"] == mock_loop -@minimum_python_37 +@minimum_python_38 @patch("asyncio.get_running_loop") def test_sentry_task_factory_with_factory(mock_get_running_loop): mock_loop = mock_get_running_loop.return_value @@ -361,12 +351,11 @@ def test_sentry_task_factory_context_with_factory(mock_get_running_loop): assert task_factory_kwargs["context"] == mock_context -@minimum_python_37 -@pytest.mark.asyncio +@minimum_python_38 +@pytest.mark.asyncio(loop_scope="module") async def test_span_origin( sentry_init, capture_events, - event_loop, ): sentry_init( integrations=[AsyncioIntegration()], @@ -377,11 +366,11 @@ async def test_span_origin( with sentry_sdk.start_transaction(name="something"): tasks = [ - event_loop.create_task(foo()), + asyncio.create_task(foo()), ] await asyncio.wait(tasks, return_when=asyncio.FIRST_EXCEPTION) - sentry_sdk.flush() + sentry_sdk.flush() (event,) = events diff --git a/tests/integrations/grpc/test_grpc_aio.py b/tests/integrations/grpc/test_grpc_aio.py index fff22626d9..9ce9aef6a5 100644 --- a/tests/integrations/grpc/test_grpc_aio.py +++ b/tests/integrations/grpc/test_grpc_aio.py @@ -21,22 +21,14 @@ AIO_PORT += os.getpid() % 100 # avoid port conflicts when running tests in parallel -@pytest.fixture(scope="function") -def event_loop(request): - """Create an instance of the default event loop for each test case.""" - loop = asyncio.new_event_loop() - yield loop - loop.close() - - @pytest_asyncio.fixture(scope="function") -async def grpc_server(sentry_init, event_loop): +async def grpc_server(sentry_init): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) server = grpc.aio.server() server.add_insecure_port("[::]:{}".format(AIO_PORT)) add_gRPCTestServiceServicer_to_server(TestService, server) - await event_loop.create_task(server.start()) + await asyncio.create_task(server.start()) try: yield server @@ -45,12 +37,12 @@ async def grpc_server(sentry_init, event_loop): @pytest.mark.asyncio -async def test_noop_for_unimplemented_method(event_loop, sentry_init, capture_events): +async def test_noop_for_unimplemented_method(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) server = grpc.aio.server() server.add_insecure_port("[::]:{}".format(AIO_PORT)) - await event_loop.create_task(server.start()) + await asyncio.create_task(server.start()) events = capture_events() try: From 3d8445c0339f61903ade6be72c3e3d5890503b39 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 2 Dec 2024 16:21:32 +0100 Subject: [PATCH 052/591] Revert "Fix spans for streaming responses in WSGI based frameworks (#3798)" (#3836) This reverts commit da206237473aeb38d911d9cd86f40bd928a2a350. (PR #3798) Having a timer thread on each request is too much overhead on high volume servers. --- sentry_sdk/integrations/wsgi.py | 135 ++++++------------ sentry_sdk/tracing_utils.py | 18 --- tests/integrations/django/test_basic.py | 46 +++--- tests/integrations/flask/test_flask.py | 22 +-- .../strawberry/test_strawberry.py | 43 ++---- tests/integrations/wsgi/test_wsgi.py | 79 ---------- 6 files changed, 73 insertions(+), 270 deletions(-) diff --git a/sentry_sdk/integrations/wsgi.py b/sentry_sdk/integrations/wsgi.py index 751735f462..50deae10c5 100644 --- a/sentry_sdk/integrations/wsgi.py +++ b/sentry_sdk/integrations/wsgi.py @@ -1,19 +1,19 @@ import sys from functools import partial -from threading import Timer import sentry_sdk from sentry_sdk._werkzeug import get_host, _get_headers from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP -from sentry_sdk.scope import should_send_default_pii, use_isolation_scope, use_scope +from sentry_sdk.scope import should_send_default_pii from sentry_sdk.integrations._wsgi_common import ( DEFAULT_HTTP_METHODS_TO_CAPTURE, _filter_headers, + nullcontext, ) from sentry_sdk.sessions import track_session +from sentry_sdk.scope import use_isolation_scope from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_ROUTE -from sentry_sdk.tracing_utils import finish_running_transaction from sentry_sdk.utils import ( ContextVar, capture_internal_exceptions, @@ -46,9 +46,6 @@ def __call__(self, status, response_headers, exc_info=None): # type: ignore pass -MAX_TRANSACTION_DURATION_SECONDS = 5 * 60 - - _wsgi_middleware_applied = ContextVar("sentry_wsgi_middleware_applied") @@ -101,7 +98,6 @@ def __call__(self, environ, start_response): _wsgi_middleware_applied.set(True) try: with sentry_sdk.isolation_scope() as scope: - current_scope = sentry_sdk.get_current_scope() with track_session(scope, session_mode="request"): with capture_internal_exceptions(): scope.clear_breadcrumbs() @@ -113,7 +109,6 @@ def __call__(self, environ, start_response): ) method = environ.get("REQUEST_METHOD", "").upper() - transaction = None if method in self.http_methods_to_capture: transaction = continue_trace( @@ -124,43 +119,27 @@ def __call__(self, environ, start_response): origin=self.span_origin, ) - timer = None - if transaction is not None: + with ( sentry_sdk.start_transaction( transaction, custom_sampling_context={"wsgi_environ": environ}, - ).__enter__() - timer = Timer( - MAX_TRANSACTION_DURATION_SECONDS, - _finish_long_running_transaction, - args=(current_scope, scope), ) - timer.start() - - try: - response = self.app( - environ, - partial( - _sentry_start_response, - start_response, - transaction, - ), - ) - except BaseException: - exc_info = sys.exc_info() - _capture_exception(exc_info) - finish_running_transaction(current_scope, exc_info, timer) - reraise(*exc_info) - + if transaction is not None + else nullcontext() + ): + try: + response = self.app( + environ, + partial( + _sentry_start_response, start_response, transaction + ), + ) + except BaseException: + reraise(*_capture_exception()) finally: _wsgi_middleware_applied.set(False) - return _ScopedResponse( - response=response, - current_scope=current_scope, - isolation_scope=scope, - timer=timer, - ) + return _ScopedResponse(scope, response) def _sentry_start_response( # type: ignore @@ -222,13 +201,13 @@ def get_client_ip(environ): return environ.get("REMOTE_ADDR") -def _capture_exception(exc_info=None): - # type: (Optional[ExcInfo]) -> ExcInfo +def _capture_exception(): + # type: () -> ExcInfo """ Captures the current exception and sends it to Sentry. Returns the ExcInfo tuple to it can be reraised afterwards. """ - exc_info = exc_info or sys.exc_info() + exc_info = sys.exc_info() e = exc_info[1] # SystemExit(0) is the only uncaught exception that is expected behavior @@ -246,7 +225,7 @@ def _capture_exception(exc_info=None): class _ScopedResponse: """ - Use separate scopes for each response chunk. + Users a separate scope for each response chunk. This will make WSGI apps more tolerant against: - WSGI servers streaming responses from a different thread/from @@ -255,54 +234,37 @@ class _ScopedResponse: - WSGI servers streaming responses interleaved from the same thread """ - __slots__ = ("_response", "_current_scope", "_isolation_scope", "_timer") + __slots__ = ("_response", "_scope") - def __init__( - self, - response, # type: Iterator[bytes] - current_scope, # type: sentry_sdk.scope.Scope - isolation_scope, # type: sentry_sdk.scope.Scope - timer=None, # type: Optional[Timer] - ): - # type: (...) -> None + def __init__(self, scope, response): + # type: (sentry_sdk.scope.Scope, Iterator[bytes]) -> None + self._scope = scope self._response = response - self._current_scope = current_scope - self._isolation_scope = isolation_scope - self._timer = timer def __iter__(self): # type: () -> Iterator[bytes] iterator = iter(self._response) - try: - while True: - with use_isolation_scope(self._isolation_scope): - with use_scope(self._current_scope): - try: - chunk = next(iterator) - except StopIteration: - break - except BaseException: - reraise(*_capture_exception()) - - yield chunk + while True: + with use_isolation_scope(self._scope): + try: + chunk = next(iterator) + except StopIteration: + break + except BaseException: + reraise(*_capture_exception()) - finally: - with use_isolation_scope(self._isolation_scope): - with use_scope(self._current_scope): - finish_running_transaction(timer=self._timer) + yield chunk def close(self): # type: () -> None - with use_isolation_scope(self._isolation_scope): - with use_scope(self._current_scope): - try: - finish_running_transaction(timer=self._timer) - self._response.close() # type: ignore - except AttributeError: - pass - except BaseException: - reraise(*_capture_exception()) + with use_isolation_scope(self._scope): + try: + self._response.close() # type: ignore + except AttributeError: + pass + except BaseException: + reraise(*_capture_exception()) def _make_wsgi_event_processor(environ, use_x_forwarded_for): @@ -346,18 +308,3 @@ def event_processor(event, hint): return event return event_processor - - -def _finish_long_running_transaction(current_scope, isolation_scope): - # type: (sentry_sdk.scope.Scope, sentry_sdk.scope.Scope) -> None - """ - Make sure we don't keep transactions open for too long. - Triggered after MAX_TRANSACTION_DURATION_SECONDS have passed. - """ - try: - with use_isolation_scope(isolation_scope): - with use_scope(current_scope): - finish_running_transaction() - except AttributeError: - # transaction is not there anymore - pass diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 969e0812e4..0459563776 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -36,9 +36,6 @@ from types import FrameType - from sentry_sdk._types import ExcInfo - from threading import Timer - SENTRY_TRACE_REGEX = re.compile( "^[ \t]*" # whitespace @@ -742,18 +739,3 @@ def get_current_span(scope=None): if TYPE_CHECKING: from sentry_sdk.tracing import Span - - -def finish_running_transaction(scope=None, exc_info=None, timer=None): - # type: (Optional[sentry_sdk.Scope], Optional[ExcInfo], Optional[Timer]) -> None - if timer is not None: - timer.cancel() - - current_scope = scope or sentry_sdk.get_current_scope() - if current_scope.transaction is not None and hasattr( - current_scope.transaction, "_context_manager_state" - ): - if exc_info is not None: - current_scope.transaction.__exit__(*exc_info) - else: - current_scope.transaction.__exit__(None, None, None) diff --git a/tests/integrations/django/test_basic.py b/tests/integrations/django/test_basic.py index 243431fdf5..0e3f700105 100644 --- a/tests/integrations/django/test_basic.py +++ b/tests/integrations/django/test_basic.py @@ -51,7 +51,7 @@ def test_view_exceptions(sentry_init, client, capture_exceptions, capture_events sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - unpack_werkzeug_response(client.get(reverse("view_exc"))) + client.get(reverse("view_exc")) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -72,9 +72,7 @@ def test_ensures_x_forwarded_header_is_honored_in_sdk_when_enabled_in_django( sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - unpack_werkzeug_response( - client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) - ) + client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -93,9 +91,7 @@ def test_ensures_x_forwarded_header_is_not_honored_when_unenabled_in_django( sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() events = capture_events() - unpack_werkzeug_response( - client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) - ) + client.get(reverse("view_exc"), headers={"X_FORWARDED_HOST": "example.com"}) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -107,7 +103,7 @@ def test_ensures_x_forwarded_header_is_not_honored_when_unenabled_in_django( def test_middleware_exceptions(sentry_init, client, capture_exceptions): sentry_init(integrations=[DjangoIntegration()], send_default_pii=True) exceptions = capture_exceptions() - unpack_werkzeug_response(client.get(reverse("middleware_exc"))) + client.get(reverse("middleware_exc")) (error,) = exceptions assert isinstance(error, ZeroDivisionError) @@ -161,7 +157,7 @@ def test_has_trace_if_performance_enabled(sentry_init, client, capture_events): traces_sample_rate=1.0, ) events = capture_events() - unpack_werkzeug_response(client.head(reverse("view_exc_with_msg"))) + client.head(reverse("view_exc_with_msg")) (msg_event, error_event, transaction_event) = events @@ -217,10 +213,8 @@ def test_trace_from_headers_if_performance_enabled(sentry_init, client, capture_ trace_id = "582b43a4192642f0b136d5159a501701" sentry_trace_header = "{}-{}-{}".format(trace_id, "6e8f22c393e68f19", 1) - unpack_werkzeug_response( - client.head( - reverse("view_exc_with_msg"), headers={"sentry-trace": sentry_trace_header} - ) + client.head( + reverse("view_exc_with_msg"), headers={"sentry-trace": sentry_trace_header} ) (msg_event, error_event, transaction_event) = events @@ -934,7 +928,7 @@ def test_render_spans(sentry_init, client, capture_events, render_span_tree): for url, expected_line in views_tests: events = capture_events() - unpack_werkzeug_response(client.get(url)) + client.get(url) transaction = events[0] assert expected_line in render_span_tree(transaction) @@ -973,7 +967,7 @@ def test_middleware_spans(sentry_init, client, capture_events, render_span_tree) ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("message"))) + client.get(reverse("message")) message, transaction = events @@ -990,7 +984,7 @@ def test_middleware_spans_disabled(sentry_init, client, capture_events): ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("message"))) + client.get(reverse("message")) message, transaction = events @@ -1014,7 +1008,7 @@ def test_signals_spans(sentry_init, client, capture_events, render_span_tree): ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("message"))) + client.get(reverse("message")) message, transaction = events @@ -1037,7 +1031,7 @@ def test_signals_spans_disabled(sentry_init, client, capture_events): ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("message"))) + client.get(reverse("message")) message, transaction = events @@ -1067,7 +1061,7 @@ def test_signals_spans_filtering(sentry_init, client, capture_events, render_spa ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("send_myapp_custom_signal"))) + client.get(reverse("send_myapp_custom_signal")) (transaction,) = events @@ -1192,7 +1186,7 @@ def test_span_origin(sentry_init, client, capture_events): ) events = capture_events() - unpack_werkzeug_response(client.get(reverse("view_with_signal"))) + client.get(reverse("view_with_signal")) (transaction,) = events @@ -1217,9 +1211,9 @@ def test_transaction_http_method_default(sentry_init, client, capture_events): ) events = capture_events() - unpack_werkzeug_response(client.get("/nomessage")) - unpack_werkzeug_response(client.options("/nomessage")) - unpack_werkzeug_response(client.head("/nomessage")) + client.get("/nomessage") + client.options("/nomessage") + client.head("/nomessage") (event,) = events @@ -1241,9 +1235,9 @@ def test_transaction_http_method_custom(sentry_init, client, capture_events): ) events = capture_events() - unpack_werkzeug_response(client.get("/nomessage")) - unpack_werkzeug_response(client.options("/nomessage")) - unpack_werkzeug_response(client.head("/nomessage")) + client.get("/nomessage") + client.options("/nomessage") + client.head("/nomessage") assert len(events) == 2 diff --git a/tests/integrations/flask/test_flask.py b/tests/integrations/flask/test_flask.py index e2c37aa5f7..6febb12b8b 100644 --- a/tests/integrations/flask/test_flask.py +++ b/tests/integrations/flask/test_flask.py @@ -394,8 +394,6 @@ def index(): client = app.test_client() response = client.post("/", data=data) assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() event, transaction_event = events @@ -748,8 +746,6 @@ def hi_tx(): with app.test_client() as client: response = client.get("/message_tx") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() message_event, transaction_event = events @@ -942,9 +938,7 @@ def test_response_status_code_not_found_in_transaction_context( envelopes = capture_envelopes() client = app.test_client() - response = client.get("/not-existing-route") - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() + client.get("/not-existing-route") sentry_sdk.get_client().flush() @@ -989,21 +983,14 @@ def test_transaction_http_method_default( events = capture_events() client = app.test_client() - response = client.get("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() response = client.options("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() response = client.head("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() (event,) = events @@ -1033,21 +1020,14 @@ def test_transaction_http_method_custom( events = capture_events() client = app.test_client() - response = client.get("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() response = client.options("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() response = client.head("/nomessage") assert response.status_code == 200 - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - response.close() assert len(events) == 2 diff --git a/tests/integrations/strawberry/test_strawberry.py b/tests/integrations/strawberry/test_strawberry.py index 0aab78f443..7b40b238d2 100644 --- a/tests/integrations/strawberry/test_strawberry.py +++ b/tests/integrations/strawberry/test_strawberry.py @@ -198,10 +198,7 @@ def test_capture_request_if_available_and_send_pii_is_on( client = client_factory(schema) query = "query ErrorQuery { error }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post( - "/graphql", json={"query": query, "operationName": "ErrorQuery"} - ).close() + client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) assert len(events) == 1 @@ -256,10 +253,7 @@ def test_do_not_capture_request_if_send_pii_is_off( client = client_factory(schema) query = "query ErrorQuery { error }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post( - "/graphql", json={"query": query, "operationName": "ErrorQuery"} - ).close() + client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) assert len(events) == 1 @@ -299,8 +293,7 @@ def test_breadcrumb_no_operation_name( client = client_factory(schema) query = "{ error }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={"query": query}).close() + client.post("/graphql", json={"query": query}) assert len(events) == 1 @@ -339,10 +332,7 @@ def test_capture_transaction_on_error( client = client_factory(schema) query = "query ErrorQuery { error }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post( - "/graphql", json={"query": query, "operationName": "ErrorQuery"} - ).close() + client.post("/graphql", json={"query": query, "operationName": "ErrorQuery"}) assert len(events) == 2 (_, transaction_event) = events @@ -419,10 +409,7 @@ def test_capture_transaction_on_success( client = client_factory(schema) query = "query GreetingQuery { hello }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post( - "/graphql", json={"query": query, "operationName": "GreetingQuery"} - ).close() + client.post("/graphql", json={"query": query, "operationName": "GreetingQuery"}) assert len(events) == 1 (transaction_event,) = events @@ -499,8 +486,7 @@ def test_transaction_no_operation_name( client = client_factory(schema) query = "{ hello }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={"query": query}).close() + client.post("/graphql", json={"query": query}) assert len(events) == 1 (transaction_event,) = events @@ -580,8 +566,7 @@ def test_transaction_mutation( client = client_factory(schema) query = 'mutation Change { change(attribute: "something") }' - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={"query": query}).close() + client.post("/graphql", json={"query": query}) assert len(events) == 1 (transaction_event,) = events @@ -656,8 +641,7 @@ def test_handle_none_query_gracefully( client_factory = request.getfixturevalue(client_factory) client = client_factory(schema) - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={}).close() + client.post("/graphql", json={}) assert len(events) == 0, "expected no events to be sent to Sentry" @@ -689,8 +673,7 @@ def test_span_origin( client = client_factory(schema) query = 'mutation Change { change(attribute: "something") }' - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={"query": query}).close() + client.post("/graphql", json={"query": query}) (event,) = events @@ -732,10 +715,7 @@ def test_span_origin2( client = client_factory(schema) query = "query GreetingQuery { hello }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post( - "/graphql", json={"query": query, "operationName": "GreetingQuery"} - ).close() + client.post("/graphql", json={"query": query, "operationName": "GreetingQuery"}) (event,) = events @@ -777,8 +757,7 @@ def test_span_origin3( client = client_factory(schema) query = "subscription { messageAdded { content } }" - # Close the response to ensure the WSGI cycle is complete and the transaction is finished - client.post("/graphql", json={"query": query}).close() + client.post("/graphql", json={"query": query}) (event,) = events diff --git a/tests/integrations/wsgi/test_wsgi.py b/tests/integrations/wsgi/test_wsgi.py index a4f5ca0623..656fc1757f 100644 --- a/tests/integrations/wsgi/test_wsgi.py +++ b/tests/integrations/wsgi/test_wsgi.py @@ -1,9 +1,7 @@ -import time from collections import Counter from unittest import mock import pytest -from sentry_sdk.utils import datetime_from_isoformat from werkzeug.test import Client import sentry_sdk @@ -497,80 +495,3 @@ def dogpark(environ, start_response): (event,) = events assert event["contexts"]["trace"]["origin"] == "auto.dogpark.deluxe" - - -def test_long_running_transaction_finished(sentry_init, capture_events): - # we allow transactions to be 0.5 seconds as a maximum - new_max_duration = 0.5 - - with mock.patch.object( - sentry_sdk.integrations.wsgi, - "MAX_TRANSACTION_DURATION_SECONDS", - new_max_duration, - ): - - def generate_content(): - # This response will take 1.5 seconds to generate - for _ in range(15): - time.sleep(0.1) - yield "ok" - - def long_running_app(environ, start_response): - start_response("200 OK", []) - return generate_content() - - sentry_init(send_default_pii=True, traces_sample_rate=1.0) - app = SentryWsgiMiddleware(long_running_app) - - events = capture_events() - - client = Client(app) - response = client.get("/") - _ = response.get_data() - - (transaction,) = events - - transaction_duration = ( - datetime_from_isoformat(transaction["timestamp"]) - - datetime_from_isoformat(transaction["start_timestamp"]) - ).total_seconds() - assert ( - transaction_duration <= new_max_duration * 1.02 - ) # we allow 2% margin for processing the request - - -def test_long_running_transaction_timer_canceled(sentry_init, capture_events): - # we allow transactions to be 0.5 seconds as a maximum - new_max_duration = 0.5 - - with mock.patch.object( - sentry_sdk.integrations.wsgi, - "MAX_TRANSACTION_DURATION_SECONDS", - new_max_duration, - ): - with mock.patch( - "sentry_sdk.integrations.wsgi._finish_long_running_transaction" - ) as mock_finish: - - def generate_content(): - # This response will take 0.3 seconds to generate - for _ in range(3): - time.sleep(0.1) - yield "ok" - - def long_running_app(environ, start_response): - start_response("200 OK", []) - return generate_content() - - sentry_init(send_default_pii=True, traces_sample_rate=1.0) - app = SentryWsgiMiddleware(long_running_app) - - events = capture_events() - - client = Client(app) - response = client.get("/") - _ = response.get_data() - - (transaction,) = events - - mock_finish.assert_not_called() From dfb84cc499335fdbf674fa32b8247316faf087f1 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 3 Dec 2024 16:49:53 +0100 Subject: [PATCH 053/591] Test with celery 5.5.0rc3 (#3842) --- tox.ini | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/tox.ini b/tox.ini index 0ecd2b697b..8c6f9eda86 100644 --- a/tox.ini +++ b/tox.ini @@ -75,7 +75,7 @@ envlist = {py3.6,py3.8}-celery-v{4} {py3.6,py3.8}-celery-v{5.0} {py3.7,py3.10}-celery-v{5.1,5.2} - {py3.8,py3.11,py3.12}-celery-v{5.3,5.4} + {py3.8,py3.11,py3.12}-celery-v{5.3,5.4,5.5} {py3.8,py3.12,py3.13}-celery-latest # Chalice @@ -383,6 +383,8 @@ deps = celery-v5.2: Celery~=5.2.0 celery-v5.3: Celery~=5.3.0 celery-v5.4: Celery~=5.4.0 + # TODO: update when stable is out + celery-v5.5: Celery==5.5.0rc3 celery-latest: Celery celery: newrelic From 3e43a91b0e7f90f73a4165f7b58d5a10567e19bc Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 4 Dec 2024 15:41:04 +0100 Subject: [PATCH 054/591] Improve ray tests (#3846) * Make ray tests actually test something and show that actors are not supported --- tests/integrations/ray/test_ray.py | 167 ++++++++++++++++------------- 1 file changed, 92 insertions(+), 75 deletions(-) diff --git a/tests/integrations/ray/test_ray.py b/tests/integrations/ray/test_ray.py index 02c08c2a9e..95ab4ad0fa 100644 --- a/tests/integrations/ray/test_ray.py +++ b/tests/integrations/ray/test_ray.py @@ -39,8 +39,27 @@ def setup_sentry(transport=None): ) +def read_error_from_log(job_id): + log_dir = "/tmp/ray/session_latest/logs/" + log_file = [ + f + for f in os.listdir(log_dir) + if "worker" in f and job_id in f and f.endswith(".out") + ][0] + with open(os.path.join(log_dir, log_file), "r") as file: + lines = file.readlines() + + try: + # parse error object from log line + error = json.loads(lines[4][:-1]) + except IndexError: + error = None + + return error + + @pytest.mark.forked -def test_ray_tracing(): +def test_tracing_in_ray_tasks(): setup_sentry() ray.init( @@ -50,6 +69,7 @@ def test_ray_tracing(): } ) + # Setup ray task @ray.remote def example_task(): with sentry_sdk.start_span(op="task", name="example task step"): @@ -62,63 +82,42 @@ def example_task(): client_envelope = sentry_sdk.get_client().transport.envelopes[0] client_transaction = client_envelope.get_transaction_event() + assert client_transaction["transaction"] == "ray test transaction" + assert client_transaction["transaction_info"] == {"source": "custom"} + worker_envelope = worker_envelopes[0] worker_transaction = worker_envelope.get_transaction_event() - assert ( - client_transaction["contexts"]["trace"]["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] + worker_transaction["transaction"] + == "tests.integrations.ray.test_ray.test_tracing_in_ray_tasks..example_task" ) + assert worker_transaction["transaction_info"] == {"source": "task"} - for span in client_transaction["spans"]: - assert ( - span["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - ) - - for span in worker_transaction["spans"]: - assert ( - span["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - ) - - -@pytest.mark.forked -def test_ray_spans(): - setup_sentry() - - ray.init( - runtime_env={ - "worker_process_setup_hook": setup_sentry, - "working_dir": "./", - } + (span,) = client_transaction["spans"] + assert span["op"] == "queue.submit.ray" + assert span["origin"] == "auto.queue.ray" + assert ( + span["description"] + == "tests.integrations.ray.test_ray.test_tracing_in_ray_tasks..example_task" ) + assert span["parent_span_id"] == client_transaction["contexts"]["trace"]["span_id"] + assert span["trace_id"] == client_transaction["contexts"]["trace"]["trace_id"] - @ray.remote - def example_task(): - return sentry_sdk.get_client().transport.envelopes + (span,) = worker_transaction["spans"] + assert span["op"] == "task" + assert span["origin"] == "manual" + assert span["description"] == "example task step" + assert span["parent_span_id"] == worker_transaction["contexts"]["trace"]["span_id"] + assert span["trace_id"] == worker_transaction["contexts"]["trace"]["trace_id"] - with sentry_sdk.start_transaction(op="task", name="ray test transaction"): - worker_envelopes = ray.get(example_task.remote()) - - client_envelope = sentry_sdk.get_client().transport.envelopes[0] - client_transaction = client_envelope.get_transaction_event() - worker_envelope = worker_envelopes[0] - worker_transaction = worker_envelope.get_transaction_event() - - for span in client_transaction["spans"]: - assert span["op"] == "queue.submit.ray" - assert span["origin"] == "auto.queue.ray" - - for span in worker_transaction["spans"]: - assert span["op"] == "queue.task.ray" - assert span["origin"] == "auto.queue.ray" + assert ( + client_transaction["contexts"]["trace"]["trace_id"] + == worker_transaction["contexts"]["trace"]["trace_id"] + ) @pytest.mark.forked -def test_ray_errors(): +def test_errors_in_ray_tasks(): setup_sentry_with_logging_transport() ray.init( @@ -128,6 +127,7 @@ def test_ray_errors(): } ) + # Setup ray task @ray.remote def example_task(): 1 / 0 @@ -138,30 +138,19 @@ def example_task(): ray.get(future) job_id = future.job_id().hex() - - # Read the worker log output containing the error - log_dir = "/tmp/ray/session_latest/logs/" - log_file = [ - f - for f in os.listdir(log_dir) - if "worker" in f and job_id in f and f.endswith(".out") - ][0] - with open(os.path.join(log_dir, log_file), "r") as file: - lines = file.readlines() - # parse error object from log line - error = json.loads(lines[4][:-1]) + error = read_error_from_log(job_id) assert error["level"] == "error" assert ( error["transaction"] - == "tests.integrations.ray.test_ray.test_ray_errors..example_task" - ) # its in the worker, not the client thus not "ray test transaction" + == "tests.integrations.ray.test_ray.test_errors_in_ray_tasks..example_task" + ) assert error["exception"]["values"][0]["mechanism"]["type"] == "ray" assert not error["exception"]["values"][0]["mechanism"]["handled"] @pytest.mark.forked -def test_ray_actor(): +def test_tracing_in_ray_actors(): setup_sentry() ray.init( @@ -171,13 +160,14 @@ def test_ray_actor(): } ) + # Setup ray actor @ray.remote class Counter: def __init__(self): self.n = 0 def increment(self): - with sentry_sdk.start_span(op="task", name="example task step"): + with sentry_sdk.start_span(op="task", name="example actor execution"): self.n += 1 return sentry_sdk.get_client().transport.envelopes @@ -186,20 +176,47 @@ def increment(self): counter = Counter.remote() worker_envelopes = ray.get(counter.increment.remote()) - # Currently no transactions/spans are captured in actors - assert worker_envelopes == [] - client_envelope = sentry_sdk.get_client().transport.envelopes[0] client_transaction = client_envelope.get_transaction_event() - assert ( - client_transaction["contexts"]["trace"]["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] + # Spans for submitting the actor task are not created (actors are not supported yet) + assert client_transaction["spans"] == [] + + # Transaction are not yet created when executing ray actors (actors are not supported yet) + assert worker_envelopes == [] + + +@pytest.mark.forked +def test_errors_in_ray_actors(): + setup_sentry_with_logging_transport() + + ray.init( + runtime_env={ + "worker_process_setup_hook": setup_sentry_with_logging_transport, + "working_dir": "./", + } ) - for span in client_transaction["spans"]: - assert ( - span["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - == client_transaction["contexts"]["trace"]["trace_id"] - ) + # Setup ray actor + @ray.remote + class Counter: + def __init__(self): + self.n = 0 + + def increment(self): + with sentry_sdk.start_span(op="task", name="example actor execution"): + 1 / 0 + + return sentry_sdk.get_client().transport.envelopes + + with sentry_sdk.start_transaction(op="task", name="ray test transaction"): + with pytest.raises(ZeroDivisionError): + counter = Counter.remote() + future = counter.increment.remote() + ray.get(future) + + job_id = future.job_id().hex() + error = read_error_from_log(job_id) + + # We do not capture errors in ray actors yet + assert error is None From 50ad148803e372bdaea4815884788c28a4897974 Mon Sep 17 00:00:00 2001 From: Florian Dellekart <60044734+fdellekart@users.noreply.github.com> Date: Thu, 5 Dec 2024 12:57:09 +0100 Subject: [PATCH 055/591] =?UTF-8?q?fix(grpc):=20Return=20proper=20metadata?= =?UTF-8?q?=20object=20instead=20of=20list=20in=E2=80=A6=20(#3205)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * fix(grpc): Return propagate proper metadata object instead of list in client interceptor Fixes #2509 * fix(grpc): Transform metadata into Metadata object in case it's a tuple Up until version 1.65.0 of grpcio, the metadata was not guaranteed to arrive as the type specified in annotations but could be a tuple. To support versions before that we check and transform it here. * docs(grpc): Add comment about workaround --------- Co-authored-by: Anton Pirker Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/integrations/grpc/aio/client.py | 23 ++++++++++------------ 1 file changed, 10 insertions(+), 13 deletions(-) diff --git a/sentry_sdk/integrations/grpc/aio/client.py b/sentry_sdk/integrations/grpc/aio/client.py index e8adeba05e..ff3c213176 100644 --- a/sentry_sdk/integrations/grpc/aio/client.py +++ b/sentry_sdk/integrations/grpc/aio/client.py @@ -6,6 +6,7 @@ ClientCallDetails, UnaryUnaryCall, UnaryStreamCall, + Metadata, ) from google.protobuf.message import Message @@ -19,23 +20,19 @@ class ClientInterceptor: def _update_client_call_details_metadata_from_scope( client_call_details: ClientCallDetails, ) -> ClientCallDetails: - metadata = ( - list(client_call_details.metadata) if client_call_details.metadata else [] - ) + if client_call_details.metadata is None: + client_call_details = client_call_details._replace(metadata=Metadata()) + elif not isinstance(client_call_details.metadata, Metadata): + # This is a workaround for a GRPC bug, which was fixed in grpcio v1.60.0 + # See https://github.com/grpc/grpc/issues/34298. + client_call_details = client_call_details._replace( + metadata=Metadata.from_tuple(client_call_details.metadata) + ) for ( key, value, ) in sentry_sdk.get_current_scope().iter_trace_propagation_headers(): - metadata.append((key, value)) - - client_call_details = ClientCallDetails( - method=client_call_details.method, - timeout=client_call_details.timeout, - metadata=metadata, - credentials=client_call_details.credentials, - wait_for_ready=client_call_details.wait_for_ready, - ) - + client_call_details.metadata.add(key, value) return client_call_details From cda51274de6b11c59a496d610907e4656fa99fd7 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 5 Dec 2024 14:29:06 +0100 Subject: [PATCH 056/591] Add missing stack frames (#3673) Add a new `init()` option `add_full_stack` (default `False`), when set to `True` it will add all the missing frames from the beginning of the execution to the stack trace sent to Sentry. Also adds another option `max_stack_frames` (default `100`) to limit the number of frames sent. The limitation is only enforced when `add_full_stack=True` to not change behavior for existing users. Fixes #3646 --- sentry_sdk/consts.py | 5 ++ sentry_sdk/utils.py | 82 +++++++++++++++++++++++-- tests/test_full_stack_frames.py | 103 ++++++++++++++++++++++++++++++++ 3 files changed, 185 insertions(+), 5 deletions(-) create mode 100644 tests/test_full_stack_frames.py diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 488743b579..6750e85f99 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -6,6 +6,9 @@ # up top to prevent circular import due to integration import DEFAULT_MAX_VALUE_LENGTH = 1024 +DEFAULT_MAX_STACK_FRAMES = 100 +DEFAULT_ADD_FULL_STACK = False + # Also needs to be at the top to prevent circular import class EndpointType(Enum): @@ -551,6 +554,8 @@ def __init__( cert_file=None, # type: Optional[str] key_file=None, # type: Optional[str] custom_repr=None, # type: Optional[Callable[..., Optional[str]]] + add_full_stack=DEFAULT_ADD_FULL_STACK, # type: bool + max_stack_frames=DEFAULT_MAX_STACK_FRAMES, # type: Optional[int] ): # type: (...) -> None pass diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 4d07974809..ae6e7538ac 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -26,7 +26,12 @@ import sentry_sdk from sentry_sdk._compat import PY37 -from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH, EndpointType +from sentry_sdk.consts import ( + DEFAULT_ADD_FULL_STACK, + DEFAULT_MAX_STACK_FRAMES, + DEFAULT_MAX_VALUE_LENGTH, + EndpointType, +) from typing import TYPE_CHECKING @@ -737,6 +742,7 @@ def single_exception_from_error_tuple( exception_id=None, # type: Optional[int] parent_id=None, # type: Optional[int] source=None, # type: Optional[str] + full_stack=None, # type: Optional[list[dict[str, Any]]] ): # type: (...) -> Dict[str, Any] """ @@ -804,10 +810,15 @@ def single_exception_from_error_tuple( custom_repr=custom_repr, ) for tb in iter_stacks(tb) - ] + ] # type: List[Dict[str, Any]] if frames: - exception_value["stacktrace"] = {"frames": frames} + if not full_stack: + new_frames = frames + else: + new_frames = merge_stack_frames(frames, full_stack, client_options) + + exception_value["stacktrace"] = {"frames": new_frames} return exception_value @@ -862,6 +873,7 @@ def exceptions_from_error( exception_id=0, # type: int parent_id=0, # type: int source=None, # type: Optional[str] + full_stack=None, # type: Optional[list[dict[str, Any]]] ): # type: (...) -> Tuple[int, List[Dict[str, Any]]] """ @@ -881,6 +893,7 @@ def exceptions_from_error( exception_id=exception_id, parent_id=parent_id, source=source, + full_stack=full_stack, ) exceptions = [parent] @@ -906,6 +919,7 @@ def exceptions_from_error( mechanism=mechanism, exception_id=exception_id, source="__cause__", + full_stack=full_stack, ) exceptions.extend(child_exceptions) @@ -927,6 +941,7 @@ def exceptions_from_error( mechanism=mechanism, exception_id=exception_id, source="__context__", + full_stack=full_stack, ) exceptions.extend(child_exceptions) @@ -943,6 +958,7 @@ def exceptions_from_error( exception_id=exception_id, parent_id=parent_id, source="exceptions[%s]" % idx, + full_stack=full_stack, ) exceptions.extend(child_exceptions) @@ -953,6 +969,7 @@ def exceptions_from_error_tuple( exc_info, # type: ExcInfo client_options=None, # type: Optional[Dict[str, Any]] mechanism=None, # type: Optional[Dict[str, Any]] + full_stack=None, # type: Optional[list[dict[str, Any]]] ): # type: (...) -> List[Dict[str, Any]] exc_type, exc_value, tb = exc_info @@ -970,6 +987,7 @@ def exceptions_from_error_tuple( mechanism=mechanism, exception_id=0, parent_id=0, + full_stack=full_stack, ) else: @@ -977,7 +995,12 @@ def exceptions_from_error_tuple( for exc_type, exc_value, tb in walk_exception_chain(exc_info): exceptions.append( single_exception_from_error_tuple( - exc_type, exc_value, tb, client_options, mechanism + exc_type=exc_type, + exc_value=exc_value, + tb=tb, + client_options=client_options, + mechanism=mechanism, + full_stack=full_stack, ) ) @@ -1096,6 +1119,46 @@ def exc_info_from_error(error): return exc_info +def merge_stack_frames(frames, full_stack, client_options): + # type: (List[Dict[str, Any]], List[Dict[str, Any]], Optional[Dict[str, Any]]) -> List[Dict[str, Any]] + """ + Add the missing frames from full_stack to frames and return the merged list. + """ + frame_ids = { + ( + frame["abs_path"], + frame["context_line"], + frame["lineno"], + frame["function"], + ) + for frame in frames + } + + new_frames = [ + stackframe + for stackframe in full_stack + if ( + stackframe["abs_path"], + stackframe["context_line"], + stackframe["lineno"], + stackframe["function"], + ) + not in frame_ids + ] + new_frames.extend(frames) + + # Limit the number of frames + max_stack_frames = ( + client_options.get("max_stack_frames", DEFAULT_MAX_STACK_FRAMES) + if client_options + else None + ) + if max_stack_frames is not None: + new_frames = new_frames[len(new_frames) - max_stack_frames :] + + return new_frames + + def event_from_exception( exc_info, # type: Union[BaseException, ExcInfo] client_options=None, # type: Optional[Dict[str, Any]] @@ -1104,12 +1167,21 @@ def event_from_exception( # type: (...) -> Tuple[Event, Dict[str, Any]] exc_info = exc_info_from_error(exc_info) hint = event_hint_with_exc_info(exc_info) + + if client_options and client_options.get("add_full_stack", DEFAULT_ADD_FULL_STACK): + full_stack = current_stacktrace( + include_local_variables=client_options["include_local_variables"], + max_value_length=client_options["max_value_length"], + )["frames"] + else: + full_stack = None + return ( { "level": "error", "exception": { "values": exceptions_from_error_tuple( - exc_info, client_options, mechanism + exc_info, client_options, mechanism, full_stack ) }, }, diff --git a/tests/test_full_stack_frames.py b/tests/test_full_stack_frames.py new file mode 100644 index 0000000000..ad0826cd10 --- /dev/null +++ b/tests/test_full_stack_frames.py @@ -0,0 +1,103 @@ +import sentry_sdk + + +def test_full_stack_frames_default(sentry_init, capture_events): + sentry_init() + events = capture_events() + + def foo(): + try: + bar() + except Exception as e: + sentry_sdk.capture_exception(e) + + def bar(): + raise Exception("This is a test exception") + + foo() + + (event,) = events + frames = event["exception"]["values"][0]["stacktrace"]["frames"] + + assert len(frames) == 2 + assert frames[-1]["function"] == "bar" + assert frames[-2]["function"] == "foo" + + +def test_full_stack_frames_enabled(sentry_init, capture_events): + sentry_init( + add_full_stack=True, + ) + events = capture_events() + + def foo(): + try: + bar() + except Exception as e: + sentry_sdk.capture_exception(e) + + def bar(): + raise Exception("This is a test exception") + + foo() + + (event,) = events + frames = event["exception"]["values"][0]["stacktrace"]["frames"] + + assert len(frames) > 2 + assert frames[-1]["function"] == "bar" + assert frames[-2]["function"] == "foo" + assert frames[-3]["function"] == "foo" + assert frames[-4]["function"] == "test_full_stack_frames_enabled" + + +def test_full_stack_frames_enabled_truncated(sentry_init, capture_events): + sentry_init( + add_full_stack=True, + max_stack_frames=3, + ) + events = capture_events() + + def foo(): + try: + bar() + except Exception as e: + sentry_sdk.capture_exception(e) + + def bar(): + raise Exception("This is a test exception") + + foo() + + (event,) = events + frames = event["exception"]["values"][0]["stacktrace"]["frames"] + + assert len(frames) == 3 + assert frames[-1]["function"] == "bar" + assert frames[-2]["function"] == "foo" + assert frames[-3]["function"] == "foo" + + +def test_full_stack_frames_default_no_truncation_happening(sentry_init, capture_events): + sentry_init( + max_stack_frames=1, # this is ignored if add_full_stack=False (which is the default) + ) + events = capture_events() + + def foo(): + try: + bar() + except Exception as e: + sentry_sdk.capture_exception(e) + + def bar(): + raise Exception("This is a test exception") + + foo() + + (event,) = events + frames = event["exception"]["values"][0]["stacktrace"]["frames"] + + assert len(frames) == 2 + assert frames[-1]["function"] == "bar" + assert frames[-2]["function"] == "foo" From 5891717b1470f0aa29193a9eb6cf0d899f8ba776 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 5 Dec 2024 14:29:42 +0100 Subject: [PATCH 057/591] Script for checking if our instrumented libs are python 3.13 compatible (#3425) A simple script that parses all libraries we test against from our `tox.ini` and then checks PyPI if this library already supports the newest Python version (currently 3.13) --- scripts/ready_yet/main.py | 124 +++++++++++++++++++++++++++++ scripts/ready_yet/requirements.txt | 3 + scripts/ready_yet/run.sh | 16 ++++ 3 files changed, 143 insertions(+) create mode 100644 scripts/ready_yet/main.py create mode 100644 scripts/ready_yet/requirements.txt create mode 100755 scripts/ready_yet/run.sh diff --git a/scripts/ready_yet/main.py b/scripts/ready_yet/main.py new file mode 100644 index 0000000000..bba97d0c98 --- /dev/null +++ b/scripts/ready_yet/main.py @@ -0,0 +1,124 @@ +import time +import re +import sys + +import requests + +from collections import defaultdict + +from pathlib import Path + +from tox.config.cli.parse import get_options +from tox.session.state import State +from tox.config.sets import CoreConfigSet +from tox.config.source.tox_ini import ToxIni + +PYTHON_VERSION = "3.13" + +MATCH_LIB_SENTRY_REGEX = r"py[\d\.]*-(.*)-.*" + +PYPI_PROJECT_URL = "https://pypi.python.org/pypi/%7Bproject%7D/json" +PYPI_VERSION_URL = "https://pypi.python.org/pypi/%7Bproject%7D/%7Bversion%7D/json" + + +def get_tox_envs(tox_ini_path: Path) -> list: + tox_ini = ToxIni(tox_ini_path) + conf = State(get_options(), []).conf + tox_section = next(tox_ini.sections()) + core_config_set = CoreConfigSet( + conf, tox_section, tox_ini_path.parent, tox_ini_path + ) + ( + core_config_set.loaders.extend( + tox_ini.get_loaders( + tox_section, + base=[], + override_map=defaultdict(list, {}), + conf=core_config_set, + ) + ) + ) + return core_config_set.load("env_list") + + +def get_libs(tox_ini: Path, regex: str) -> set: + libs = set() + for env in get_tox_envs(tox_ini): + match = re.match(regex, env) + if match: + libs.add(match.group(1)) + + return sorted(libs) + + +def main(): + """ + Check if libraries in our tox.ini are ready for Python version defined in `PYTHON_VERSION`. + """ + print(f"Checking libs from tox.ini for Python {PYTHON_VERSION} compatibility:") + + ready = set() + not_ready = set() + not_found = set() + + tox_ini = Path(__file__).parent.parent.parent.joinpath("tox.ini") + + libs = get_libs(tox_ini, MATCH_LIB_SENTRY_REGEX) + + for lib in libs: + print(".", end="") + sys.stdout.flush() + + # Get latest version of lib + url = PYPI_PROJECT_URL.format(project=lib) + pypi_data = requests.get(url) + + if pypi_data.status_code != 200: + not_found.add(lib) + continue + + latest_version = pypi_data.json()["info"]["version"] + + # Get supported Python version of latest version of lib + url = PYPI_PROJECT_URL.format(project=lib, version=latest_version) + pypi_data = requests.get(url) + + if pypi_data.status_code != 200: + continue + + classifiers = pypi_data.json()["info"]["classifiers"] + + if f"Programming Language :: Python :: {PYTHON_VERSION}" in classifiers: + ready.add(lib) + else: + not_ready.add(lib) + + # cut pypi some slack + time.sleep(0.1) + + # Print report + print("\n") + print(f"\nReady for Python {PYTHON_VERSION}:") + if len(ready) == 0: + print("- None ") + + for x in sorted(ready): + print(f"- {x}") + + print(f"\nNOT ready for Python {PYTHON_VERSION}:") + if len(not_ready) == 0: + print("- None ") + + for x in sorted(not_ready): + print(f"- {x}") + + print("\nNot found on PyPI:") + if len(not_found) == 0: + print("- None ") + + for x in sorted(not_found): + print(f"- {x}") + + +if __name__ == "__main__": + main() diff --git a/scripts/ready_yet/requirements.txt b/scripts/ready_yet/requirements.txt new file mode 100644 index 0000000000..e0590b89c6 --- /dev/null +++ b/scripts/ready_yet/requirements.txt @@ -0,0 +1,3 @@ +requests +pathlib +tox \ No newline at end of file diff --git a/scripts/ready_yet/run.sh b/scripts/ready_yet/run.sh new file mode 100755 index 0000000000..f32bd7bdda --- /dev/null +++ b/scripts/ready_yet/run.sh @@ -0,0 +1,16 @@ +#!/usr/bin/env bash + +# exit on first error +set -xe + +reset + +# create and activate virtual environment +python -m venv .venv +source .venv/bin/activate + +# Install (or update) requirements +python -m pip install -r requirements.txt + +# Run the script +python main.py \ No newline at end of file From 31fdcfaee7e871802f8ffef72847884e28472969 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Thu, 5 Dec 2024 13:58:22 +0000 Subject: [PATCH 058/591] fix(django): Fix errors when instrumenting Django cache (#3855) I was testing Spotlight with Sentry and realized things started to get slow and crashy. It looks like sometimes `args` is just an empty array on cache's `_instruments_call` causing lots of exceptions being thrown. This patch fixes that with explicit length checks and also adds a note for the missing instrumentation for `get_or_set` method. This might be related to #2122 and #3300. --- sentry_sdk/integrations/django/caching.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/integrations/django/caching.py b/sentry_sdk/integrations/django/caching.py index 39d1679183..7985611761 100644 --- a/sentry_sdk/integrations/django/caching.py +++ b/sentry_sdk/integrations/django/caching.py @@ -75,11 +75,12 @@ def _instrument_call( span.set_data(SPANDATA.CACHE_HIT, True) else: span.set_data(SPANDATA.CACHE_HIT, False) - else: - try: + else: # TODO: We don't handle `get_or_set` which we should + arg_count = len(args) + if arg_count >= 2: # 'set' command item_size = len(str(args[1])) - except IndexError: + elif arg_count == 1: # 'set_many' command item_size = len(str(args[0])) From 5a097705411842c48358b5a797fd92723a853019 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Thu, 5 Dec 2024 14:06:41 +0000 Subject: [PATCH 059/591] fix(spotlight): Don't give up on Spotlight on 3 errors (#3856) Current Spotlight error handling logic gives up sending events to Spotlight after 3 errors. This doesn't make much sense because: 1. Since there is no back off or retry mechanism, even a very brief server hiccup or restart turns off Spotlight reporting 2. Once this shut off kicks in, there is no way to turn it back on except for a server restart I added a note for future work for retries and some short buffer. --- sentry_sdk/spotlight.py | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index 806ba5a09e..a94c691723 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -42,11 +42,6 @@ def __init__(self, url): def capture_envelope(self, envelope): # type: (Envelope) -> None - if self.tries > 3: - sentry_logger.warning( - "Too many errors sending to Spotlight, stop sending events there." - ) - return body = io.BytesIO() envelope.serialize_into(body) try: @@ -60,7 +55,7 @@ def capture_envelope(self, envelope): ) req.close() except Exception as e: - self.tries += 1 + # TODO: Implement buffering and retrying with exponential backoff sentry_logger.warning(str(e)) From 7a6d460bd14433c3d3f03efa6a4b3f924105adc6 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 5 Dec 2024 15:49:17 +0100 Subject: [PATCH 060/591] Copy scope.client reference as well (#3857) --- sentry_sdk/scope.py | 1 + tests/test_scope.py | 6 +----- 2 files changed, 2 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 34ccc7f940..bb45143c48 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -225,6 +225,7 @@ def __copy__(self): rv = object.__new__(self.__class__) # type: Scope rv._type = self._type + rv.client = self.client rv._level = self._level rv._name = self._name rv._fingerprint = self._fingerprint diff --git a/tests/test_scope.py b/tests/test_scope.py index 374a354446..a03eb07a99 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -19,10 +19,6 @@ ) -SLOTS_NOT_COPIED = {"client"} -"""__slots__ that are not copied when copying a Scope object.""" - - def test_copying(): s1 = Scope() s1.fingerprint = {} @@ -43,7 +39,7 @@ def test_all_slots_copied(): scope_copy = copy.copy(scope) # Check all attributes are copied - for attr in set(Scope.__slots__) - SLOTS_NOT_COPIED: + for attr in set(Scope.__slots__): assert getattr(scope_copy, attr) == getattr(scope, attr) From c591b64d5075628d5fa5351ed4307182981e9bd5 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 5 Dec 2024 14:51:42 +0000 Subject: [PATCH 061/591] release: 2.19.1 --- CHANGELOG.md | 20 ++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 23 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index dbb35eb1eb..d1d0a78ce8 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,25 @@ # Changelog +## 2.19.1 + +### Various fixes & improvements + +- Copy scope.client reference as well (#3857) by @sl0thentr0py +- fix(spotlight): Don't give up on Spotlight on 3 errors (#3856) by @BYK +- fix(django): Fix errors when instrumenting Django cache (#3855) by @BYK +- Script for checking if our instrumented libs are python 3.13 compatible (#3425) by @antonpirker +- Add missing stack frames (#3673) by @antonpirker +- fix(grpc): Return proper metadata object instead of list in… (#3205) by @fdellekart +- Improve ray tests (#3846) by @antonpirker +- Test with celery 5.5.0rc3 (#3842) by @sentrivana +- Revert "Fix spans for streaming responses in WSGI based frameworks (#3798)" (#3836) by @antonpirker +- Fix asyncio testing setup (#3832) by @sl0thentr0py +- build(deps): bump codecov/codecov-action from 5.0.2 to 5.0.7 (#3821) by @dependabot +- Fix CI (#3834) by @sentrivana +- ref(flags): rename launch darkly hook to match JS SDK (#3743) by @aliu39 +- Use new clickhouse gh action (#3826) by @antonpirker +- Fix spans for streaming responses in WSGI based frameworks (#3798) by @antonpirker + ## 2.19.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 55d5295381..4f5c210322 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.19.0" +release = "2.19.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 6750e85f99..f338543dee 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -581,4 +581,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.19.0" +VERSION = "2.19.1" diff --git a/setup.py b/setup.py index fda3daa229..7782d57a36 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.19.0", + version="2.19.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 231a6a1d5eb5026415542ef2c2355e468bc69f66 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 5 Dec 2024 15:53:50 +0100 Subject: [PATCH 062/591] Update CHANGELOG.md --- CHANGELOG.md | 22 ++++++++++------------ 1 file changed, 10 insertions(+), 12 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index d1d0a78ce8..eb45f28c7e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,21 +4,19 @@ ### Various fixes & improvements -- Copy scope.client reference as well (#3857) by @sl0thentr0py -- fix(spotlight): Don't give up on Spotlight on 3 errors (#3856) by @BYK -- fix(django): Fix errors when instrumenting Django cache (#3855) by @BYK -- Script for checking if our instrumented libs are python 3.13 compatible (#3425) by @antonpirker +- Fix errors when instrumenting Django cache (#3855) by @BYK +- Copy `scope.client` reference as well (#3857) by @sl0thentr0py +- Don't give up on Spotlight on 3 errors (#3856) by @BYK - Add missing stack frames (#3673) by @antonpirker -- fix(grpc): Return proper metadata object instead of list in… (#3205) by @fdellekart -- Improve ray tests (#3846) by @antonpirker -- Test with celery 5.5.0rc3 (#3842) by @sentrivana -- Revert "Fix spans for streaming responses in WSGI based frameworks (#3798)" (#3836) by @antonpirker +- Fix wrong metadata type in async gRPC interceptor (#3205) by @fdellekart +- Rename launch darkly hook to match JS SDK (#3743) by @aliu39 +- Script for checking if our instrumented libs are Python 3.13 compatible (#3425) by @antonpirker +- Improve Ray tests (#3846) by @antonpirker +- Test with Celery `5.5.0rc3` (#3842) by @sentrivana - Fix asyncio testing setup (#3832) by @sl0thentr0py -- build(deps): bump codecov/codecov-action from 5.0.2 to 5.0.7 (#3821) by @dependabot +- Bump `codecov/codecov-action` from `5.0.2` to `5.0.7` (#3821) by @dependabot - Fix CI (#3834) by @sentrivana -- ref(flags): rename launch darkly hook to match JS SDK (#3743) by @aliu39 -- Use new clickhouse gh action (#3826) by @antonpirker -- Fix spans for streaming responses in WSGI based frameworks (#3798) by @antonpirker +- Use new ClickHouse GH action (#3826) by @antonpirker ## 2.19.0 From 7ab7fe67496fce2396edcb5bc8a64645601a1218 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 5 Dec 2024 16:16:49 +0100 Subject: [PATCH 063/591] Cleanup chalice test environment (#3858) --- tox.ini | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/tox.ini b/tox.ini index 8c6f9eda86..d3bd83cb03 100644 --- a/tox.ini +++ b/tox.ini @@ -391,11 +391,9 @@ deps = {py3.7}-celery: importlib-metadata<5.0 # Chalice + chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 chalice-latest: chalice - chalice: pytest-chalice==0.0.5 - - {py3.7,py3.8}-chalice: botocore~=1.31 # Clickhouse Driver clickhouse_driver-v0.2.0: clickhouse_driver~=0.2.0 From 8f9461e1a0bc497e6333b4d955561a904beb9dae Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Fri, 6 Dec 2024 02:11:03 -0600 Subject: [PATCH 064/591] Deepcopy and ensure get_all function always terminates (#3861) @aliu39 discovered that under certain circumstances a process can get stuck in an infinite loop. Andrew fixed this by using `deepcopy` which prevents the infinite loop and fixes a bug where the LRU returns incorrect results. Additionally, I've added a terminating loop in case there are any future bugs we've missed. Closes: https://github.com/getsentry/sentry-python/issues/3862 Out of precaution, we disabled flagpole evaluation tracking Sentry while we wait for this to be merged. --- sentry_sdk/_lru_cache.py | 14 +++++++++++--- tests/test_lru_cache.py | 18 ++++++++++++++++++ 2 files changed, 29 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/_lru_cache.py b/sentry_sdk/_lru_cache.py index ec557b1093..825c773529 100644 --- a/sentry_sdk/_lru_cache.py +++ b/sentry_sdk/_lru_cache.py @@ -62,7 +62,7 @@ """ -from copy import copy +from copy import copy, deepcopy SENTINEL = object() @@ -95,7 +95,7 @@ def __copy__(self): cache = LRUCache(self.max_size) cache.full = self.full cache.cache = copy(self.cache) - cache.root = copy(self.root) + cache.root = deepcopy(self.root) return cache def set(self, key, value): @@ -167,7 +167,15 @@ def get(self, key, default=None): def get_all(self): nodes = [] node = self.root[NEXT] - while node is not self.root: + + # To ensure the loop always terminates we iterate to the maximum + # size of the LRU cache. + for _ in range(self.max_size): + # The cache may not be full. We exit early if we've wrapped + # around to the head. + if node is self.root: + break nodes.append((node[KEY], node[VALUE])) node = node[NEXT] + return nodes diff --git a/tests/test_lru_cache.py b/tests/test_lru_cache.py index 3e9c0ac964..cab9bbc7eb 100644 --- a/tests/test_lru_cache.py +++ b/tests/test_lru_cache.py @@ -1,4 +1,5 @@ import pytest +from copy import copy from sentry_sdk._lru_cache import LRUCache @@ -58,3 +59,20 @@ def test_cache_get_all(): assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] cache.get(1) assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] + + +def test_cache_copy(): + cache = LRUCache(3) + cache.set(0, 0) + cache.set(1, 1) + + copied = copy(cache) + cache.set(2, 2) + cache.set(3, 3) + assert copied.get_all() == [(0, 0), (1, 1)] + assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] + + copied = copy(cache) + cache.get(1) + assert copied.get_all() == [(1, 1), (2, 2), (3, 3)] + assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] From 163762f107710cdd1c36040a54806418f3ec4c8c Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Fri, 6 Dec 2024 08:12:00 +0000 Subject: [PATCH 065/591] release: 2.19.2 --- CHANGELOG.md | 7 +++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 10 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index eb45f28c7e..af4eb04fef 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,12 @@ # Changelog +## 2.19.2 + +### Various fixes & improvements + +- Deepcopy and ensure get_all function always terminates (#3861) by @cmanallen +- Cleanup chalice test environment (#3858) by @antonpirker + ## 2.19.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 4f5c210322..3ecdbe2e68 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.19.1" +release = "2.19.2" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index f338543dee..0bb71cb98d 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -581,4 +581,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.19.1" +VERSION = "2.19.2" diff --git a/setup.py b/setup.py index 7782d57a36..da3adcab42 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.19.1", + version="2.19.2", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 26479b22d51cc9544e4c1bf515fc8590f83589bc Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 6 Dec 2024 10:04:31 +0100 Subject: [PATCH 066/591] Use stdlib pathlib in ready-yet script (#3863) --- scripts/ready_yet/requirements.txt | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/scripts/ready_yet/requirements.txt b/scripts/ready_yet/requirements.txt index e0590b89c6..69f9472fa5 100644 --- a/scripts/ready_yet/requirements.txt +++ b/scripts/ready_yet/requirements.txt @@ -1,3 +1,2 @@ requests -pathlib -tox \ No newline at end of file +tox From 6448c709b840f37ca40b297fd64a99467f05d39b Mon Sep 17 00:00:00 2001 From: Jeffrey Hung <17494876+Jeffreyhung@users.noreply.github.com> Date: Wed, 11 Dec 2024 04:05:57 -0800 Subject: [PATCH 067/591] Replace release bot with GH app (#3868) --- .github/workflows/release.yml | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 268f62c4cc..2cd3dfb2ac 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -18,14 +18,20 @@ jobs: runs-on: ubuntu-latest name: "Release a new version" steps: + - name: Get auth token + id: token + uses: actions/create-github-app-token@5d869da34e18e7287c1daad50e0b8ea0f506ce69 # v1.11.0 + with: + app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} + private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} - uses: actions/checkout@v4.2.2 with: - token: ${{ secrets.GH_RELEASE_PAT }} + token: ${{ steps.token.outputs.token }} fetch-depth: 0 - name: Prepare release uses: getsentry/action-prepare-release@v1 env: - GITHUB_TOKEN: ${{ secrets.GH_RELEASE_PAT }} + GITHUB_TOKEN: ${{ steps.token.outputs.token }} with: version: ${{ github.event.inputs.version }} force: ${{ github.event.inputs.force }} From 1239499b5d6274f997a890650a516f6c5538a188 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Fri, 13 Dec 2024 11:26:43 +0000 Subject: [PATCH 068/591] fix(spotlight): Make Django middleware init even more defensive (#3870) I just got faced with a situation where even trying to do `settings.DEBUG` may trigger a Django exception if the settings are not loaded yet, hence widening the `capture_internal_exceptions()` scope for this. --- sentry_sdk/spotlight.py | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index a94c691723..1555afb829 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -210,13 +210,13 @@ def setup_spotlight(options): if not isinstance(url, str): return None - if ( - settings is not None - and settings.DEBUG - and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_ON_ERROR", "1")) - and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_MIDDLEWARE", "1")) - ): - with capture_internal_exceptions(): + with capture_internal_exceptions(): + if ( + settings is not None + and settings.DEBUG + and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_ON_ERROR", "1")) + and env_to_bool(os.environ.get("SENTRY_SPOTLIGHT_MIDDLEWARE", "1")) + ): middleware = settings.MIDDLEWARE if DJANGO_SPOTLIGHT_MIDDLEWARE_PATH not in middleware: settings.MIDDLEWARE = type(middleware)( From 81b806321fed9715d0c7ff227bdf22c9f1178ce9 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Sat, 14 Dec 2024 00:55:25 +0000 Subject: [PATCH 069/591] fix(spotlight): Use the spotlight_url passed into the SDK when loading Spotlight (#3871) When we inject spotlight, we don't set the correct sidecar URL. This is an issue when a user defines a custom sidecar URL where we are able to load Spotlight UI from the correct URL but don't tell it the correct sidecar URL, making it non-functional. --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/spotlight.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index 1555afb829..a783b155a1 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -66,7 +66,8 @@ def capture_envelope(self, envelope): SPOTLIGHT_JS_ENTRY_PATH = "/assets/main.js" SPOTLIGHT_JS_SNIPPET_PATTERN = ( - '' + "\n" + '\n' ) SPOTLIGHT_ERROR_PAGE_SNIPPET = ( '\n' @@ -113,7 +114,8 @@ def spotlight_script(self): ) urllib.request.urlopen(req) self._spotlight_script = SPOTLIGHT_JS_SNIPPET_PATTERN.format( - spotlight_js_url + spotlight_url=self._spotlight_url, + spotlight_js_url=spotlight_js_url, ) except urllib.error.URLError as err: sentry_logger.debug( From 2666022f490dfe3f94db80059535818b37e76839 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 17 Dec 2024 15:33:04 +0100 Subject: [PATCH 070/591] Fix CI (#3878) --- tox.ini | 1 + 1 file changed, 1 insertion(+) diff --git a/tox.ini b/tox.ini index d3bd83cb03..9ccc4dc0eb 100644 --- a/tox.ini +++ b/tox.ini @@ -603,6 +603,7 @@ deps = quart-v0.16: quart~=0.16.0 quart-v0.19: Werkzeug>=3.0.0 quart-v0.19: quart~=0.19.0 + {py3.8}-quart: taskgroup==0.0.0a4 quart-latest: quart # Ray From 4e69cb7f56880ba5f1a0041c80cdf2b773ed7deb Mon Sep 17 00:00:00 2001 From: Patrick Arminio Date: Wed, 18 Dec 2024 10:52:05 +0000 Subject: [PATCH 071/591] =?UTF-8?q?=E2=9C=A8=20Add=20Typer=20integration?= =?UTF-8?q?=20(#3869)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-misc.yml | 10 +++- requirements-linting.txt | 1 + .../split-tox-gh-actions.py | 1 + sentry_sdk/integrations/typer.py | 60 +++++++++++++++++++ tests/integrations/typer/__init__.py | 3 + tests/integrations/typer/test_typer.py | 52 ++++++++++++++++ tox.ini | 9 +++ 7 files changed, 135 insertions(+), 1 deletion(-) create mode 100644 sentry_sdk/integrations/typer.py create mode 100644 tests/integrations/typer/__init__.py create mode 100644 tests/integrations/typer/test_typer.py diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index fb76a854fb..b88b256384 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -27,7 +27,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.8","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -73,6 +73,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-trytond-latest" + - name: Test typer latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-typer-latest" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -153,6 +157,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-trytond" + - name: Test typer pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-typer" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/requirements-linting.txt b/requirements-linting.txt index c9d4bd7f5c..c3f39ecd1f 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -17,3 +17,4 @@ pre-commit # local linting httpcore openfeature-sdk launchdarkly-server-sdk +typer diff --git a/scripts/split-tox-gh-actions/split-tox-gh-actions.py b/scripts/split-tox-gh-actions/split-tox-gh-actions.py index c4b8f3e5e5..26d13390c2 100755 --- a/scripts/split-tox-gh-actions/split-tox-gh-actions.py +++ b/scripts/split-tox-gh-actions/split-tox-gh-actions.py @@ -132,6 +132,7 @@ "potel", "pure_eval", "trytond", + "typer", ], } diff --git a/sentry_sdk/integrations/typer.py b/sentry_sdk/integrations/typer.py new file mode 100644 index 0000000000..8879d6d0d0 --- /dev/null +++ b/sentry_sdk/integrations/typer.py @@ -0,0 +1,60 @@ +import sentry_sdk +from sentry_sdk.utils import ( + capture_internal_exceptions, + event_from_exception, +) +from sentry_sdk.integrations import Integration, DidNotEnable + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Callable + from typing import Any + from typing import Type + from typing import Optional + + from types import TracebackType + + Excepthook = Callable[ + [Type[BaseException], BaseException, Optional[TracebackType]], + Any, + ] + +try: + import typer +except ImportError: + raise DidNotEnable("Typer not installed") + + +class TyperIntegration(Integration): + identifier = "typer" + + @staticmethod + def setup_once(): + # type: () -> None + typer.main.except_hook = _make_excepthook(typer.main.except_hook) # type: ignore + + +def _make_excepthook(old_excepthook): + # type: (Excepthook) -> Excepthook + def sentry_sdk_excepthook(type_, value, traceback): + # type: (Type[BaseException], BaseException, Optional[TracebackType]) -> None + integration = sentry_sdk.get_client().get_integration(TyperIntegration) + + # Note: If we replace this with ensure_integration_enabled then + # we break the exceptiongroup backport; + # See: https://github.com/getsentry/sentry-python/issues/3097 + if integration is None: + return old_excepthook(type_, value, traceback) + + with capture_internal_exceptions(): + event, hint = event_from_exception( + (type_, value, traceback), + client_options=sentry_sdk.get_client().options, + mechanism={"type": "typer", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + return old_excepthook(type_, value, traceback) + + return sentry_sdk_excepthook diff --git a/tests/integrations/typer/__init__.py b/tests/integrations/typer/__init__.py new file mode 100644 index 0000000000..3b7c8011ea --- /dev/null +++ b/tests/integrations/typer/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("typer") diff --git a/tests/integrations/typer/test_typer.py b/tests/integrations/typer/test_typer.py new file mode 100644 index 0000000000..34ac0a7c8c --- /dev/null +++ b/tests/integrations/typer/test_typer.py @@ -0,0 +1,52 @@ +import subprocess +import sys +from textwrap import dedent +import pytest + +from typer.testing import CliRunner + +runner = CliRunner() + + +def test_catch_exceptions(tmpdir): + app = tmpdir.join("app.py") + + app.write( + dedent( + """ + import typer + from unittest import mock + + from sentry_sdk import init, transport + from sentry_sdk.integrations.typer import TyperIntegration + + def capture_envelope(self, envelope): + print("capture_envelope was called") + event = envelope.get_event() + if event is not None: + print(event) + + transport.HttpTransport.capture_envelope = capture_envelope + + init("http://foobar@localhost/123", integrations=[TyperIntegration()]) + + app = typer.Typer() + + @app.command() + def test(): + print("test called") + raise Exception("pollo") + + app() + """ + ) + ) + + with pytest.raises(subprocess.CalledProcessError) as excinfo: + subprocess.check_output([sys.executable, str(app)], stderr=subprocess.STDOUT) + + output = excinfo.value.output + + assert b"capture_envelope was called" in output + assert b"test called" in output + assert b"pollo" in output diff --git a/tox.ini b/tox.ini index 9ccc4dc0eb..717ea62141 100644 --- a/tox.ini +++ b/tox.ini @@ -287,6 +287,10 @@ envlist = {py3.8,py3.11,py3.12}-trytond-v{7} {py3.8,py3.12,py3.13}-trytond-latest + # Typer + {py3.7,py3.12,py3.13}-typer-v{0.15} + {py3.7,py3.12,py3.13}-typer-latest + [testenv] deps = # if you change requirements-testing.txt and your change is not being reflected @@ -724,6 +728,10 @@ deps = trytond-v7: trytond~=7.0 trytond-latest: trytond + # Typer + typer-v0.15: typer~=0.15.0 + typer-latest: typer + setenv = PYTHONDONTWRITEBYTECODE=1 OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES @@ -786,6 +794,7 @@ setenv = strawberry: TESTPATH=tests/integrations/strawberry tornado: TESTPATH=tests/integrations/tornado trytond: TESTPATH=tests/integrations/trytond + typer: TESTPATH=tests/integrations/typer socket: TESTPATH=tests/integrations/socket passenv = From 50222ca2a6c680bb0e712b3bc8a1813d83fa55a0 Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Wed, 18 Dec 2024 20:33:36 -0800 Subject: [PATCH 072/591] feat(flags): Add integration for custom tracking of flag evaluations (#3860) * Add new integration and unit tests * Test flag values for LD and OF threaded/asyncio, not just flag names * update ffIntegration test to be e2e, and fix LRU copy bug * make a helper fixture and test error processor in original thread * Move api to top-level, rename to add_flag * Add docstrs * Rename to add_feature_flag * Rm extra import in test_lru_cache * Revert lru comment * Type annotate * Review comments * Update launchdarkly and openfeature tests to be e2e * Update docstrs * Skip threading test for <3.7 * Skip ffs asyncio test if 3.6 * undo 'skip threading test' * Try commenting out asyncio * Use importorskip * Import order --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/featureflags.py | 44 ++++++ tests/conftest.py | 11 ++ tests/integrations/featureflags/__init__.py | 0 .../featureflags/test_featureflags.py | 133 ++++++++++++++++++ .../launchdarkly/test_launchdarkly.py | 119 +++++++++++++--- .../openfeature/test_openfeature.py | 113 ++++++++++++--- 6 files changed, 377 insertions(+), 43 deletions(-) create mode 100644 sentry_sdk/integrations/featureflags.py create mode 100644 tests/integrations/featureflags/__init__.py create mode 100644 tests/integrations/featureflags/test_featureflags.py diff --git a/sentry_sdk/integrations/featureflags.py b/sentry_sdk/integrations/featureflags.py new file mode 100644 index 0000000000..46947eec72 --- /dev/null +++ b/sentry_sdk/integrations/featureflags.py @@ -0,0 +1,44 @@ +from sentry_sdk.flag_utils import flag_error_processor + +import sentry_sdk +from sentry_sdk.integrations import Integration + + +class FeatureFlagsIntegration(Integration): + """ + Sentry integration for capturing feature flags on error events. To manually buffer flag data, + call `integrations.featureflags.add_feature_flag`. We recommend you do this on each flag + evaluation. + + See the [feature flag documentation](https://develop.sentry.dev/sdk/expected-features/#feature-flags) + for more information. + + @example + ``` + import sentry_sdk + from sentry_sdk.integrations.featureflags import FeatureFlagsIntegration, add_feature_flag + + sentry_sdk.init(dsn="my_dsn", integrations=[FeatureFlagsIntegration()]); + + add_feature_flag('my-flag', true); + sentry_sdk.capture_exception(Exception('broke')); // 'my-flag' should be captured on this Sentry event. + ``` + """ + + identifier = "featureflags" + + @staticmethod + def setup_once(): + # type: () -> None + scope = sentry_sdk.get_current_scope() + scope.add_error_processor(flag_error_processor) + + +def add_feature_flag(flag, result): + # type: (str, bool) -> None + """ + Records a flag and its value to be sent on subsequent error events by FeatureFlagsIntegration. + We recommend you do this on flag evaluations. Flags are buffered per Sentry scope. + """ + flags = sentry_sdk.get_current_scope().flags + flags.set(flag, result) diff --git a/tests/conftest.py b/tests/conftest.py index 64527c1e36..c0383d94b7 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -184,6 +184,17 @@ def reset_integrations(): _installed_integrations.clear() +@pytest.fixture +def uninstall_integration(): + """Use to force the next call to sentry_init to re-install/setup an integration.""" + + def inner(identifier): + _processed_integrations.discard(identifier) + _installed_integrations.discard(identifier) + + return inner + + @pytest.fixture def sentry_init(request): def inner(*a, **kw): diff --git a/tests/integrations/featureflags/__init__.py b/tests/integrations/featureflags/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/featureflags/test_featureflags.py b/tests/integrations/featureflags/test_featureflags.py new file mode 100644 index 0000000000..539e910607 --- /dev/null +++ b/tests/integrations/featureflags/test_featureflags.py @@ -0,0 +1,133 @@ +import concurrent.futures as cf +import sys + +import pytest + +import sentry_sdk +from sentry_sdk.integrations.featureflags import ( + FeatureFlagsIntegration, + add_feature_flag, +) + + +def test_featureflags_integration(sentry_init, capture_events, uninstall_integration): + uninstall_integration(FeatureFlagsIntegration.identifier) + sentry_init(integrations=[FeatureFlagsIntegration()]) + + add_feature_flag("hello", False) + add_feature_flag("world", True) + add_feature_flag("other", False) + + events = capture_events() + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "world", "result": True}, + {"flag": "other", "result": False}, + ] + } + + +def test_featureflags_integration_threaded( + sentry_init, capture_events, uninstall_integration +): + uninstall_integration(FeatureFlagsIntegration.identifier) + sentry_init(integrations=[FeatureFlagsIntegration()]) + events = capture_events() + + # Capture an eval before we split isolation scopes. + add_feature_flag("hello", False) + + def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + add_feature_flag(flag_key, False) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + # Run tasks in separate threads + with cf.ThreadPoolExecutor(max_workers=2) as pool: + pool.map(task, ["world", "other"]) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "world", "result": False}, + ] + } + + +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_featureflags_integration_asyncio( + sentry_init, capture_events, uninstall_integration +): + asyncio = pytest.importorskip("asyncio") + + uninstall_integration(FeatureFlagsIntegration.identifier) + sentry_init(integrations=[FeatureFlagsIntegration()]) + events = capture_events() + + # Capture an eval before we split isolation scopes. + add_feature_flag("hello", False) + + async def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + add_feature_flag(flag_key, False) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + async def runner(): + return asyncio.gather(task("world"), task("other")) + + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "world", "result": False}, + ] + } diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index acbe764104..f66a4219ec 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -1,9 +1,7 @@ -import asyncio import concurrent.futures as cf +import sys import ldclient - -import sentry_sdk import pytest from ldclient import LDClient @@ -11,6 +9,7 @@ from ldclient.context import Context from ldclient.integrations.test_data import TestData +import sentry_sdk from sentry_sdk.integrations import DidNotEnable from sentry_sdk.integrations.launchdarkly import LaunchDarklyIntegration @@ -19,9 +18,13 @@ "use_global_client", (False, True), ) -def test_launchdarkly_integration(sentry_init, use_global_client): +def test_launchdarkly_integration( + sentry_init, use_global_client, capture_events, uninstall_integration +): td = TestData.data_source() config = Config("sdk-key", update_processor_class=td) + + uninstall_integration(LaunchDarklyIntegration.identifier) if use_global_client: ldclient.set_config(config) sentry_init(integrations=[LaunchDarklyIntegration()]) @@ -39,25 +42,38 @@ def test_launchdarkly_integration(sentry_init, use_global_client): client.variation("world", Context.create("user1", "user"), False) client.variation("other", Context.create("user2", "user"), False) - assert sentry_sdk.get_current_scope().flags.get() == [ - {"flag": "hello", "result": True}, - {"flag": "world", "result": True}, - {"flag": "other", "result": False}, - ] + events = capture_events() + sentry_sdk.capture_exception(Exception("something wrong!")) + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": True}, + {"flag": "other", "result": False}, + ] + } -def test_launchdarkly_integration_threaded(sentry_init): + +def test_launchdarkly_integration_threaded( + sentry_init, capture_events, uninstall_integration +): td = TestData.data_source() client = LDClient(config=Config("sdk-key", update_processor_class=td)) - sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) context = Context.create("user1") + uninstall_integration(LaunchDarklyIntegration.identifier) + sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) + events = capture_events() + def task(flag_key): # Creates a new isolation scope for the thread. # This means the evaluations in each task are captured separately. with sentry_sdk.isolation_scope(): client.variation(flag_key, context, False) - return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) td.update(td.flag("hello").variation_for_all(True)) td.update(td.flag("world").variation_for_all(False)) @@ -65,34 +81,91 @@ def task(flag_key): client.variation("hello", context, False) with cf.ThreadPoolExecutor(max_workers=2) as pool: - results = list(pool.map(task, ["world", "other"])) - - assert results[0] == ["hello", "world"] - assert results[1] == ["hello", "other"] + pool.map(task, ["world", "other"]) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } + + +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_launchdarkly_integration_asyncio( + sentry_init, capture_events, uninstall_integration +): + """Assert concurrently evaluated flags do not pollute one another.""" + asyncio = pytest.importorskip("asyncio") -def test_launchdarkly_integration_asyncio(sentry_init): - """Assert concurrently evaluated flags do not pollute one another.""" td = TestData.data_source() client = LDClient(config=Config("sdk-key", update_processor_class=td)) - sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) context = Context.create("user1") + uninstall_integration(LaunchDarklyIntegration.identifier) + sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) + events = capture_events() + async def task(flag_key): with sentry_sdk.isolation_scope(): client.variation(flag_key, context, False) - return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) async def runner(): return asyncio.gather(task("world"), task("other")) td.update(td.flag("hello").variation_for_all(True)) td.update(td.flag("world").variation_for_all(False)) + # Capture an eval before we split isolation scopes. client.variation("hello", context, False) - results = asyncio.run(runner()).result() - assert results[0] == ["hello", "world"] - assert results[1] == ["hello", "other"] + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } def test_launchdarkly_integration_did_not_enable(monkeypatch): diff --git a/tests/integrations/openfeature/test_openfeature.py b/tests/integrations/openfeature/test_openfeature.py index 24e7857f9a..c180211c3f 100644 --- a/tests/integrations/openfeature/test_openfeature.py +++ b/tests/integrations/openfeature/test_openfeature.py @@ -1,13 +1,17 @@ -import asyncio import concurrent.futures as cf -import sentry_sdk +import sys + +import pytest from openfeature import api from openfeature.provider.in_memory_provider import InMemoryFlag, InMemoryProvider + +import sentry_sdk from sentry_sdk.integrations.openfeature import OpenFeatureIntegration -def test_openfeature_integration(sentry_init): +def test_openfeature_integration(sentry_init, capture_events, uninstall_integration): + uninstall_integration(OpenFeatureIntegration.identifier) sentry_init(integrations=[OpenFeatureIntegration()]) flags = { @@ -21,15 +25,25 @@ def test_openfeature_integration(sentry_init): client.get_boolean_value("world", default_value=False) client.get_boolean_value("other", default_value=True) - assert sentry_sdk.get_current_scope().flags.get() == [ - {"flag": "hello", "result": True}, - {"flag": "world", "result": False}, - {"flag": "other", "result": True}, - ] + events = capture_events() + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + {"flag": "other", "result": True}, + ] + } -def test_openfeature_integration_threaded(sentry_init): +def test_openfeature_integration_threaded( + sentry_init, capture_events, uninstall_integration +): + uninstall_integration(OpenFeatureIntegration.identifier) sentry_init(integrations=[OpenFeatureIntegration()]) + events = capture_events() flags = { "hello": InMemoryFlag("on", {"on": True, "off": False}), @@ -37,6 +51,7 @@ def test_openfeature_integration_threaded(sentry_init): } api.set_provider(InMemoryProvider(flags)) + # Capture an eval before we split isolation scopes. client = api.get_client() client.get_boolean_value("hello", default_value=False) @@ -44,37 +59,95 @@ def task(flag): # Create a new isolation scope for the thread. This means the flags with sentry_sdk.isolation_scope(): client.get_boolean_value(flag, default_value=False) - return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag) + sentry_sdk.capture_exception(Exception("something wrong!")) + # Run tasks in separate threads with cf.ThreadPoolExecutor(max_workers=2) as pool: - results = list(pool.map(task, ["world", "other"])) + pool.map(task, ["world", "other"]) - assert results[0] == ["hello", "world"] - assert results[1] == ["hello", "other"] + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } -def test_openfeature_integration_asyncio(sentry_init): +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_openfeature_integration_asyncio( + sentry_init, capture_events, uninstall_integration +): """Assert concurrently evaluated flags do not pollute one another.""" + asyncio = pytest.importorskip("asyncio") + + uninstall_integration(OpenFeatureIntegration.identifier) + sentry_init(integrations=[OpenFeatureIntegration()]) + events = capture_events() + async def task(flag): with sentry_sdk.isolation_scope(): client.get_boolean_value(flag, default_value=False) - return [f["flag"] for f in sentry_sdk.get_current_scope().flags.get()] + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag) + sentry_sdk.capture_exception(Exception("something wrong!")) async def runner(): return asyncio.gather(task("world"), task("other")) - sentry_init(integrations=[OpenFeatureIntegration()]) - flags = { "hello": InMemoryFlag("on", {"on": True, "off": False}), "world": InMemoryFlag("off", {"on": True, "off": False}), } api.set_provider(InMemoryProvider(flags)) + # Capture an eval before we split isolation scopes. client = api.get_client() client.get_boolean_value("hello", default_value=False) - results = asyncio.run(runner()).result() - assert results[0] == ["hello", "world"] - assert results[1] == ["hello", "other"] + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } From fe4b88b8505376ace7c6f8750f83fd2af383190f Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 19 Dec 2024 14:00:09 +0100 Subject: [PATCH 073/591] Add github workflow to comment on issues when a fix was released (#3866) --- .github/workflows/release-comment-issues.yml | 31 ++++++++++++++++++++ 1 file changed, 31 insertions(+) create mode 100644 .github/workflows/release-comment-issues.yml diff --git a/.github/workflows/release-comment-issues.yml b/.github/workflows/release-comment-issues.yml new file mode 100644 index 0000000000..d31c61dced --- /dev/null +++ b/.github/workflows/release-comment-issues.yml @@ -0,0 +1,31 @@ +name: "Automation: Notify issues for release" +on: + release: + types: + - published + workflow_dispatch: + inputs: + version: + description: Which version to notify issues for + required: false + +# This workflow is triggered when a release is published +jobs: + release-comment-issues: + runs-on: ubuntu-20.04 + name: Notify issues + steps: + - name: Get version + id: get_version + run: echo "version=${{ github.event.inputs.version || github.event.release.tag_name }}" >> $GITHUB_OUTPUT + + - name: Comment on linked issues that are mentioned in release + if: | + steps.get_version.outputs.version != '' + && !contains(steps.get_version.outputs.version, 'a') + && !contains(steps.get_version.outputs.version, 'b') + && !contains(steps.get_version.outputs.version, 'rc') + uses: getsentry/release-comment-issues-gh-action@v1 + with: + github_token: ${{ secrets.GITHUB_TOKEN }} + version: ${{ steps.get_version.outputs.version }} \ No newline at end of file From 54aede36f9d3942c1069b47b20b88f01cb461fb5 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 19 Dec 2024 15:34:49 +0100 Subject: [PATCH 074/591] build(deps): bump codecov/codecov-action from 5.0.7 to 5.1.1 (#3867) Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.0.7 to 5.1.1. - [Release notes](https://github.com/codecov/codecov-action/releases) - [Changelog](https://github.com/codecov/codecov-action/blob/main/CHANGELOG.md) - [Commits](https://github.com/codecov/codecov-action/compare/v5.0.7...v5.1.1) --- updated-dependencies: - dependency-name: codecov/codecov-action dependency-type: direct:production update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-misc.yml | 4 ++-- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 4 ++-- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split-tox-gh-actions/templates/test_group.jinja | 2 +- 12 files changed, 21 insertions(+), 21 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 5d1b05add8..8be64736c1 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -78,7 +78,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -150,7 +150,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index d2ce22f326..6eed3a3ab1 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 8fdd4a0649..677385e405 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 8294b9480e..9c476553f5 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -62,7 +62,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 0d9a7bbd7d..cbaa2c32d2 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -101,7 +101,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -196,7 +196,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 30480efe2e..d582717fff 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index b88b256384..00b1286362 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -90,7 +90,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -174,7 +174,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 0a51866164..8f6bd9fd61 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -74,7 +74,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 695c338721..74c868d9b9 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 6e172182b3..5be067a36b 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -178,7 +178,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index f9f2651cb8..7ce0399a13 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -98,7 +98,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -190,7 +190,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split-tox-gh-actions/templates/test_group.jinja index 522be6dc5c..7225bbbfe5 100644 --- a/scripts/split-tox-gh-actions/templates/test_group.jinja +++ b/scripts/split-tox-gh-actions/templates/test_group.jinja @@ -92,7 +92,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.0.7 + uses: codecov/codecov-action@v5.1.1 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From 6e4cc36fbb66a09f4272176fc8972368e1028ae8 Mon Sep 17 00:00:00 2001 From: seyoon-lim Date: Fri, 20 Dec 2024 16:43:19 +0900 Subject: [PATCH 075/591] Support SparkIntegration activation after SparkContext created (#3411) --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/spark/spark_driver.py | 121 +++++++---- tests/integrations/asgi/test_asgi.py | 1 - tests/integrations/spark/test_spark.py | 202 ++++++++++-------- 3 files changed, 189 insertions(+), 135 deletions(-) diff --git a/sentry_sdk/integrations/spark/spark_driver.py b/sentry_sdk/integrations/spark/spark_driver.py index c6470f2302..a86f16344d 100644 --- a/sentry_sdk/integrations/spark/spark_driver.py +++ b/sentry_sdk/integrations/spark/spark_driver.py @@ -9,6 +9,7 @@ from typing import Optional from sentry_sdk._types import Event, Hint + from pyspark import SparkContext class SparkIntegration(Integration): @@ -17,7 +18,7 @@ class SparkIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - patch_spark_context_init() + _setup_sentry_tracing() def _set_app_properties(): @@ -37,7 +38,7 @@ def _set_app_properties(): def _start_sentry_listener(sc): - # type: (Any) -> None + # type: (SparkContext) -> None """ Start java gateway server to add custom `SparkListener` """ @@ -49,7 +50,51 @@ def _start_sentry_listener(sc): sc._jsc.sc().addSparkListener(listener) -def patch_spark_context_init(): +def _add_event_processor(sc): + # type: (SparkContext) -> None + scope = sentry_sdk.get_isolation_scope() + + @scope.add_event_processor + def process_event(event, hint): + # type: (Event, Hint) -> Optional[Event] + with capture_internal_exceptions(): + if sentry_sdk.get_client().get_integration(SparkIntegration) is None: + return event + + if sc._active_spark_context is None: + return event + + event.setdefault("user", {}).setdefault("id", sc.sparkUser()) + + event.setdefault("tags", {}).setdefault( + "executor.id", sc._conf.get("spark.executor.id") + ) + event["tags"].setdefault( + "spark-submit.deployMode", + sc._conf.get("spark.submit.deployMode"), + ) + event["tags"].setdefault("driver.host", sc._conf.get("spark.driver.host")) + event["tags"].setdefault("driver.port", sc._conf.get("spark.driver.port")) + event["tags"].setdefault("spark_version", sc.version) + event["tags"].setdefault("app_name", sc.appName) + event["tags"].setdefault("application_id", sc.applicationId) + event["tags"].setdefault("master", sc.master) + event["tags"].setdefault("spark_home", sc.sparkHome) + + event.setdefault("extra", {}).setdefault("web_url", sc.uiWebUrl) + + return event + + +def _activate_integration(sc): + # type: (SparkContext) -> None + + _start_sentry_listener(sc) + _set_app_properties() + _add_event_processor(sc) + + +def _patch_spark_context_init(): # type: () -> None from pyspark import SparkContext @@ -59,51 +104,22 @@ def patch_spark_context_init(): def _sentry_patched_spark_context_init(self, *args, **kwargs): # type: (SparkContext, *Any, **Any) -> Optional[Any] rv = spark_context_init(self, *args, **kwargs) - _start_sentry_listener(self) - _set_app_properties() - - scope = sentry_sdk.get_isolation_scope() - - @scope.add_event_processor - def process_event(event, hint): - # type: (Event, Hint) -> Optional[Event] - with capture_internal_exceptions(): - if sentry_sdk.get_client().get_integration(SparkIntegration) is None: - return event - - if self._active_spark_context is None: - return event - - event.setdefault("user", {}).setdefault("id", self.sparkUser()) - - event.setdefault("tags", {}).setdefault( - "executor.id", self._conf.get("spark.executor.id") - ) - event["tags"].setdefault( - "spark-submit.deployMode", - self._conf.get("spark.submit.deployMode"), - ) - event["tags"].setdefault( - "driver.host", self._conf.get("spark.driver.host") - ) - event["tags"].setdefault( - "driver.port", self._conf.get("spark.driver.port") - ) - event["tags"].setdefault("spark_version", self.version) - event["tags"].setdefault("app_name", self.appName) - event["tags"].setdefault("application_id", self.applicationId) - event["tags"].setdefault("master", self.master) - event["tags"].setdefault("spark_home", self.sparkHome) - - event.setdefault("extra", {}).setdefault("web_url", self.uiWebUrl) - - return event - + _activate_integration(self) return rv SparkContext._do_init = _sentry_patched_spark_context_init +def _setup_sentry_tracing(): + # type: () -> None + from pyspark import SparkContext + + if SparkContext._active_spark_context is not None: + _activate_integration(SparkContext._active_spark_context) + return + _patch_spark_context_init() + + class SparkListener: def onApplicationEnd(self, applicationEnd): # noqa: N802,N803 # type: (Any) -> None @@ -208,10 +224,21 @@ class Java: class SentryListener(SparkListener): + def _add_breadcrumb( + self, + level, # type: str + message, # type: str + data=None, # type: Optional[dict[str, Any]] + ): + # type: (...) -> None + sentry_sdk.get_global_scope().add_breadcrumb( + level=level, message=message, data=data + ) + def onJobStart(self, jobStart): # noqa: N802,N803 # type: (Any) -> None message = "Job {} Started".format(jobStart.jobId()) - sentry_sdk.add_breadcrumb(level="info", message=message) + self._add_breadcrumb(level="info", message=message) _set_app_properties() def onJobEnd(self, jobEnd): # noqa: N802,N803 @@ -227,14 +254,14 @@ def onJobEnd(self, jobEnd): # noqa: N802,N803 level = "warning" message = "Job {} Failed".format(jobEnd.jobId()) - sentry_sdk.add_breadcrumb(level=level, message=message, data=data) + self._add_breadcrumb(level=level, message=message, data=data) def onStageSubmitted(self, stageSubmitted): # noqa: N802,N803 # type: (Any) -> None stage_info = stageSubmitted.stageInfo() message = "Stage {} Submitted".format(stage_info.stageId()) data = {"attemptId": stage_info.attemptId(), "name": stage_info.name()} - sentry_sdk.add_breadcrumb(level="info", message=message, data=data) + self._add_breadcrumb(level="info", message=message, data=data) _set_app_properties() def onStageCompleted(self, stageCompleted): # noqa: N802,N803 @@ -255,4 +282,4 @@ def onStageCompleted(self, stageCompleted): # noqa: N802,N803 message = "Stage {} Completed".format(stage_info.stageId()) level = "info" - sentry_sdk.add_breadcrumb(level=level, message=message, data=data) + self._add_breadcrumb(level=level, message=message, data=data) diff --git a/tests/integrations/asgi/test_asgi.py b/tests/integrations/asgi/test_asgi.py index e0a3900a38..f3bc7147bf 100644 --- a/tests/integrations/asgi/test_asgi.py +++ b/tests/integrations/asgi/test_asgi.py @@ -128,7 +128,6 @@ async def app(scope, receive, send): @pytest.fixture def asgi3_custom_transaction_app(): - async def app(scope, receive, send): sentry_sdk.get_current_scope().set_transaction_name("foobar", source="custom") await send( diff --git a/tests/integrations/spark/test_spark.py b/tests/integrations/spark/test_spark.py index 58c8862ee2..44ba9f8728 100644 --- a/tests/integrations/spark/test_spark.py +++ b/tests/integrations/spark/test_spark.py @@ -1,6 +1,7 @@ import pytest import sys from unittest.mock import patch + from sentry_sdk.integrations.spark.spark_driver import ( _set_app_properties, _start_sentry_listener, @@ -18,8 +19,22 @@ ################ -def test_set_app_properties(): - spark_context = SparkContext(appName="Testing123") +@pytest.fixture(scope="function") +def sentry_init_with_reset(sentry_init): + from sentry_sdk.integrations import _processed_integrations + + yield lambda: sentry_init(integrations=[SparkIntegration()]) + _processed_integrations.remove("spark") + + +@pytest.fixture(scope="function") +def create_spark_context(): + yield lambda: SparkContext(appName="Testing123") + SparkContext._active_spark_context.stop() + + +def test_set_app_properties(create_spark_context): + spark_context = create_spark_context() _set_app_properties() assert spark_context.getLocalProperty("sentry_app_name") == "Testing123" @@ -30,9 +45,8 @@ def test_set_app_properties(): ) -def test_start_sentry_listener(): - spark_context = SparkContext.getOrCreate() - +def test_start_sentry_listener(create_spark_context): + spark_context = create_spark_context() gateway = spark_context._gateway assert gateway._callback_server is None @@ -41,9 +55,28 @@ def test_start_sentry_listener(): assert gateway._callback_server is not None -def test_initialize_spark_integration(sentry_init): - sentry_init(integrations=[SparkIntegration()]) - SparkContext.getOrCreate() +@patch("sentry_sdk.integrations.spark.spark_driver._patch_spark_context_init") +def test_initialize_spark_integration_before_spark_context_init( + mock_patch_spark_context_init, + sentry_init_with_reset, + create_spark_context, +): + sentry_init_with_reset() + create_spark_context() + + mock_patch_spark_context_init.assert_called_once() + + +@patch("sentry_sdk.integrations.spark.spark_driver._activate_integration") +def test_initialize_spark_integration_after_spark_context_init( + mock_activate_integration, + create_spark_context, + sentry_init_with_reset, +): + create_spark_context() + sentry_init_with_reset() + + mock_activate_integration.assert_called_once() @pytest.fixture @@ -54,88 +87,83 @@ def sentry_listener(): return listener -@pytest.fixture -def mock_add_breadcrumb(): - with patch("sentry_sdk.add_breadcrumb") as mock: - yield mock - - -def test_sentry_listener_on_job_start(sentry_listener, mock_add_breadcrumb): +def test_sentry_listener_on_job_start(sentry_listener): listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: - class MockJobStart: - def jobId(self): # noqa: N802 - return "sample-job-id-start" + class MockJobStart: + def jobId(self): # noqa: N802 + return "sample-job-id-start" - mock_job_start = MockJobStart() - listener.onJobStart(mock_job_start) + mock_job_start = MockJobStart() + listener.onJobStart(mock_job_start) - mock_add_breadcrumb.assert_called_once() - mock_hub = mock_add_breadcrumb.call_args + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args - assert mock_hub.kwargs["level"] == "info" - assert "sample-job-id-start" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["level"] == "info" + assert "sample-job-id-start" in mock_hub.kwargs["message"] @pytest.mark.parametrize( "job_result, level", [("JobSucceeded", "info"), ("JobFailed", "warning")] ) -def test_sentry_listener_on_job_end( - sentry_listener, mock_add_breadcrumb, job_result, level -): +def test_sentry_listener_on_job_end(sentry_listener, job_result, level): listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: - class MockJobResult: - def toString(self): # noqa: N802 - return job_result + class MockJobResult: + def toString(self): # noqa: N802 + return job_result - class MockJobEnd: - def jobId(self): # noqa: N802 - return "sample-job-id-end" + class MockJobEnd: + def jobId(self): # noqa: N802 + return "sample-job-id-end" - def jobResult(self): # noqa: N802 - result = MockJobResult() - return result + def jobResult(self): # noqa: N802 + result = MockJobResult() + return result - mock_job_end = MockJobEnd() - listener.onJobEnd(mock_job_end) + mock_job_end = MockJobEnd() + listener.onJobEnd(mock_job_end) - mock_add_breadcrumb.assert_called_once() - mock_hub = mock_add_breadcrumb.call_args + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args - assert mock_hub.kwargs["level"] == level - assert mock_hub.kwargs["data"]["result"] == job_result - assert "sample-job-id-end" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["level"] == level + assert mock_hub.kwargs["data"]["result"] == job_result + assert "sample-job-id-end" in mock_hub.kwargs["message"] -def test_sentry_listener_on_stage_submitted(sentry_listener, mock_add_breadcrumb): +def test_sentry_listener_on_stage_submitted(sentry_listener): listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: - class StageInfo: - def stageId(self): # noqa: N802 - return "sample-stage-id-submit" + class StageInfo: + def stageId(self): # noqa: N802 + return "sample-stage-id-submit" - def name(self): - return "run-job" + def name(self): + return "run-job" - def attemptId(self): # noqa: N802 - return 14 + def attemptId(self): # noqa: N802 + return 14 - class MockStageSubmitted: - def stageInfo(self): # noqa: N802 - stageinf = StageInfo() - return stageinf + class MockStageSubmitted: + def stageInfo(self): # noqa: N802 + stageinf = StageInfo() + return stageinf - mock_stage_submitted = MockStageSubmitted() - listener.onStageSubmitted(mock_stage_submitted) + mock_stage_submitted = MockStageSubmitted() + listener.onStageSubmitted(mock_stage_submitted) - mock_add_breadcrumb.assert_called_once() - mock_hub = mock_add_breadcrumb.call_args + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args - assert mock_hub.kwargs["level"] == "info" - assert "sample-stage-id-submit" in mock_hub.kwargs["message"] - assert mock_hub.kwargs["data"]["attemptId"] == 14 - assert mock_hub.kwargs["data"]["name"] == "run-job" + assert mock_hub.kwargs["level"] == "info" + assert "sample-stage-id-submit" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["data"]["attemptId"] == 14 + assert mock_hub.kwargs["data"]["name"] == "run-job" @pytest.fixture @@ -175,39 +203,39 @@ def stageInfo(self): # noqa: N802 def test_sentry_listener_on_stage_completed_success( - sentry_listener, mock_add_breadcrumb, get_mock_stage_completed + sentry_listener, get_mock_stage_completed ): listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: + mock_stage_completed = get_mock_stage_completed(failure_reason=False) + listener.onStageCompleted(mock_stage_completed) - mock_stage_completed = get_mock_stage_completed(failure_reason=False) - listener.onStageCompleted(mock_stage_completed) - - mock_add_breadcrumb.assert_called_once() - mock_hub = mock_add_breadcrumb.call_args + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args - assert mock_hub.kwargs["level"] == "info" - assert "sample-stage-id-submit" in mock_hub.kwargs["message"] - assert mock_hub.kwargs["data"]["attemptId"] == 14 - assert mock_hub.kwargs["data"]["name"] == "run-job" - assert "reason" not in mock_hub.kwargs["data"] + assert mock_hub.kwargs["level"] == "info" + assert "sample-stage-id-submit" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["data"]["attemptId"] == 14 + assert mock_hub.kwargs["data"]["name"] == "run-job" + assert "reason" not in mock_hub.kwargs["data"] def test_sentry_listener_on_stage_completed_failure( - sentry_listener, mock_add_breadcrumb, get_mock_stage_completed + sentry_listener, get_mock_stage_completed ): listener = sentry_listener - - mock_stage_completed = get_mock_stage_completed(failure_reason=True) - listener.onStageCompleted(mock_stage_completed) - - mock_add_breadcrumb.assert_called_once() - mock_hub = mock_add_breadcrumb.call_args - - assert mock_hub.kwargs["level"] == "warning" - assert "sample-stage-id-submit" in mock_hub.kwargs["message"] - assert mock_hub.kwargs["data"]["attemptId"] == 14 - assert mock_hub.kwargs["data"]["name"] == "run-job" - assert mock_hub.kwargs["data"]["reason"] == "failure-reason" + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: + mock_stage_completed = get_mock_stage_completed(failure_reason=True) + listener.onStageCompleted(mock_stage_completed) + + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args + + assert mock_hub.kwargs["level"] == "warning" + assert "sample-stage-id-submit" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["data"]["attemptId"] == 14 + assert mock_hub.kwargs["data"]["name"] == "run-job" + assert mock_hub.kwargs["data"]["reason"] == "failure-reason" ################ From 8ced6609e6fcc95855f43cf9fc1d94b59836b57f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 20 Dec 2024 10:15:48 +0100 Subject: [PATCH 076/591] Rename scripts (#3885) --- .github/workflows/ci.yml | 4 ++-- .github/workflows/test-integrations-ai.yml | 6 ++++-- .github/workflows/test-integrations-aws.yml | 6 ++++-- .github/workflows/test-integrations-cloud.yml | 6 ++++-- .github/workflows/test-integrations-common.yml | 6 ++++-- .github/workflows/test-integrations-dbs.yml | 6 ++++-- .github/workflows/test-integrations-graphql.yml | 6 ++++-- .github/workflows/test-integrations-misc.yml | 6 ++++-- .github/workflows/test-integrations-network.yml | 6 ++++-- .github/workflows/test-integrations-tasks.yml | 6 ++++-- .github/workflows/test-integrations-web-1.yml | 6 ++++-- .github/workflows/test-integrations-web-2.yml | 6 ++++-- ...er-versions.sh => aws-delete-lambda-layer-versions.sh} | 0 scripts/split_tox_gh_actions/__init__.py | 0 scripts/split_tox_gh_actions/requirements.txt | 1 + .../split_tox_gh_actions.py} | 8 ++++---- .../templates/base.jinja | 6 ++++-- .../templates/check_permissions.jinja | 0 .../templates/check_required.jinja | 0 .../templates/test_group.jinja | 0 20 files changed, 55 insertions(+), 30 deletions(-) rename scripts/{aws-delete-lamba-layer-versions.sh => aws-delete-lambda-layer-versions.sh} (100%) create mode 100644 scripts/split_tox_gh_actions/__init__.py create mode 100644 scripts/split_tox_gh_actions/requirements.txt rename scripts/{split-tox-gh-actions/split-tox-gh-actions.py => split_tox_gh_actions/split_tox_gh_actions.py} (96%) rename scripts/{split-tox-gh-actions => split_tox_gh_actions}/templates/base.jinja (87%) rename scripts/{split-tox-gh-actions => split_tox_gh_actions}/templates/check_permissions.jinja (100%) rename scripts/{split-tox-gh-actions => split_tox_gh_actions}/templates/check_required.jinja (100%) rename scripts/{split-tox-gh-actions => split_tox_gh_actions}/templates/test_group.jinja (100%) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index ed035b4ab0..7ef6604e39 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -45,8 +45,8 @@ jobs: python-version: 3.12 - run: | - pip install jinja2 - python scripts/split-tox-gh-actions/split-tox-gh-actions.py --fail-on-changes + pip install -r scripts/split_tox_gh_actions/requirements.txt + python scripts/split_tox_gh_actions/split_tox_gh_actions.py --fail-on-changes build_lambda_layer: name: Build Package diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 8be64736c1..c5e1f6b87e 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test AI on: push: diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index 6eed3a3ab1..54610f1abd 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test AWS on: push: diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 677385e405..f72fec9f9f 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Cloud on: push: diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 9c476553f5..0837c60c30 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Common on: push: diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index cbaa2c32d2..a4aefa6a51 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test DBs on: push: diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index d582717fff..ab7e81dcd6 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test GraphQL on: push: diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 00b1286362..1a4e910383 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Misc on: push: diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 8f6bd9fd61..f41fd86b29 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Network on: push: diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 74c868d9b9..9910b75568 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Tasks on: push: diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 5be067a36b..fb7a9247d5 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Web 1 on: push: diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 7ce0399a13..1910d5999e 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja name: Test Web 2 on: push: diff --git a/scripts/aws-delete-lamba-layer-versions.sh b/scripts/aws-delete-lambda-layer-versions.sh similarity index 100% rename from scripts/aws-delete-lamba-layer-versions.sh rename to scripts/aws-delete-lambda-layer-versions.sh diff --git a/scripts/split_tox_gh_actions/__init__.py b/scripts/split_tox_gh_actions/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/scripts/split_tox_gh_actions/requirements.txt b/scripts/split_tox_gh_actions/requirements.txt new file mode 100644 index 0000000000..7f7afbf3bf --- /dev/null +++ b/scripts/split_tox_gh_actions/requirements.txt @@ -0,0 +1 @@ +jinja2 diff --git a/scripts/split-tox-gh-actions/split-tox-gh-actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py similarity index 96% rename from scripts/split-tox-gh-actions/split-tox-gh-actions.py rename to scripts/split_tox_gh_actions/split_tox_gh_actions.py index 26d13390c2..1b53093c5e 100755 --- a/scripts/split-tox-gh-actions/split-tox-gh-actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -8,7 +8,7 @@ Whenever tox.ini is changed, this script needs to be run. Usage: - python split-tox-gh-actions.py [--fail-on-changes] + python split_tox_gh_actions.py [--fail-on-changes] If the parameter `--fail-on-changes` is set, the script will raise a RuntimeError in case the yaml files have been changed by the scripts execution. This is used in CI to check if the yaml files @@ -158,7 +158,7 @@ def main(fail_on_changes): if missing_frameworks: raise RuntimeError( "Please add the following frameworks to the corresponding group " - "in `GROUPS` in `scripts/split-tox-gh-actions/split-tox-gh-actions.py: " + "in `GROUPS` in `scripts/split_tox_gh_actions/split_tox_gh_actions.py: " + ", ".join(missing_frameworks) ) @@ -176,9 +176,9 @@ def main(fail_on_changes): if old_hash != new_hash: raise RuntimeError( "The yaml configuration files have changed. This means that either `tox.ini` " - "or one of the constants in `split-tox-gh-actions.py` has changed " + "or one of the constants in `split_tox_gh_actions.py` has changed " "but the changes have not been propagated to the GitHub actions config files. " - "Please run `python scripts/split-tox-gh-actions/split-tox-gh-actions.py` " + "Please run `python scripts/split_tox_gh_actions/split_tox_gh_actions.py` " "locally and commit the changes of the yaml configuration files to continue. " ) diff --git a/scripts/split-tox-gh-actions/templates/base.jinja b/scripts/split_tox_gh_actions/templates/base.jinja similarity index 87% rename from scripts/split-tox-gh-actions/templates/base.jinja rename to scripts/split_tox_gh_actions/templates/base.jinja index 23f051de42..16dbc04a76 100644 --- a/scripts/split-tox-gh-actions/templates/base.jinja +++ b/scripts/split_tox_gh_actions/templates/base.jinja @@ -1,5 +1,7 @@ -# Do not edit this file. This file is generated automatically by executing -# python scripts/split-tox-gh-actions/split-tox-gh-actions.py +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja {% with lowercase_group=group | replace(" ", "_") | lower %} name: Test {{ group }} diff --git a/scripts/split-tox-gh-actions/templates/check_permissions.jinja b/scripts/split_tox_gh_actions/templates/check_permissions.jinja similarity index 100% rename from scripts/split-tox-gh-actions/templates/check_permissions.jinja rename to scripts/split_tox_gh_actions/templates/check_permissions.jinja diff --git a/scripts/split-tox-gh-actions/templates/check_required.jinja b/scripts/split_tox_gh_actions/templates/check_required.jinja similarity index 100% rename from scripts/split-tox-gh-actions/templates/check_required.jinja rename to scripts/split_tox_gh_actions/templates/check_required.jinja diff --git a/scripts/split-tox-gh-actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja similarity index 100% rename from scripts/split-tox-gh-actions/templates/test_group.jinja rename to scripts/split_tox_gh_actions/templates/test_group.jinja From f6281f557fe62c847a0aca95eb666129e893cf32 Mon Sep 17 00:00:00 2001 From: ffelixg <142172984+ffelixg@users.noreply.github.com> Date: Fri, 20 Dec 2024 12:34:12 +0100 Subject: [PATCH 077/591] Fix lru cache copying (#3883) A simpler and better LRU Cache implementation that prevents data leaking between copied caches. Fixes #3852 --------- Co-authored-by: Anton Pirker --- sentry_sdk/_lru_cache.py | 195 +++++++-------------------------------- tests/test_lru_cache.py | 37 +++++++- tests/test_scope.py | 22 +++++ 3 files changed, 93 insertions(+), 161 deletions(-) diff --git a/sentry_sdk/_lru_cache.py b/sentry_sdk/_lru_cache.py index 825c773529..09eae27df2 100644 --- a/sentry_sdk/_lru_cache.py +++ b/sentry_sdk/_lru_cache.py @@ -1,181 +1,56 @@ -""" -A fork of Python 3.6's stdlib lru_cache (found in Python's 'cpython/Lib/functools.py') -adapted into a data structure for single threaded uses. +from typing import TYPE_CHECKING -https://github.com/python/cpython/blob/v3.6.12/Lib/functools.py +if TYPE_CHECKING: + from typing import Any -Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, -2011, 2012, 2013, 2014, 2015, 2016, 2017, 2018, 2019, 2020 Python Software Foundation; - -All Rights Reserved - - -PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2 --------------------------------------------- - -1. This LICENSE AGREEMENT is between the Python Software Foundation -("PSF"), and the Individual or Organization ("Licensee") accessing and -otherwise using this software ("Python") in source or binary form and -its associated documentation. - -2. Subject to the terms and conditions of this License Agreement, PSF hereby -grants Licensee a nonexclusive, royalty-free, world-wide license to reproduce, -analyze, test, perform and/or display publicly, prepare derivative works, -distribute, and otherwise use Python alone or in any derivative version, -provided, however, that PSF's License Agreement and PSF's notice of copyright, -i.e., "Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, -2011, 2012, 2013, 2014, 2015, 2016, 2017, 2018, 2019, 2020 Python Software Foundation; -All Rights Reserved" are retained in Python alone or in any derivative version -prepared by Licensee. - -3. In the event Licensee prepares a derivative work that is based on -or incorporates Python or any part thereof, and wants to make -the derivative work available to others as provided herein, then -Licensee hereby agrees to include in any such work a brief summary of -the changes made to Python. - -4. PSF is making Python available to Licensee on an "AS IS" -basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR -IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND -DISCLAIMS ANY REPRESENTATION OR WARRANTY OF MERCHANTABILITY OR FITNESS -FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT -INFRINGE ANY THIRD PARTY RIGHTS. - -5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON -FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL DAMAGES OR LOSS AS -A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, -OR ANY DERIVATIVE THEREOF, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. - -6. This License Agreement will automatically terminate upon a material -breach of its terms and conditions. - -7. Nothing in this License Agreement shall be deemed to create any -relationship of agency, partnership, or joint venture between PSF and -Licensee. This License Agreement does not grant permission to use PSF -trademarks or trade name in a trademark sense to endorse or promote -products or services of Licensee, or any third party. - -8. By copying, installing or otherwise using Python, Licensee -agrees to be bound by the terms and conditions of this License -Agreement. - -""" - -from copy import copy, deepcopy - -SENTINEL = object() - - -# aliases to the entries in a node -PREV = 0 -NEXT = 1 -KEY = 2 -VALUE = 3 +_SENTINEL = object() class LRUCache: def __init__(self, max_size): - assert max_size > 0 - + # type: (int) -> None + if max_size <= 0: + raise AssertionError(f"invalid max_size: {max_size}") self.max_size = max_size - self.full = False - - self.cache = {} - - # root of the circularly linked list to keep track of - # the least recently used key - self.root = [] # type: ignore - # the node looks like [PREV, NEXT, KEY, VALUE] - self.root[:] = [self.root, self.root, None, None] - + self._data = {} # type: dict[Any, Any] self.hits = self.misses = 0 + self.full = False def __copy__(self): - cache = LRUCache(self.max_size) - cache.full = self.full - cache.cache = copy(self.cache) - cache.root = deepcopy(self.root) - return cache + # type: () -> LRUCache + new = LRUCache(max_size=self.max_size) + new.hits = self.hits + new.misses = self.misses + new.full = self.full + new._data = self._data.copy() + return new def set(self, key, value): - link = self.cache.get(key, SENTINEL) - - if link is not SENTINEL: - # have to move the node to the front of the linked list - link_prev, link_next, _key, _value = link - - # first remove the node from the lsnked list - link_prev[NEXT] = link_next - link_next[PREV] = link_prev - - # insert the node between the root and the last - last = self.root[PREV] - last[NEXT] = self.root[PREV] = link - link[PREV] = last - link[NEXT] = self.root - - # update the value - link[VALUE] = value - + # type: (Any, Any) -> None + current = self._data.pop(key, _SENTINEL) + if current is not _SENTINEL: + self._data[key] = value elif self.full: - # reuse the root node, so update its key/value - old_root = self.root - old_root[KEY] = key - old_root[VALUE] = value - - self.root = old_root[NEXT] - old_key = self.root[KEY] - - self.root[KEY] = self.root[VALUE] = None - - del self.cache[old_key] - - self.cache[key] = old_root - + self._data.pop(next(iter(self._data))) + self._data[key] = value else: - # insert new node after last - last = self.root[PREV] - link = [last, self.root, key, value] - last[NEXT] = self.root[PREV] = self.cache[key] = link - self.full = len(self.cache) >= self.max_size + self._data[key] = value + self.full = len(self._data) >= self.max_size def get(self, key, default=None): - link = self.cache.get(key, SENTINEL) - - if link is SENTINEL: + # type: (Any, Any) -> Any + try: + ret = self._data.pop(key) + except KeyError: self.misses += 1 - return default - - # have to move the node to the front of the linked list - link_prev, link_next, _key, _value = link - - # first remove the node from the lsnked list - link_prev[NEXT] = link_next - link_next[PREV] = link_prev - - # insert the node between the root and the last - last = self.root[PREV] - last[NEXT] = self.root[PREV] = link - link[PREV] = last - link[NEXT] = self.root - - self.hits += 1 + ret = default + else: + self.hits += 1 + self._data[key] = ret - return link[VALUE] + return ret def get_all(self): - nodes = [] - node = self.root[NEXT] - - # To ensure the loop always terminates we iterate to the maximum - # size of the LRU cache. - for _ in range(self.max_size): - # The cache may not be full. We exit early if we've wrapped - # around to the head. - if node is self.root: - break - nodes.append((node[KEY], node[VALUE])) - node = node[NEXT] - - return nodes + # type: () -> list[tuple[Any, Any]] + return list(self._data.items()) diff --git a/tests/test_lru_cache.py b/tests/test_lru_cache.py index cab9bbc7eb..1a54ed83d3 100644 --- a/tests/test_lru_cache.py +++ b/tests/test_lru_cache.py @@ -1,5 +1,5 @@ import pytest -from copy import copy +from copy import copy, deepcopy from sentry_sdk._lru_cache import LRUCache @@ -76,3 +76,38 @@ def test_cache_copy(): cache.get(1) assert copied.get_all() == [(1, 1), (2, 2), (3, 3)] assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] + + +def test_cache_deepcopy(): + cache = LRUCache(3) + cache.set(0, 0) + cache.set(1, 1) + + copied = deepcopy(cache) + cache.set(2, 2) + cache.set(3, 3) + assert copied.get_all() == [(0, 0), (1, 1)] + assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] + + copied = deepcopy(cache) + cache.get(1) + assert copied.get_all() == [(1, 1), (2, 2), (3, 3)] + assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] + + +def test_cache_pollution(): + cache1 = LRUCache(max_size=2) + cache1.set(1, True) + cache2 = copy(cache1) + cache2.set(1, False) + assert cache1.get(1) is True + assert cache2.get(1) is False + + +def test_cache_pollution_deepcopy(): + cache1 = LRUCache(max_size=2) + cache1.set(1, True) + cache2 = deepcopy(cache1) + cache2.set(1, False) + assert cache1.get(1) is True + assert cache2.get(1) is False diff --git a/tests/test_scope.py b/tests/test_scope.py index a03eb07a99..9b16dc4344 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -43,6 +43,28 @@ def test_all_slots_copied(): assert getattr(scope_copy, attr) == getattr(scope, attr) +def test_scope_flags_copy(): + # Assert forking creates a deepcopy of the flag buffer. The new + # scope is free to mutate without consequence to the old scope. The + # old scope is free to mutate without consequence to the new scope. + old_scope = Scope() + old_scope.flags.set("a", True) + + new_scope = old_scope.fork() + new_scope.flags.set("a", False) + old_scope.flags.set("b", True) + new_scope.flags.set("c", True) + + assert old_scope.flags.get() == [ + {"flag": "a", "result": True}, + {"flag": "b", "result": True}, + ] + assert new_scope.flags.get() == [ + {"flag": "a", "result": False}, + {"flag": "c", "result": True}, + ] + + def test_merging(sentry_init, capture_events): sentry_init() From 00c5961cadd23ded77982b085d36ce526ca8ece3 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 23 Dec 2024 09:43:02 +0100 Subject: [PATCH 078/591] build(deps): bump codecov/codecov-action from 5.1.1 to 5.1.2 (#3892) * build(deps): bump codecov/codecov-action from 5.1.1 to 5.1.2 Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.1.1 to 5.1.2. - [Release notes](https://github.com/codecov/codecov-action/releases) - [Changelog](https://github.com/codecov/codecov-action/blob/main/CHANGELOG.md) - [Commits](https://github.com/codecov/codecov-action/compare/v5.1.1...v5.1.2) --- updated-dependencies: - dependency-name: codecov/codecov-action dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] * Updated template * Update linting config to work with new mypy version --------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-misc.yml | 4 ++-- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 4 ++-- .github/workflows/test-integrations-web-2.yml | 4 ++-- .../split_tox_gh_actions/templates/test_group.jinja | 2 +- sentry_sdk/client.py | 6 +++--- sentry_sdk/integrations/rust_tracing.py | 10 +++++----- 14 files changed, 29 insertions(+), 29 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index c5e1f6b87e..2fd6995a5f 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -80,7 +80,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -152,7 +152,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index 54610f1abd..f83e3379f6 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -99,7 +99,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index f72fec9f9f..9e34dc6b2b 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -144,7 +144,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 0837c60c30..f1806597af 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -64,7 +64,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index a4aefa6a51..d9bea0611b 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index ab7e81dcd6..7138204e16 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -144,7 +144,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 1a4e910383..79b7ba020d 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -92,7 +92,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -176,7 +176,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index f41fd86b29..1b9ee3c529 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -144,7 +144,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 9910b75568..0f97146d6d 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index fb7a9247d5..53206f764f 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 1910d5999e..f1fbec6c67 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -100,7 +100,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -192,7 +192,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 7225bbbfe5..186d70c9fd 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -92,7 +92,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.1.1 + uses: codecov/codecov-action@v5.1.2 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index db2cc19110..cf345c41f9 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -532,7 +532,7 @@ def _prepare_event( for key in "release", "environment", "server_name", "dist": if event.get(key) is None and self.options[key] is not None: - event[key] = str(self.options[key]).strip() # type: ignore[literal-required] + event[key] = str(self.options[key]).strip() if event.get("sdk") is None: sdk_info = dict(SDK_INFO) sdk_info["integrations"] = sorted(self.integrations.keys()) @@ -581,7 +581,7 @@ def _prepare_event( self.transport.record_lost_event( "before_send", data_category="error" ) - event = new_event # type: ignore + event = new_event before_send_transaction = self.options["before_send_transaction"] if ( @@ -611,7 +611,7 @@ def _prepare_event( reason="before_send", data_category="span", quantity=spans_delta ) - event = new_event # type: ignore + event = new_event return event diff --git a/sentry_sdk/integrations/rust_tracing.py b/sentry_sdk/integrations/rust_tracing.py index ae52c850c3..e4c211814f 100644 --- a/sentry_sdk/integrations/rust_tracing.py +++ b/sentry_sdk/integrations/rust_tracing.py @@ -44,11 +44,11 @@ class RustTracingLevel(Enum): - Trace: str = "TRACE" - Debug: str = "DEBUG" - Info: str = "INFO" - Warn: str = "WARN" - Error: str = "ERROR" + Trace = "TRACE" + Debug = "DEBUG" + Info = "INFO" + Warn = "WARN" + Error = "ERROR" class EventTypeMapping(Enum): From 60fb6fc4eacb3b4e8fffd81a0a6079e0ea31bfcf Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 23 Dec 2024 08:56:54 +0000 Subject: [PATCH 079/591] build(deps): bump actions/create-github-app-token from 1.11.0 to 1.11.1 (#3893) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.0 to 1.11.1. - [Release notes](https://github.com/actions/create-github-app-token/releases) - [Commits](https://github.com/actions/create-github-app-token/compare/5d869da34e18e7287c1daad50e0b8ea0f506ce69...c1a285145b9d317df6ced56c09f525b5c2b6f755) --- updated-dependencies: - dependency-name: actions/create-github-app-token dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 2cd3dfb2ac..6450150138 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@5d869da34e18e7287c1daad50e0b8ea0f506ce69 # v1.11.0 + uses: actions/create-github-app-token@c1a285145b9d317df6ced56c09f525b5c2b6f755 # v1.11.1 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From c3516db643af20396ea981393431646f1a3ef123 Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Mon, 23 Dec 2024 02:02:20 -0800 Subject: [PATCH 080/591] ref(flags): register LD hook in setup instead of init, and don't check for initialization (#3890) --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/launchdarkly.py | 14 ++++++------- .../launchdarkly/test_launchdarkly.py | 21 +++++++++---------- 2 files changed, 17 insertions(+), 18 deletions(-) diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index a9eef9e1a9..066464cc22 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -20,6 +20,7 @@ class LaunchDarklyIntegration(Integration): identifier = "launchdarkly" + _ld_client = None # type: LDClient | None def __init__(self, ld_client=None): # type: (LDClient | None) -> None @@ -27,20 +28,19 @@ def __init__(self, ld_client=None): :param client: An initialized LDClient instance. If a client is not provided, this integration will attempt to use the shared global instance. """ + self.__class__._ld_client = ld_client + + @staticmethod + def setup_once(): + # type: () -> None try: - client = ld_client or ldclient.get() + client = LaunchDarklyIntegration._ld_client or ldclient.get() except Exception as exc: raise DidNotEnable("Error getting LaunchDarkly client. " + repr(exc)) - if not client.is_initialized(): - raise DidNotEnable("LaunchDarkly client is not initialized.") - # Register the flag collection hook with the LD client. client.add_hook(LaunchDarklyHook()) - @staticmethod - def setup_once(): - # type: () -> None scope = sentry_sdk.get_current_scope() scope.add_error_processor(flag_error_processor) diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index f66a4219ec..e7576bb469 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -168,10 +168,14 @@ async def runner(): } -def test_launchdarkly_integration_did_not_enable(monkeypatch): - # Client is not passed in and set_config wasn't called. - # TODO: Bad practice to access internals like this. We can skip this test, or remove this - # case entirely (force user to pass in a client instance). +def test_launchdarkly_integration_did_not_enable(sentry_init, uninstall_integration): + """ + Setup should fail when using global client and ldclient.set_config wasn't called. + + We're accessing ldclient internals to set up this test, so it might break if launchdarkly's + implementation changes. + """ + ldclient._reset_client() try: ldclient.__lock.lock() @@ -179,11 +183,6 @@ def test_launchdarkly_integration_did_not_enable(monkeypatch): finally: ldclient.__lock.unlock() + uninstall_integration(LaunchDarklyIntegration.identifier) with pytest.raises(DidNotEnable): - LaunchDarklyIntegration() - - # Client not initialized. - client = LDClient(config=Config("sdk-key")) - monkeypatch.setattr(client, "is_initialized", lambda: False) - with pytest.raises(DidNotEnable): - LaunchDarklyIntegration(ld_client=client) + sentry_init(integrations=[LaunchDarklyIntegration()]) From bb85c26a2b877965c5e0a0cd841b7f676ec2533e Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Mon, 23 Dec 2024 04:37:17 -0600 Subject: [PATCH 081/591] Fix cache pollution from mutable reference (#3887) - Removes manual overrides of copy behavior and leaves it up to the caller. - E.g. a future use case may require a non-deepcopy. If we override copy they would have to remove the dunder copy, update every implementation which relies copy, before finally creating their own copy implementation. - Deepcopies the flag buffer. - Though we do not cache mutable references yet we may soon and so this foot gun should be removed from possibility. - Removes "copy" test coverage from `test_lru_cache.py`. We're no longer assuming copy usage and leave it up to the caller. - The existing test in `tests/test_scope.py` covers the cache pollution case [originally mentioned here](https://github.com/getsentry/sentry-python/issues/3852). - The mutable cache pollution case is not covered because we do not currently cache mutable objects. In general a generic class should assume as few implementation details as possible. If we leave the existing copy method someone may assume copy semantics and rely on it in a way that is inappropriate. Closes: https://github.com/getsentry/sentry-python/issues/3886 Co-authored-by: Anton Pirker --- sentry_sdk/_lru_cache.py | 9 ------- sentry_sdk/flag_utils.py | 7 ------ sentry_sdk/scope.py | 4 +-- tests/test_lru_cache.py | 53 ---------------------------------------- 4 files changed, 2 insertions(+), 71 deletions(-) diff --git a/sentry_sdk/_lru_cache.py b/sentry_sdk/_lru_cache.py index 09eae27df2..cbadd9723b 100644 --- a/sentry_sdk/_lru_cache.py +++ b/sentry_sdk/_lru_cache.py @@ -17,15 +17,6 @@ def __init__(self, max_size): self.hits = self.misses = 0 self.full = False - def __copy__(self): - # type: () -> LRUCache - new = LRUCache(max_size=self.max_size) - new.hits = self.hits - new.misses = self.misses - new.full = self.full - new._data = self._data.copy() - return new - def set(self, key, value): # type: (Any, Any) -> None current = self._data.pop(key, _SENTINEL) diff --git a/sentry_sdk/flag_utils.py b/sentry_sdk/flag_utils.py index 2b345a7f0b..cf4800e855 100644 --- a/sentry_sdk/flag_utils.py +++ b/sentry_sdk/flag_utils.py @@ -1,4 +1,3 @@ -from copy import copy from typing import TYPE_CHECKING import sentry_sdk @@ -25,12 +24,6 @@ def clear(self): # type: () -> None self.buffer = LRUCache(self.capacity) - def __copy__(self): - # type: () -> FlagBuffer - buffer = FlagBuffer(capacity=self.capacity) - buffer.buffer = copy(self.buffer) - return buffer - def get(self): # type: () -> list[FlagData] return [{"flag": key, "result": value} for key, value in self.buffer.get_all()] diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index bb45143c48..cf72fabdd1 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -1,7 +1,7 @@ import os import sys import warnings -from copy import copy +from copy import copy, deepcopy from collections import deque from contextlib import contextmanager from enum import Enum @@ -252,7 +252,7 @@ def __copy__(self): rv._last_event_id = self._last_event_id - rv._flags = copy(self._flags) + rv._flags = deepcopy(self._flags) return rv diff --git a/tests/test_lru_cache.py b/tests/test_lru_cache.py index 1a54ed83d3..3e9c0ac964 100644 --- a/tests/test_lru_cache.py +++ b/tests/test_lru_cache.py @@ -1,5 +1,4 @@ import pytest -from copy import copy, deepcopy from sentry_sdk._lru_cache import LRUCache @@ -59,55 +58,3 @@ def test_cache_get_all(): assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] cache.get(1) assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] - - -def test_cache_copy(): - cache = LRUCache(3) - cache.set(0, 0) - cache.set(1, 1) - - copied = copy(cache) - cache.set(2, 2) - cache.set(3, 3) - assert copied.get_all() == [(0, 0), (1, 1)] - assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] - - copied = copy(cache) - cache.get(1) - assert copied.get_all() == [(1, 1), (2, 2), (3, 3)] - assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] - - -def test_cache_deepcopy(): - cache = LRUCache(3) - cache.set(0, 0) - cache.set(1, 1) - - copied = deepcopy(cache) - cache.set(2, 2) - cache.set(3, 3) - assert copied.get_all() == [(0, 0), (1, 1)] - assert cache.get_all() == [(1, 1), (2, 2), (3, 3)] - - copied = deepcopy(cache) - cache.get(1) - assert copied.get_all() == [(1, 1), (2, 2), (3, 3)] - assert cache.get_all() == [(2, 2), (3, 3), (1, 1)] - - -def test_cache_pollution(): - cache1 = LRUCache(max_size=2) - cache1.set(1, True) - cache2 = copy(cache1) - cache2.set(1, False) - assert cache1.get(1) is True - assert cache2.get(1) is False - - -def test_cache_pollution_deepcopy(): - cache1 = LRUCache(max_size=2) - cache1.set(1, True) - cache2 = deepcopy(cache1) - cache2.set(1, False) - assert cache1.get(1) is True - assert cache2.get(1) is False From fd224946e084ad6bf6e55d6c4216cb8399e15c7e Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Tue, 7 Jan 2025 01:56:08 -0800 Subject: [PATCH 082/591] fix(flags): fix/refactor flaky launchdarkly tests (#3896) Fixes flakes ([example](https://github.com/getsentry/sentry-python/actions/runs/12465223145/job/34790658871?pr=3887)) caused by background processes in `LDClient` trying to connect to a non-existent server (we're mocking the flag data through `TestData`). --- .../launchdarkly/test_launchdarkly.py | 41 +++++++++++++------ 1 file changed, 28 insertions(+), 13 deletions(-) diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index e7576bb469..9b2bbb6b86 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -22,7 +22,12 @@ def test_launchdarkly_integration( sentry_init, use_global_client, capture_events, uninstall_integration ): td = TestData.data_source() - config = Config("sdk-key", update_processor_class=td) + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(True)) + # Disable background requests as we aren't using a server. + config = Config( + "sdk-key", update_processor_class=td, diagnostic_opt_out=True, send_events=False + ) uninstall_integration(LaunchDarklyIntegration.identifier) if use_global_client: @@ -33,10 +38,6 @@ def test_launchdarkly_integration( client = LDClient(config=config) sentry_init(integrations=[LaunchDarklyIntegration(ld_client=client)]) - # Set test values - td.update(td.flag("hello").variation_for_all(True)) - td.update(td.flag("world").variation_for_all(True)) - # Evaluate client.variation("hello", Context.create("my-org", "organization"), False) client.variation("world", Context.create("user1", "user"), False) @@ -59,7 +60,16 @@ def test_launchdarkly_integration_threaded( sentry_init, capture_events, uninstall_integration ): td = TestData.data_source() - client = LDClient(config=Config("sdk-key", update_processor_class=td)) + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(True)) + client = LDClient( + config=Config( + "sdk-key", + update_processor_class=td, + diagnostic_opt_out=True, # Disable background requests as we aren't using a server. + send_events=False, + ) + ) context = Context.create("user1") uninstall_integration(LaunchDarklyIntegration.identifier) @@ -75,8 +85,6 @@ def task(flag_key): sentry_sdk.set_tag("task_id", flag_key) sentry_sdk.capture_exception(Exception("something wrong!")) - td.update(td.flag("hello").variation_for_all(True)) - td.update(td.flag("world").variation_for_all(False)) # Capture an eval before we split isolation scopes. client.variation("hello", context, False) @@ -104,7 +112,7 @@ def task(flag_key): assert events[2]["contexts"]["flags"] == { "values": [ {"flag": "hello", "result": True}, - {"flag": "world", "result": False}, + {"flag": "world", "result": True}, ] } @@ -118,7 +126,16 @@ def test_launchdarkly_integration_asyncio( asyncio = pytest.importorskip("asyncio") td = TestData.data_source() - client = LDClient(config=Config("sdk-key", update_processor_class=td)) + td.update(td.flag("hello").variation_for_all(True)) + td.update(td.flag("world").variation_for_all(True)) + client = LDClient( + config=Config( + "sdk-key", + update_processor_class=td, + diagnostic_opt_out=True, # Disable background requests as we aren't using a server. + send_events=False, + ) + ) context = Context.create("user1") uninstall_integration(LaunchDarklyIntegration.identifier) @@ -135,8 +152,6 @@ async def task(flag_key): async def runner(): return asyncio.gather(task("world"), task("other")) - td.update(td.flag("hello").variation_for_all(True)) - td.update(td.flag("world").variation_for_all(False)) # Capture an eval before we split isolation scopes. client.variation("hello", context, False) @@ -163,7 +178,7 @@ async def runner(): assert events[2]["contexts"]["flags"] == { "values": [ {"flag": "hello", "result": True}, - {"flag": "world", "result": False}, + {"flag": "world", "result": True}, ] } From 235f5586056acdb1eedf70f73ddea8c962d57301 Mon Sep 17 00:00:00 2001 From: danmr <136265172+danmr@users.noreply.github.com> Date: Tue, 7 Jan 2025 14:21:43 +0300 Subject: [PATCH 083/591] fix: preserve ARQ enqueue_job __kwdefaults__ after patching (#3903) Co-authored-by: Marukhin Daniil --- sentry_sdk/integrations/arq.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index d568714fe2..d61499139b 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -71,6 +71,7 @@ def setup_once(): def patch_enqueue_job(): # type: () -> None old_enqueue_job = ArqRedis.enqueue_job + original_kwdefaults = old_enqueue_job.__kwdefaults__ async def _sentry_enqueue_job(self, function, *args, **kwargs): # type: (ArqRedis, str, *Any, **Any) -> Optional[Job] @@ -83,6 +84,7 @@ async def _sentry_enqueue_job(self, function, *args, **kwargs): ): return await old_enqueue_job(self, function, *args, **kwargs) + _sentry_enqueue_job.__kwdefaults__ = original_kwdefaults ArqRedis.enqueue_job = _sentry_enqueue_job From 7f73c9edcf87b95163437a7aff3a7ed828ec11d9 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 7 Jan 2025 13:38:12 +0100 Subject: [PATCH 084/591] Update test matrix for Sanic (#3904) Fixes the failing test suite. --- .github/workflows/test-integrations-web-2.yml | 2 +- tox.ini | 10 ++++------ 2 files changed, 5 insertions(+), 7 deletions(-) diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index f1fbec6c67..39c1eba535 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/tox.ini b/tox.ini index 717ea62141..37273b2a35 100644 --- a/tox.ini +++ b/tox.ini @@ -247,9 +247,8 @@ envlist = # Sanic {py3.6,py3.7}-sanic-v{0.8} {py3.6,py3.8}-sanic-v{20} - {py3.7,py3.11}-sanic-v{22} - {py3.7,py3.11}-sanic-v{23} - {py3.8,py3.11,py3.12}-sanic-latest + {py3.8,py3.11,py3.12}-sanic-v{24.6} + {py3.9,py3.12,py3.13}-sanic-latest # Spark {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} @@ -652,13 +651,12 @@ deps = # Sanic sanic: websockets<11.0 sanic: aiohttp - sanic-v{22,23}: sanic_testing + sanic-v{24.6}: sanic_testing sanic-latest: sanic_testing {py3.6}-sanic: aiocontextvars==0.2.1 sanic-v0.8: sanic~=0.8.0 sanic-v20: sanic~=20.0 - sanic-v22: sanic~=22.0 - sanic-v23: sanic~=23.0 + sanic-v24.6: sanic~=24.6.0 sanic-latest: sanic # Spark From 8fa6d3d814c76faf72098e4f4ba2d2207e87f5b9 Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Tue, 7 Jan 2025 07:12:47 -0600 Subject: [PATCH 085/591] =?UTF-8?q?Revert=20"ref(flags):=20register=20LD?= =?UTF-8?q?=20hook=20in=20setup=20instead=20of=20init,=20and=20don't=20che?= =?UTF-8?q?c=E2=80=A6"=20(#3900)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Mutating a class attribute on `__init__` violates encapsulation and will lead to strange errors. We need to rethink how we want to implement this before we merge any code. A simple reproduction of the issue: ```python >>> class X: ... y = 0 ... def __init__(self, z): ... self.__class__.y = z ... >>> a = X(1) >>> b = X(2) >>> X.y 2 >>> a.y 2 >>> b.y 2 ``` Reverts getsentry/sentry-python#3890 This reverts commit c3516db643af20396ea981393431646f1a3ef123. Co-authored-by: Anton Pirker --- sentry_sdk/integrations/launchdarkly.py | 14 ++++++------- .../launchdarkly/test_launchdarkly.py | 21 ++++++++++--------- 2 files changed, 18 insertions(+), 17 deletions(-) diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index 066464cc22..a9eef9e1a9 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -20,7 +20,6 @@ class LaunchDarklyIntegration(Integration): identifier = "launchdarkly" - _ld_client = None # type: LDClient | None def __init__(self, ld_client=None): # type: (LDClient | None) -> None @@ -28,19 +27,20 @@ def __init__(self, ld_client=None): :param client: An initialized LDClient instance. If a client is not provided, this integration will attempt to use the shared global instance. """ - self.__class__._ld_client = ld_client - - @staticmethod - def setup_once(): - # type: () -> None try: - client = LaunchDarklyIntegration._ld_client or ldclient.get() + client = ld_client or ldclient.get() except Exception as exc: raise DidNotEnable("Error getting LaunchDarkly client. " + repr(exc)) + if not client.is_initialized(): + raise DidNotEnable("LaunchDarkly client is not initialized.") + # Register the flag collection hook with the LD client. client.add_hook(LaunchDarklyHook()) + @staticmethod + def setup_once(): + # type: () -> None scope = sentry_sdk.get_current_scope() scope.add_error_processor(flag_error_processor) diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index 9b2bbb6b86..20566ce09a 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -183,14 +183,10 @@ async def runner(): } -def test_launchdarkly_integration_did_not_enable(sentry_init, uninstall_integration): - """ - Setup should fail when using global client and ldclient.set_config wasn't called. - - We're accessing ldclient internals to set up this test, so it might break if launchdarkly's - implementation changes. - """ - +def test_launchdarkly_integration_did_not_enable(monkeypatch): + # Client is not passed in and set_config wasn't called. + # TODO: Bad practice to access internals like this. We can skip this test, or remove this + # case entirely (force user to pass in a client instance). ldclient._reset_client() try: ldclient.__lock.lock() @@ -198,6 +194,11 @@ def test_launchdarkly_integration_did_not_enable(sentry_init, uninstall_integrat finally: ldclient.__lock.unlock() - uninstall_integration(LaunchDarklyIntegration.identifier) with pytest.raises(DidNotEnable): - sentry_init(integrations=[LaunchDarklyIntegration()]) + LaunchDarklyIntegration() + + # Client not initialized. + client = LDClient(config=Config("sdk-key")) + monkeypatch.setattr(client, "is_initialized", lambda: False) + with pytest.raises(DidNotEnable): + LaunchDarklyIntegration(ld_client=client) From bf65ede42172dd9bc6718b69e3ea9a9dd417c93d Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Tue, 7 Jan 2025 05:27:08 -0800 Subject: [PATCH 086/591] ref(flags): Beter naming for featureflags module and identifier (#3902) Co-authored-by: Anton Pirker --- sentry_sdk/integrations/{featureflags.py => feature_flags.py} | 4 ++-- .../integrations/{featureflags => feature_flags}/__init__.py | 0 .../test_feature_flags.py} | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) rename sentry_sdk/integrations/{featureflags.py => feature_flags.py} (91%) rename tests/integrations/{featureflags => feature_flags}/__init__.py (100%) rename tests/integrations/{featureflags/test_featureflags.py => feature_flags/test_feature_flags.py} (98%) diff --git a/sentry_sdk/integrations/featureflags.py b/sentry_sdk/integrations/feature_flags.py similarity index 91% rename from sentry_sdk/integrations/featureflags.py rename to sentry_sdk/integrations/feature_flags.py index 46947eec72..2aeabffbfa 100644 --- a/sentry_sdk/integrations/featureflags.py +++ b/sentry_sdk/integrations/feature_flags.py @@ -16,7 +16,7 @@ class FeatureFlagsIntegration(Integration): @example ``` import sentry_sdk - from sentry_sdk.integrations.featureflags import FeatureFlagsIntegration, add_feature_flag + from sentry_sdk.integrations.feature_flags import FeatureFlagsIntegration, add_feature_flag sentry_sdk.init(dsn="my_dsn", integrations=[FeatureFlagsIntegration()]); @@ -25,7 +25,7 @@ class FeatureFlagsIntegration(Integration): ``` """ - identifier = "featureflags" + identifier = "feature_flags" @staticmethod def setup_once(): diff --git a/tests/integrations/featureflags/__init__.py b/tests/integrations/feature_flags/__init__.py similarity index 100% rename from tests/integrations/featureflags/__init__.py rename to tests/integrations/feature_flags/__init__.py diff --git a/tests/integrations/featureflags/test_featureflags.py b/tests/integrations/feature_flags/test_feature_flags.py similarity index 98% rename from tests/integrations/featureflags/test_featureflags.py rename to tests/integrations/feature_flags/test_feature_flags.py index 539e910607..ca6ac16949 100644 --- a/tests/integrations/featureflags/test_featureflags.py +++ b/tests/integrations/feature_flags/test_feature_flags.py @@ -4,7 +4,7 @@ import pytest import sentry_sdk -from sentry_sdk.integrations.featureflags import ( +from sentry_sdk.integrations.feature_flags import ( FeatureFlagsIntegration, add_feature_flag, ) From c6a89d64db965fe0ece6de10df38ab936af8f5e4 Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Tue, 7 Jan 2025 06:17:03 -0800 Subject: [PATCH 087/591] feat(flags): add Unleash feature flagging integration (#3888) Adds an integration for tracking flag evaluations from [Unleash](https://www.getunleash.io/) customers. Implementation Unleash has no native support for evaluation hooks/listeners, unless the user opts in for each flag. Therefore we decided to patch the `is_enabled` and `get_variant` methods on the `UnleashClient` class. The methods are wrapped and the only side effect is writing to Sentry scope, so users shouldn't see any change in behavior. We patch one `UnleashClient` instance instead of the whole class. The reasons for this are described in - https://github.com/getsentry/sentry-python/pull/3895 It's also safer to not modify the unleash import. References - https://develop.sentry.dev/sdk/expected-features/#feature-flags - https://docs.getunleash.io/reference/sdks/python for methods we're patching/wrapping --------- Co-authored-by: Anton Pirker Co-authored-by: Colton Allen --- .github/workflows/test-integrations-misc.yml | 8 + requirements-linting.txt | 1 + .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/unleash.py | 55 ++++ setup.py | 1 + tests/conftest.py | 1 + tests/integrations/unleash/__init__.py | 3 + tests/integrations/unleash/test_unleash.py | 308 ++++++++++++++++++ tests/integrations/unleash/testutils.py | 77 +++++ tox.ini | 17 +- 10 files changed, 468 insertions(+), 4 deletions(-) create mode 100644 sentry_sdk/integrations/unleash.py create mode 100644 tests/integrations/unleash/__init__.py create mode 100644 tests/integrations/unleash/test_unleash.py create mode 100644 tests/integrations/unleash/testutils.py diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 79b7ba020d..d524863423 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -79,6 +79,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-typer-latest" + - name: Test unleash latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-unleash-latest" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -163,6 +167,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-typer" + - name: Test unleash pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-unleash" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/requirements-linting.txt b/requirements-linting.txt index c3f39ecd1f..4227acc26a 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -17,4 +17,5 @@ pre-commit # local linting httpcore openfeature-sdk launchdarkly-server-sdk +UnleashClient typer diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 1b53093c5e..743677daf4 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -133,6 +133,7 @@ "pure_eval", "trytond", "typer", + "unleash", ], } diff --git a/sentry_sdk/integrations/unleash.py b/sentry_sdk/integrations/unleash.py new file mode 100644 index 0000000000..33b0a4b9dc --- /dev/null +++ b/sentry_sdk/integrations/unleash.py @@ -0,0 +1,55 @@ +from functools import wraps +from typing import Any + +import sentry_sdk +from sentry_sdk.flag_utils import flag_error_processor +from sentry_sdk.integrations import Integration, DidNotEnable + +try: + from UnleashClient import UnleashClient +except ImportError: + raise DidNotEnable("UnleashClient is not installed") + + +class UnleashIntegration(Integration): + identifier = "unleash" + + @staticmethod + def setup_once(): + # type: () -> None + # Wrap and patch evaluation methods (instance methods) + old_is_enabled = UnleashClient.is_enabled + old_get_variant = UnleashClient.get_variant + + @wraps(old_is_enabled) + def sentry_is_enabled(self, feature, *args, **kwargs): + # type: (UnleashClient, str, *Any, **Any) -> Any + enabled = old_is_enabled(self, feature, *args, **kwargs) + + # We have no way of knowing what type of unleash feature this is, so we have to treat + # it as a boolean / toggle feature. + flags = sentry_sdk.get_current_scope().flags + flags.set(feature, enabled) + + return enabled + + @wraps(old_get_variant) + def sentry_get_variant(self, feature, *args, **kwargs): + # type: (UnleashClient, str, *Any, **Any) -> Any + variant = old_get_variant(self, feature, *args, **kwargs) + enabled = variant.get("enabled", False) + + # Payloads are not always used as the feature's value for application logic. They + # may be used for metrics or debugging context instead. Therefore, we treat every + # variant as a boolean toggle, using the `enabled` field. + flags = sentry_sdk.get_current_scope().flags + flags.set(feature, enabled) + + return variant + + UnleashClient.is_enabled = sentry_is_enabled # type: ignore + UnleashClient.get_variant = sentry_get_variant # type: ignore + + # Error processor + scope = sentry_sdk.get_current_scope() + scope.add_error_processor(flag_error_processor) diff --git a/setup.py b/setup.py index da3adcab42..9e24d59d21 100644 --- a/setup.py +++ b/setup.py @@ -80,6 +80,7 @@ def get_file_text(file_name): "starlette": ["starlette>=0.19.1"], "starlite": ["starlite>=1.48"], "tornado": ["tornado>=6"], + "unleash": ["UnleashClient>=6.0.1"], }, entry_points={ "opentelemetry_propagator": [ diff --git a/tests/conftest.py b/tests/conftest.py index c0383d94b7..b5ab7aa804 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -10,6 +10,7 @@ import pytest import jsonschema + try: import gevent except ImportError: diff --git a/tests/integrations/unleash/__init__.py b/tests/integrations/unleash/__init__.py new file mode 100644 index 0000000000..33cff3e65a --- /dev/null +++ b/tests/integrations/unleash/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("UnleashClient") diff --git a/tests/integrations/unleash/test_unleash.py b/tests/integrations/unleash/test_unleash.py new file mode 100644 index 0000000000..9a7a3f57bd --- /dev/null +++ b/tests/integrations/unleash/test_unleash.py @@ -0,0 +1,308 @@ +import concurrent.futures as cf +import sys +from random import random +from unittest import mock +from UnleashClient import UnleashClient + +import pytest + +import sentry_sdk +from sentry_sdk.integrations.unleash import UnleashIntegration +from tests.integrations.unleash.testutils import mock_unleash_client + + +def test_is_enabled(sentry_init, capture_events, uninstall_integration): + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) + client.is_enabled("hello") + client.is_enabled("world") + client.is_enabled("other") + + events = capture_events() + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + {"flag": "other", "result": False}, + ] + } + + +def test_get_variant(sentry_init, capture_events, uninstall_integration): + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + client.get_variant("no_payload_feature") + client.get_variant("string_feature") + client.get_variant("json_feature") + client.get_variant("csv_feature") + client.get_variant("number_feature") + client.get_variant("unknown_feature") + + events = capture_events() + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "no_payload_feature", "result": True}, + {"flag": "string_feature", "result": True}, + {"flag": "json_feature", "result": True}, + {"flag": "csv_feature", "result": True}, + {"flag": "number_feature", "result": True}, + {"flag": "unknown_feature", "result": False}, + ] + } + + +def test_is_enabled_threaded(sentry_init, capture_events, uninstall_integration): + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + events = capture_events() + + def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + client.is_enabled(flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + # Capture an eval before we split isolation scopes. + client.is_enabled("hello") + + with cf.ThreadPoolExecutor(max_workers=2) as pool: + pool.map(task, ["world", "other"]) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } + + +def test_get_variant_threaded(sentry_init, capture_events, uninstall_integration): + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + events = capture_events() + + def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + client.get_variant(flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + # Capture an eval before we split isolation scopes. + client.get_variant("hello") + + with cf.ThreadPoolExecutor(max_workers=2) as pool: + pool.map(task, ["no_payload_feature", "other"]) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "no_payload_feature", "result": True}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "other", "result": False}, + ] + } + + +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_is_enabled_asyncio(sentry_init, capture_events, uninstall_integration): + asyncio = pytest.importorskip("asyncio") + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + events = capture_events() + + async def task(flag_key): + with sentry_sdk.isolation_scope(): + client.is_enabled(flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + async def runner(): + return asyncio.gather(task("world"), task("other")) + + # Capture an eval before we split isolation scopes. + client.is_enabled("hello") + + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } + + +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_get_variant_asyncio(sentry_init, capture_events, uninstall_integration): + asyncio = pytest.importorskip("asyncio") + + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + client = UnleashClient() + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + events = capture_events() + + async def task(flag_key): + with sentry_sdk.isolation_scope(): + client.get_variant(flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + async def runner(): + return asyncio.gather(task("no_payload_feature"), task("other")) + + # Capture an eval before we split isolation scopes. + client.get_variant("hello") + + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "no_payload_feature", "result": True}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + {"flag": "other", "result": False}, + ] + } + + +def test_wraps_original(sentry_init, uninstall_integration): + with mock_unleash_client(): + client = UnleashClient() + + mock_is_enabled = mock.Mock(return_value=random() < 0.5) + mock_get_variant = mock.Mock(return_value={"enabled": random() < 0.5}) + client.is_enabled = mock_is_enabled + client.get_variant = mock_get_variant + + uninstall_integration(UnleashIntegration.identifier) + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + + res = client.is_enabled("test-flag", "arg", kwarg=1) + assert res == mock_is_enabled.return_value + assert mock_is_enabled.call_args == ( + ("test-flag", "arg"), + {"kwarg": 1}, + ) + + res = client.get_variant("test-flag", "arg", kwarg=1) + assert res == mock_get_variant.return_value + assert mock_get_variant.call_args == ( + ("test-flag", "arg"), + {"kwarg": 1}, + ) + + +def test_wrapper_attributes(sentry_init, uninstall_integration): + with mock_unleash_client(): + client = UnleashClient() # <- Returns a MockUnleashClient + + original_is_enabled = client.is_enabled + original_get_variant = client.get_variant + + uninstall_integration(UnleashIntegration.identifier) + sentry_init(integrations=[UnleashIntegration()]) # type: ignore + + # Mock clients methods have not lost their qualified names after decoration. + assert client.is_enabled.__name__ == "is_enabled" + assert client.is_enabled.__qualname__ == original_is_enabled.__qualname__ + assert client.get_variant.__name__ == "get_variant" + assert client.get_variant.__qualname__ == original_get_variant.__qualname__ diff --git a/tests/integrations/unleash/testutils.py b/tests/integrations/unleash/testutils.py new file mode 100644 index 0000000000..c424b34c3a --- /dev/null +++ b/tests/integrations/unleash/testutils.py @@ -0,0 +1,77 @@ +from contextlib import contextmanager +from UnleashClient import UnleashClient + + +@contextmanager +def mock_unleash_client(): + """ + Temporarily replaces UnleashClient's methods with mock implementations + for testing. + + This context manager swaps out UnleashClient's __init__, is_enabled, + and get_variant methods with mock versions from MockUnleashClient. + Original methods are restored when exiting the context. + + After mocking the client class the integration can be initialized. + The methods on the mock client class are overridden by the + integration and flag tracking proceeds as expected. + + Example: + with mock_unleash_client(): + client = UnleashClient() # Uses mock implementation + sentry_init(integrations=[UnleashIntegration()]) + """ + old_init = UnleashClient.__init__ + old_is_enabled = UnleashClient.is_enabled + old_get_variant = UnleashClient.get_variant + + UnleashClient.__init__ = MockUnleashClient.__init__ + UnleashClient.is_enabled = MockUnleashClient.is_enabled + UnleashClient.get_variant = MockUnleashClient.get_variant + + yield + + UnleashClient.__init__ = old_init + UnleashClient.is_enabled = old_is_enabled + UnleashClient.get_variant = old_get_variant + + +class MockUnleashClient: + + def __init__(self, *a, **kw): + self.features = { + "hello": True, + "world": False, + } + + self.feature_to_variant = { + "string_feature": { + "name": "variant1", + "enabled": True, + "payload": {"type": "string", "value": "val1"}, + }, + "json_feature": { + "name": "variant1", + "enabled": True, + "payload": {"type": "json", "value": '{"key1": 0.53}'}, + }, + "number_feature": { + "name": "variant1", + "enabled": True, + "payload": {"type": "number", "value": "134.5"}, + }, + "csv_feature": { + "name": "variant1", + "enabled": True, + "payload": {"type": "csv", "value": "abc 123\ncsbq 94"}, + }, + "no_payload_feature": {"name": "variant1", "enabled": True}, + } + + self.disabled_variant = {"name": "disabled", "enabled": False} + + def is_enabled(self, feature, *a, **kw): + return self.features.get(feature, False) + + def get_variant(self, feature, *a, **kw): + return self.feature_to_variant.get(feature, self.disabled_variant) diff --git a/tox.ini b/tox.ini index 37273b2a35..95c09a573e 100644 --- a/tox.ini +++ b/tox.ini @@ -168,6 +168,10 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken + # LaunchDarkly + {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 + {py3.8,py3.12,py3.13}-launchdarkly-latest + # Litestar {py3.8,py3.11}-litestar-v{2.0} {py3.8,py3.11,py3.12}-litestar-v{2.6} @@ -189,10 +193,6 @@ envlist = {py3.8,py3.12,py3.13}-openfeature-v0.7 {py3.8,py3.12,py3.13}-openfeature-latest - # LaunchDarkly - {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 - {py3.8,py3.12,py3.13}-launchdarkly-latest - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -290,6 +290,10 @@ envlist = {py3.7,py3.12,py3.13}-typer-v{0.15} {py3.7,py3.12,py3.13}-typer-latest + # Unleash + {py3.8,py3.12,py3.13}-unleash-v6.0.1 + {py3.8,py3.12,py3.13}-unleash-latest + [testenv] deps = # if you change requirements-testing.txt and your change is not being reflected @@ -571,6 +575,10 @@ deps = launchdarkly-v9.8.0: launchdarkly-server-sdk~=9.8.0 launchdarkly-latest: launchdarkly-server-sdk + # Unleash + unleash-v6.0.1: UnleashClient~=6.0.1 + unleash-latest: UnleashClient + # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -793,6 +801,7 @@ setenv = tornado: TESTPATH=tests/integrations/tornado trytond: TESTPATH=tests/integrations/trytond typer: TESTPATH=tests/integrations/typer + unleash: TESTPATH=tests/integrations/unleash socket: TESTPATH=tests/integrations/socket passenv = From 4432e26a45873080d4eaf20e769bc82f026851bb Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 9 Jan 2025 14:28:39 +0100 Subject: [PATCH 088/591] Small contribution docs update (#3909) --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. The AWS Lambda tests additionally require a maintainer to add a special label, and they will fail until this label is added. --- CONTRIBUTING.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 2f4839f8d7..085dbd6075 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -126,7 +126,7 @@ pytest -rs tests/integrations/flask/ # Replace "flask" with the specific integr ## Releasing a New Version -_(only relevant for Sentry employees)_ +_(only relevant for Python SDK core team)_ ### Prerequisites From be5327356fdae8efc77a9faa9a2ffb0773e80665 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 9 Jan 2025 15:26:50 +0100 Subject: [PATCH 089/591] Centralize minimum version checking (#3910) For [populating tox automatically](https://github.com/getsentry/sentry-python/issues/3808), we need to store min versions of frameworks/libraries in a programmatically accessible place. The obvious place for this would be in each integration; however, since integrations can't be imported unless the respective framework is installed, this couldn't be used from the script (unless we'd always install all requirements of all integrations prior to running it, which takes a non trivial amount of time). So instead I've opted for a central place within `sentry_sdk/integrations/__init__.py`. Note: the min versions probably need updating. Not sure when this was last done, but some of them look quite ancient and we probably don't support them because we'd already dropped the last Python version they'd be able to run on. --- sentry_sdk/integrations/__init__.py | 42 +++++++++++++++++++- sentry_sdk/integrations/aiohttp.py | 8 +--- sentry_sdk/integrations/anthropic.py | 9 +---- sentry_sdk/integrations/ariadne.py | 9 +---- sentry_sdk/integrations/arq.py | 8 +--- sentry_sdk/integrations/asyncpg.py | 12 +++--- sentry_sdk/integrations/boto3.py | 12 +----- sentry_sdk/integrations/bottle.py | 8 +--- sentry_sdk/integrations/celery/__init__.py | 5 +-- sentry_sdk/integrations/clickhouse_driver.py | 7 ++-- sentry_sdk/integrations/django/__init__.py | 6 +-- sentry_sdk/integrations/falcon.py | 9 +---- sentry_sdk/integrations/flask.py | 9 +---- sentry_sdk/integrations/gql.py | 11 ++--- sentry_sdk/integrations/graphene.py | 9 +---- sentry_sdk/integrations/ray.py | 9 +---- sentry_sdk/integrations/rq.py | 10 +---- sentry_sdk/integrations/sanic.py | 12 ++---- sentry_sdk/integrations/sqlalchemy.py | 12 +----- sentry_sdk/integrations/strawberry.py | 11 +---- sentry_sdk/integrations/tornado.py | 5 +-- 21 files changed, 87 insertions(+), 136 deletions(-) diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 12336a939b..683382bb9a 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -111,7 +111,6 @@ def iter_default_integrations(with_auto_enabling_integrations): "sentry_sdk.integrations.tornado.TornadoIntegration", ] - iter_default_integrations = _generate_default_integrations_iterator( integrations=_DEFAULT_INTEGRATIONS, auto_enabling_integrations=_AUTO_ENABLING_INTEGRATIONS, @@ -120,6 +119,30 @@ def iter_default_integrations(with_auto_enabling_integrations): del _generate_default_integrations_iterator +_MIN_VERSIONS = { + "aiohttp": (3, 4), + "anthropic": (0, 16), + "ariadne": (0, 20), + "arq": (0, 23), + "asyncpg": (0, 23), + "boto3": (1, 12), # this is actually the botocore version + "bottle": (0, 12), + "celery": (4, 4, 7), + "clickhouse_driver": (0, 2, 0), + "django": (1, 8), + "falcon": (1, 4), + "flask": (0, 10), + "gql": (3, 4, 1), + "graphene": (3, 3), + "ray": (2, 7, 0), + "rq": (0, 6), + "sanic": (0, 8), + "sqlalchemy": (1, 2), + "strawberry": (0, 209, 5), + "tornado": (6, 0), +} + + def setup_integrations( integrations, with_defaults=True, @@ -195,6 +218,23 @@ def setup_integrations( return integrations +def _check_minimum_version(integration, version, package=None): + # type: (type[Integration], Optional[tuple[int, ...]], Optional[str]) -> None + package = package or integration.identifier + + if version is None: + raise DidNotEnable(f"Unparsable {package} version.") + + min_version = _MIN_VERSIONS.get(integration.identifier) + if min_version is None: + return + + if version < min_version: + raise DidNotEnable( + f"Integration only supports {package} {'.'.join(map(str, min_version))} or newer." + ) + + class DidNotEnable(Exception): # noqa: N818 """ The integration could not be enabled due to a trivial user error like diff --git a/sentry_sdk/integrations/aiohttp.py b/sentry_sdk/integrations/aiohttp.py index d0226bc156..47c1272ae1 100644 --- a/sentry_sdk/integrations/aiohttp.py +++ b/sentry_sdk/integrations/aiohttp.py @@ -7,6 +7,7 @@ from sentry_sdk.consts import OP, SPANSTATUS, SPANDATA from sentry_sdk.integrations import ( _DEFAULT_FAILED_REQUEST_STATUS_CODES, + _check_minimum_version, Integration, DidNotEnable, ) @@ -91,12 +92,7 @@ def setup_once(): # type: () -> None version = parse_version(AIOHTTP_VERSION) - - if version is None: - raise DidNotEnable("Unparsable AIOHTTP version: {}".format(AIOHTTP_VERSION)) - - if version < (3, 4): - raise DidNotEnable("AIOHTTP 3.4 or newer required.") + _check_minimum_version(AioHttpIntegration, version) if not HAS_REAL_CONTEXTVARS: # We better have contextvars or we're going to leak state between diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 87e69a3113..f06d8a14db 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage from sentry_sdk.consts import OP, SPANDATA -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import ( capture_internal_exceptions, @@ -37,12 +37,7 @@ def __init__(self, include_prompts=True): def setup_once(): # type: () -> None version = package_version("anthropic") - - if version is None: - raise DidNotEnable("Unparsable anthropic version.") - - if version < (0, 16): - raise DidNotEnable("anthropic 0.16 or newer required.") + _check_minimum_version(AnthropicIntegration, version) Messages.create = _wrap_message_create(Messages.create) AsyncMessages.create = _wrap_message_create_async(AsyncMessages.create) diff --git a/sentry_sdk/integrations/ariadne.py b/sentry_sdk/integrations/ariadne.py index 70a3424a48..0336140441 100644 --- a/sentry_sdk/integrations/ariadne.py +++ b/sentry_sdk/integrations/ariadne.py @@ -2,7 +2,7 @@ import sentry_sdk from sentry_sdk import get_client, capture_event -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.integrations._wsgi_common import request_body_within_bounds from sentry_sdk.scope import should_send_default_pii @@ -36,12 +36,7 @@ class AriadneIntegration(Integration): def setup_once(): # type: () -> None version = package_version("ariadne") - - if version is None: - raise DidNotEnable("Unparsable ariadne version.") - - if version < (0, 20): - raise DidNotEnable("ariadne 0.20 or newer required.") + _check_minimum_version(AriadneIntegration, version) ignore_logger("ariadne") diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index d61499139b..a2cce8e0ff 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -2,7 +2,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANSTATUS -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_TASK @@ -55,11 +55,7 @@ def setup_once(): except (TypeError, ValueError): version = None - if version is None: - raise DidNotEnable("Unparsable arq version: {}".format(ARQ_VERSION)) - - if version < (0, 23): - raise DidNotEnable("arq 0.23 or newer required.") + _check_minimum_version(ArqIntegration, version) patch_enqueue_job() patch_run_job() diff --git a/sentry_sdk/integrations/asyncpg.py b/sentry_sdk/integrations/asyncpg.py index b05d5615ba..b6b53f4668 100644 --- a/sentry_sdk/integrations/asyncpg.py +++ b/sentry_sdk/integrations/asyncpg.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANDATA -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.tracing import Span from sentry_sdk.tracing_utils import add_query_source, record_sql_queries from sentry_sdk.utils import ( @@ -20,12 +20,6 @@ except ImportError: raise DidNotEnable("asyncpg not installed.") -# asyncpg.__version__ is a string containing the semantic version in the form of ".." -asyncpg_version = parse_version(asyncpg.__version__) - -if asyncpg_version is not None and asyncpg_version < (0, 23, 0): - raise DidNotEnable("asyncpg >= 0.23.0 required") - class AsyncPGIntegration(Integration): identifier = "asyncpg" @@ -37,6 +31,10 @@ def __init__(self, *, record_params: bool = False): @staticmethod def setup_once() -> None: + # asyncpg.__version__ is a string containing the semantic version in the form of ".." + asyncpg_version = parse_version(asyncpg.__version__) + _check_minimum_version(AsyncPGIntegration, asyncpg_version) + asyncpg.Connection.execute = _wrap_execute( asyncpg.Connection.execute, ) diff --git a/sentry_sdk/integrations/boto3.py b/sentry_sdk/integrations/boto3.py index c8da56fb14..0207341f1b 100644 --- a/sentry_sdk/integrations/boto3.py +++ b/sentry_sdk/integrations/boto3.py @@ -2,7 +2,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANDATA -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.tracing import Span from sentry_sdk.utils import ( capture_internal_exceptions, @@ -35,16 +35,8 @@ class Boto3Integration(Integration): @staticmethod def setup_once(): # type: () -> None - version = parse_version(BOTOCORE_VERSION) - - if version is None: - raise DidNotEnable( - "Unparsable botocore version: {}".format(BOTOCORE_VERSION) - ) - - if version < (1, 12): - raise DidNotEnable("Botocore 1.12 or newer is required.") + _check_minimum_version(Boto3Integration, version, "botocore") orig_init = BaseClient.__init__ diff --git a/sentry_sdk/integrations/bottle.py b/sentry_sdk/integrations/bottle.py index a2d6b51033..148b86852e 100644 --- a/sentry_sdk/integrations/bottle.py +++ b/sentry_sdk/integrations/bottle.py @@ -13,6 +13,7 @@ Integration, DidNotEnable, _DEFAULT_FAILED_REQUEST_STATUS_CODES, + _check_minimum_version, ) from sentry_sdk.integrations.wsgi import SentryWsgiMiddleware from sentry_sdk.integrations._wsgi_common import RequestExtractor @@ -72,12 +73,7 @@ def __init__( def setup_once(): # type: () -> None version = parse_version(BOTTLE_VERSION) - - if version is None: - raise DidNotEnable("Unparsable Bottle version: {}".format(BOTTLE_VERSION)) - - if version < (0, 12): - raise DidNotEnable("Bottle 0.12 or newer required.") + _check_minimum_version(BottleIntegration, version) old_app = Bottle.__call__ diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index 9a984de8c3..dc48aac0e6 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -6,7 +6,7 @@ from sentry_sdk import isolation_scope from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP, SPANSTATUS, SPANDATA -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations.celery.beat import ( _patch_beat_apply_entry, _patch_redbeat_maybe_due, @@ -79,8 +79,7 @@ def __init__( @staticmethod def setup_once(): # type: () -> None - if CELERY_VERSION < (4, 4, 7): - raise DidNotEnable("Celery 4.4.7 or newer required.") + _check_minimum_version(CeleryIntegration, CELERY_VERSION) _patch_build_tracer() _patch_task_apply_async() diff --git a/sentry_sdk/integrations/clickhouse_driver.py b/sentry_sdk/integrations/clickhouse_driver.py index daf4c2257c..2561bfad04 100644 --- a/sentry_sdk/integrations/clickhouse_driver.py +++ b/sentry_sdk/integrations/clickhouse_driver.py @@ -1,6 +1,6 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANDATA -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.tracing import Span from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import capture_internal_exceptions, ensure_integration_enabled @@ -34,9 +34,6 @@ def __getitem__(self, _): except ImportError: raise DidNotEnable("clickhouse-driver not installed.") -if clickhouse_driver.VERSION < (0, 2, 0): - raise DidNotEnable("clickhouse-driver >= 0.2.0 required") - class ClickhouseDriverIntegration(Integration): identifier = "clickhouse_driver" @@ -44,6 +41,8 @@ class ClickhouseDriverIntegration(Integration): @staticmethod def setup_once() -> None: + _check_minimum_version(ClickhouseDriverIntegration, clickhouse_driver.VERSION) + # Every query is done using the Connection's `send_query` function clickhouse_driver.connection.Connection.send_query = _wrap_start( clickhouse_driver.connection.Connection.send_query diff --git a/sentry_sdk/integrations/django/__init__.py b/sentry_sdk/integrations/django/__init__.py index e68f0cacef..54bc25675d 100644 --- a/sentry_sdk/integrations/django/__init__.py +++ b/sentry_sdk/integrations/django/__init__.py @@ -22,7 +22,7 @@ transaction_from_function, walk_exception_chain, ) -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.integrations.wsgi import SentryWsgiMiddleware from sentry_sdk.integrations._wsgi_common import ( @@ -154,9 +154,7 @@ def __init__( @staticmethod def setup_once(): # type: () -> None - - if DJANGO_VERSION < (1, 8): - raise DidNotEnable("Django 1.8 or newer is required.") + _check_minimum_version(DjangoIntegration, DJANGO_VERSION) install_sql_hook() # Patch in our custom middleware. diff --git a/sentry_sdk/integrations/falcon.py b/sentry_sdk/integrations/falcon.py index ce771d16e7..ddedcb10de 100644 --- a/sentry_sdk/integrations/falcon.py +++ b/sentry_sdk/integrations/falcon.py @@ -1,5 +1,5 @@ import sentry_sdk -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations._wsgi_common import RequestExtractor from sentry_sdk.integrations.wsgi import SentryWsgiMiddleware from sentry_sdk.tracing import SOURCE_FOR_STYLE @@ -135,12 +135,7 @@ def setup_once(): # type: () -> None version = parse_version(FALCON_VERSION) - - if version is None: - raise DidNotEnable("Unparsable Falcon version: {}".format(FALCON_VERSION)) - - if version < (1, 4): - raise DidNotEnable("Falcon 1.4 or newer required.") + _check_minimum_version(FalconIntegration, version) _patch_wsgi_app() _patch_handle_exception() diff --git a/sentry_sdk/integrations/flask.py b/sentry_sdk/integrations/flask.py index 128301ddb4..45b4f0b2b1 100644 --- a/sentry_sdk/integrations/flask.py +++ b/sentry_sdk/integrations/flask.py @@ -1,5 +1,5 @@ import sentry_sdk -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations._wsgi_common import ( DEFAULT_HTTP_METHODS_TO_CAPTURE, RequestExtractor, @@ -73,12 +73,7 @@ def __init__( def setup_once(): # type: () -> None version = package_version("flask") - - if version is None: - raise DidNotEnable("Unparsable Flask version.") - - if version < (0, 10): - raise DidNotEnable("Flask 0.10 or newer is required.") + _check_minimum_version(FlaskIntegration, version) before_render_template.connect(_add_sentry_trace) request_started.connect(_request_started) diff --git a/sentry_sdk/integrations/gql.py b/sentry_sdk/integrations/gql.py index 5074442986..d5341d2cf6 100644 --- a/sentry_sdk/integrations/gql.py +++ b/sentry_sdk/integrations/gql.py @@ -5,7 +5,7 @@ parse_version, ) -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii try: @@ -24,8 +24,6 @@ EventDataType = Dict[str, Union[str, Tuple[VariableDefinitionNode, ...]]] -MIN_GQL_VERSION = (3, 4, 1) - class GQLIntegration(Integration): identifier = "gql" @@ -34,11 +32,8 @@ class GQLIntegration(Integration): def setup_once(): # type: () -> None gql_version = parse_version(gql.__version__) - if gql_version is None or gql_version < MIN_GQL_VERSION: - raise DidNotEnable( - "GQLIntegration is only supported for GQL versions %s and above." - % ".".join(str(num) for num in MIN_GQL_VERSION) - ) + _check_minimum_version(GQLIntegration, gql_version) + _patch_execute() diff --git a/sentry_sdk/integrations/graphene.py b/sentry_sdk/integrations/graphene.py index 03731dcaaa..198aea50d2 100644 --- a/sentry_sdk/integrations/graphene.py +++ b/sentry_sdk/integrations/graphene.py @@ -2,7 +2,7 @@ import sentry_sdk from sentry_sdk.consts import OP -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import ( capture_internal_exceptions, @@ -34,12 +34,7 @@ class GrapheneIntegration(Integration): def setup_once(): # type: () -> None version = package_version("graphene") - - if version is None: - raise DidNotEnable("Unparsable graphene version.") - - if version < (3, 3): - raise DidNotEnable("graphene 3.3 or newer required.") + _check_minimum_version(GrapheneIntegration, version) _patch_graphql() diff --git a/sentry_sdk/integrations/ray.py b/sentry_sdk/integrations/ray.py index 2f5086ed92..24a28c307f 100644 --- a/sentry_sdk/integrations/ray.py +++ b/sentry_sdk/integrations/ray.py @@ -3,7 +3,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANSTATUS -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.tracing import TRANSACTION_SOURCE_TASK from sentry_sdk.utils import ( event_from_exception, @@ -136,11 +136,6 @@ class RayIntegration(Integration): def setup_once(): # type: () -> None version = package_version("ray") - - if version is None: - raise DidNotEnable("Unparsable ray version: {}".format(version)) - - if version < (2, 7, 0): - raise DidNotEnable("Ray 2.7.0 or newer required") + _check_minimum_version(RayIntegration, version) _patch_ray_remote() diff --git a/sentry_sdk/integrations/rq.py b/sentry_sdk/integrations/rq.py index 462f3ad30a..d4fca6a33b 100644 --- a/sentry_sdk/integrations/rq.py +++ b/sentry_sdk/integrations/rq.py @@ -3,7 +3,7 @@ import sentry_sdk from sentry_sdk.consts import OP from sentry_sdk.api import continue_trace -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.tracing import TRANSACTION_SOURCE_TASK from sentry_sdk.utils import ( @@ -41,14 +41,8 @@ class RqIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - version = parse_version(RQ_VERSION) - - if version is None: - raise DidNotEnable("Unparsable RQ version: {}".format(RQ_VERSION)) - - if version < (0, 6): - raise DidNotEnable("RQ 0.6 or newer is required.") + _check_minimum_version(RqIntegration, version) old_perform_job = Worker.perform_job diff --git a/sentry_sdk/integrations/sanic.py b/sentry_sdk/integrations/sanic.py index 26e29cb78c..dfcc299d42 100644 --- a/sentry_sdk/integrations/sanic.py +++ b/sentry_sdk/integrations/sanic.py @@ -6,7 +6,7 @@ import sentry_sdk from sentry_sdk import continue_trace from sentry_sdk.consts import OP -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations._wsgi_common import RequestExtractor, _filter_headers from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT, TRANSACTION_SOURCE_URL @@ -73,14 +73,8 @@ def __init__(self, unsampled_statuses=frozenset({404})): @staticmethod def setup_once(): # type: () -> None - SanicIntegration.version = parse_version(SANIC_VERSION) - - if SanicIntegration.version is None: - raise DidNotEnable("Unparsable Sanic version: {}".format(SANIC_VERSION)) - - if SanicIntegration.version < (0, 8): - raise DidNotEnable("Sanic 0.8 or newer required.") + _check_minimum_version(SanicIntegration, SanicIntegration.version) if not HAS_REAL_CONTEXTVARS: # We better have contextvars or we're going to leak state between @@ -102,7 +96,7 @@ def setup_once(): # https://github.com/huge-success/sanic/issues/1332 ignore_logger("root") - if SanicIntegration.version < (21, 9): + if SanicIntegration.version is not None and SanicIntegration.version < (21, 9): _setup_legacy_sanic() return diff --git a/sentry_sdk/integrations/sqlalchemy.py b/sentry_sdk/integrations/sqlalchemy.py index 0a54108e75..068d373053 100644 --- a/sentry_sdk/integrations/sqlalchemy.py +++ b/sentry_sdk/integrations/sqlalchemy.py @@ -1,5 +1,5 @@ from sentry_sdk.consts import SPANSTATUS, SPANDATA -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.tracing_utils import add_query_source, record_sql_queries from sentry_sdk.utils import ( capture_internal_exceptions, @@ -31,16 +31,8 @@ class SqlalchemyIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - version = parse_version(SQLALCHEMY_VERSION) - - if version is None: - raise DidNotEnable( - "Unparsable SQLAlchemy version: {}".format(SQLALCHEMY_VERSION) - ) - - if version < (1, 2): - raise DidNotEnable("SQLAlchemy 1.2 or newer required.") + _check_minimum_version(SqlalchemyIntegration, version) listen(Engine, "before_cursor_execute", _before_cursor_execute) listen(Engine, "after_cursor_execute", _after_cursor_execute) diff --git a/sentry_sdk/integrations/strawberry.py b/sentry_sdk/integrations/strawberry.py index 58860a633b..d27e0eaf1c 100644 --- a/sentry_sdk/integrations/strawberry.py +++ b/sentry_sdk/integrations/strawberry.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.consts import OP -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT @@ -75,14 +75,7 @@ def __init__(self, async_execution=None): def setup_once(): # type: () -> None version = package_version("strawberry-graphql") - - if version is None: - raise DidNotEnable( - "Unparsable strawberry-graphql version: {}".format(version) - ) - - if version < (0, 209, 5): - raise DidNotEnable("strawberry-graphql 0.209.5 or newer required.") + _check_minimum_version(StrawberryIntegration, version, "strawberry-graphql") _patch_schema_init() _patch_execute() diff --git a/sentry_sdk/integrations/tornado.py b/sentry_sdk/integrations/tornado.py index f1bd196261..b9e465c7c7 100644 --- a/sentry_sdk/integrations/tornado.py +++ b/sentry_sdk/integrations/tornado.py @@ -18,7 +18,7 @@ capture_internal_exceptions, transaction_from_function, ) -from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations._wsgi_common import ( RequestExtractor, _filter_headers, @@ -52,8 +52,7 @@ class TornadoIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - if TORNADO_VERSION < (6, 0): - raise DidNotEnable("Tornado 6.0+ required") + _check_minimum_version(TornadoIntegration, TORNADO_VERSION) if not HAS_REAL_CONTEXTVARS: # Tornado is async. We better have contextvars or we're going to leak From fa241c3425e446878f173407fd7358f38d8bd529 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 9 Jan 2025 18:07:32 +0100 Subject: [PATCH 090/591] Treat potel-base as release branch in CI (#3912) ...and remove `sentry-sdk-2.0` from the CI yamls. --- .github/workflows/ci.yml | 2 +- .github/workflows/codeql-analysis.yml | 2 +- .github/workflows/enforce-license-compliance.yml | 2 +- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-aws.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 2 +- .github/workflows/test-integrations-tasks.yml | 2 +- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/split_tox_gh_actions/templates/base.jinja | 2 +- 15 files changed, 15 insertions(+), 15 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 7ef6604e39..e8931e229e 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -5,7 +5,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index e362d1e620..d824757ee9 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -15,7 +15,7 @@ on: push: branches: - master - - sentry-sdk-2.0 + - potel-base pull_request: schedule: - cron: '18 18 * * 3' diff --git a/.github/workflows/enforce-license-compliance.yml b/.github/workflows/enforce-license-compliance.yml index ef79ed112b..5517e5347f 100644 --- a/.github/workflows/enforce-license-compliance.yml +++ b/.github/workflows/enforce-license-compliance.yml @@ -6,7 +6,7 @@ on: - master - main - release/* - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 2fd6995a5f..6e06e6067c 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index f83e3379f6..eae488776a 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base # XXX: We are using `pull_request_target` instead of `pull_request` because we want # this to run on forks with access to the secrets necessary to run the test suite. # Prefer to use `pull_request` when possible. diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 9e34dc6b2b..af089caede 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index f1806597af..d9e08bbeb8 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index d9bea0611b..f612b8fb14 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 7138204e16..d239b2ed6c 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index d524863423..5747448442 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 1b9ee3c529..ab1c5b0658 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 0f97146d6d..8ecc7ab598 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 53206f764f..2dc5f361de 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 39c1eba535..2b3204ae80 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/scripts/split_tox_gh_actions/templates/base.jinja b/scripts/split_tox_gh_actions/templates/base.jinja index 16dbc04a76..e69b6f9134 100644 --- a/scripts/split_tox_gh_actions/templates/base.jinja +++ b/scripts/split_tox_gh_actions/templates/base.jinja @@ -11,7 +11,7 @@ on: branches: - master - release/** - - sentry-sdk-2.0 + - potel-base {% if needs_github_secrets %} # XXX: We are using `pull_request_target` instead of `pull_request` because we want From 9f9ff345c6054e0623a293c1f90e6e590ceb8a9f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 13 Jan 2025 10:13:26 +0100 Subject: [PATCH 091/591] tests: Create a separate group for feature flag suites (#3911) Take feature flag tests out of Misc and into their own new Flags group. Also move Tasks down in the `GROUPS` dict do that it's alphabetized (except for misc which is at the bottom). --- .github/workflows/test-integrations-flags.yml | 163 ++++++++++++++++++ .github/workflows/test-integrations-misc.yml | 24 --- .../split_tox_gh_actions.py | 28 +-- 3 files changed, 178 insertions(+), 37 deletions(-) create mode 100644 .github/workflows/test-integrations-flags.yml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml new file mode 100644 index 0000000000..0460868473 --- /dev/null +++ b/.github/workflows/test-integrations-flags.yml @@ -0,0 +1,163 @@ +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja +name: Test Flags +on: + push: + branches: + - master + - release/** + - potel-base + pull_request: +# Cancel in progress workflows on pull_requests. +# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value +concurrency: + group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} + cancel-in-progress: true +permissions: + contents: read +env: + BUILD_CACHE_KEY: ${{ github.sha }} + CACHED_BUILD_PATHS: | + ${{ github.workspace }}/dist-serverless +jobs: + test-flags-latest: + name: Flags (latest) + timeout-minutes: 30 + runs-on: ${{ matrix.os }} + strategy: + fail-fast: false + matrix: + python-version: ["3.8","3.12","3.13"] + # python3.6 reached EOL and is no longer being supported on + # new versions of hosted runners on Github Actions + # ubuntu-20.04 is the last version that supported python3.6 + # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 + os: [ubuntu-20.04] + steps: + - uses: actions/checkout@v4.2.2 + - uses: actions/setup-python@v5 + with: + python-version: ${{ matrix.python-version }} + allow-prereleases: true + - name: Setup Test Env + run: | + pip install "coverage[toml]" tox + - name: Erase coverage + run: | + coverage erase + - name: Test launchdarkly latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-launchdarkly-latest" + - name: Test openfeature latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature-latest" + - name: Test unleash latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-unleash-latest" + - name: Generate coverage XML (Python 3.6) + if: ${{ !cancelled() && matrix.python-version == '3.6' }} + run: | + export COVERAGE_RCFILE=.coveragerc36 + coverage combine .coverage-sentry-* + coverage xml --ignore-errors + - name: Generate coverage XML + if: ${{ !cancelled() && matrix.python-version != '3.6' }} + run: | + coverage combine .coverage-sentry-* + coverage xml + - name: Upload coverage to Codecov + if: ${{ !cancelled() }} + uses: codecov/codecov-action@v5.1.2 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: coverage.xml + # make sure no plugins alter our coverage reports + plugin: noop + verbose: true + - name: Upload test results to Codecov + if: ${{ !cancelled() }} + uses: codecov/test-results-action@v1 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: .junitxml + verbose: true + test-flags-pinned: + name: Flags (pinned) + timeout-minutes: 30 + runs-on: ${{ matrix.os }} + strategy: + fail-fast: false + matrix: + python-version: ["3.8","3.12","3.13"] + # python3.6 reached EOL and is no longer being supported on + # new versions of hosted runners on Github Actions + # ubuntu-20.04 is the last version that supported python3.6 + # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 + os: [ubuntu-20.04] + steps: + - uses: actions/checkout@v4.2.2 + - uses: actions/setup-python@v5 + with: + python-version: ${{ matrix.python-version }} + allow-prereleases: true + - name: Setup Test Env + run: | + pip install "coverage[toml]" tox + - name: Erase coverage + run: | + coverage erase + - name: Test launchdarkly pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-launchdarkly" + - name: Test openfeature pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openfeature" + - name: Test unleash pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-unleash" + - name: Generate coverage XML (Python 3.6) + if: ${{ !cancelled() && matrix.python-version == '3.6' }} + run: | + export COVERAGE_RCFILE=.coveragerc36 + coverage combine .coverage-sentry-* + coverage xml --ignore-errors + - name: Generate coverage XML + if: ${{ !cancelled() && matrix.python-version != '3.6' }} + run: | + coverage combine .coverage-sentry-* + coverage xml + - name: Upload coverage to Codecov + if: ${{ !cancelled() }} + uses: codecov/codecov-action@v5.1.2 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: coverage.xml + # make sure no plugins alter our coverage reports + plugin: noop + verbose: true + - name: Upload test results to Codecov + if: ${{ !cancelled() }} + uses: codecov/test-results-action@v1 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: .junitxml + verbose: true + check_required_tests: + name: All pinned Flags tests passed + needs: test-flags-pinned + # Always run this, even if a dependent job failed + if: always() + runs-on: ubuntu-20.04 + steps: + - name: Check for failures + if: contains(needs.test-flags-pinned.result, 'failure') || contains(needs.test-flags-pinned.result, 'skipped') + run: | + echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 5747448442..9461ea506c 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -47,18 +47,10 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test launchdarkly latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-launchdarkly-latest" - name: Test loguru latest run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-loguru-latest" - - name: Test openfeature latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature-latest" - name: Test opentelemetry latest run: | set -x # print commands that are executed @@ -79,10 +71,6 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-typer-latest" - - name: Test unleash latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-unleash-latest" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -135,18 +123,10 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test launchdarkly pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-launchdarkly" - name: Test loguru pinned run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-loguru" - - name: Test openfeature pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openfeature" - name: Test opentelemetry pinned run: | set -x # print commands that are executed @@ -167,10 +147,6 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-typer" - - name: Test unleash pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-unleash" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 743677daf4..1537ad8389 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -76,16 +76,6 @@ "cloud_resource_context", "gcp", ], - "Tasks": [ - "arq", - "beam", - "celery", - "dramatiq", - "huey", - "ray", - "rq", - "spark", - ], "DBs": [ "asyncpg", "clickhouse_driver", @@ -94,6 +84,11 @@ "redis_py_cluster_legacy", "sqlalchemy", ], + "Flags": [ + "launchdarkly", + "openfeature", + "unleash", + ], "GraphQL": [ "ariadne", "gql", @@ -106,6 +101,16 @@ "httpx", "requests", ], + "Tasks": [ + "arq", + "beam", + "celery", + "dramatiq", + "huey", + "ray", + "rq", + "spark", + ], "Web 1": [ "django", "flask", @@ -125,15 +130,12 @@ "tornado", ], "Misc": [ - "launchdarkly", "loguru", - "openfeature", "opentelemetry", "potel", "pure_eval", "trytond", "typer", - "unleash", ], } From 288f69a962e4ae9e929ae1116ec683297a0a416a Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 13 Jan 2025 16:54:29 +0100 Subject: [PATCH 092/591] Moved adding of `flags` context into Scope (#3917) Using an error_processor to read data from the scope to add to the event is an anti-pattern. Moving this into `Scope.apply_to_event()`. This PR: - moves code that adds flags to an event from an error processor into the `Scope` class - moves `add_feature_flag()` function from `sentry_sdk.integrations.feature_flags` into `sentry_sdk.feature_flags` --- .../{flag_utils.py => feature_flags.py} | 20 ++++--- sentry_sdk/integrations/feature_flags.py | 44 --------------- sentry_sdk/integrations/launchdarkly.py | 4 +- sentry_sdk/integrations/openfeature.py | 4 -- sentry_sdk/integrations/unleash.py | 5 -- sentry_sdk/scope.py | 17 +++++- tests/integrations/feature_flags/__init__.py | 0 .../feature_flags => }/test_feature_flags.py | 56 +++++++++++++++---- tests/test_flag_utils.py | 43 -------------- 9 files changed, 74 insertions(+), 119 deletions(-) rename sentry_sdk/{flag_utils.py => feature_flags.py} (67%) delete mode 100644 sentry_sdk/integrations/feature_flags.py delete mode 100644 tests/integrations/feature_flags/__init__.py rename tests/{integrations/feature_flags => }/test_feature_flags.py (75%) delete mode 100644 tests/test_flag_utils.py diff --git a/sentry_sdk/flag_utils.py b/sentry_sdk/feature_flags.py similarity index 67% rename from sentry_sdk/flag_utils.py rename to sentry_sdk/feature_flags.py index cf4800e855..1187c2fa12 100644 --- a/sentry_sdk/flag_utils.py +++ b/sentry_sdk/feature_flags.py @@ -1,11 +1,10 @@ -from typing import TYPE_CHECKING - import sentry_sdk from sentry_sdk._lru_cache import LRUCache +from typing import TYPE_CHECKING + if TYPE_CHECKING: - from typing import TypedDict, Optional - from sentry_sdk._types import Event, ExcInfo + from typing import TypedDict FlagData = TypedDict("FlagData", {"flag": str, "result": bool}) @@ -33,8 +32,11 @@ def set(self, flag, result): self.buffer.set(flag, result) -def flag_error_processor(event, exc_info): - # type: (Event, ExcInfo) -> Optional[Event] - scope = sentry_sdk.get_current_scope() - event["contexts"]["flags"] = {"values": scope.flags.get()} - return event +def add_feature_flag(flag, result): + # type: (str, bool) -> None + """ + Records a flag and its value to be sent on subsequent error events. + We recommend you do this on flag evaluations. Flags are buffered per Sentry scope. + """ + flags = sentry_sdk.get_current_scope().flags + flags.set(flag, result) diff --git a/sentry_sdk/integrations/feature_flags.py b/sentry_sdk/integrations/feature_flags.py deleted file mode 100644 index 2aeabffbfa..0000000000 --- a/sentry_sdk/integrations/feature_flags.py +++ /dev/null @@ -1,44 +0,0 @@ -from sentry_sdk.flag_utils import flag_error_processor - -import sentry_sdk -from sentry_sdk.integrations import Integration - - -class FeatureFlagsIntegration(Integration): - """ - Sentry integration for capturing feature flags on error events. To manually buffer flag data, - call `integrations.featureflags.add_feature_flag`. We recommend you do this on each flag - evaluation. - - See the [feature flag documentation](https://develop.sentry.dev/sdk/expected-features/#feature-flags) - for more information. - - @example - ``` - import sentry_sdk - from sentry_sdk.integrations.feature_flags import FeatureFlagsIntegration, add_feature_flag - - sentry_sdk.init(dsn="my_dsn", integrations=[FeatureFlagsIntegration()]); - - add_feature_flag('my-flag', true); - sentry_sdk.capture_exception(Exception('broke')); // 'my-flag' should be captured on this Sentry event. - ``` - """ - - identifier = "feature_flags" - - @staticmethod - def setup_once(): - # type: () -> None - scope = sentry_sdk.get_current_scope() - scope.add_error_processor(flag_error_processor) - - -def add_feature_flag(flag, result): - # type: (str, bool) -> None - """ - Records a flag and its value to be sent on subsequent error events by FeatureFlagsIntegration. - We recommend you do this on flag evaluations. Flags are buffered per Sentry scope. - """ - flags = sentry_sdk.get_current_scope().flags - flags.set(flag, result) diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index a9eef9e1a9..cb9e911463 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -2,7 +2,6 @@ import sentry_sdk from sentry_sdk.integrations import DidNotEnable, Integration -from sentry_sdk.flag_utils import flag_error_processor try: import ldclient @@ -41,8 +40,7 @@ def __init__(self, ld_client=None): @staticmethod def setup_once(): # type: () -> None - scope = sentry_sdk.get_current_scope() - scope.add_error_processor(flag_error_processor) + pass class LaunchDarklyHook(Hook): diff --git a/sentry_sdk/integrations/openfeature.py b/sentry_sdk/integrations/openfeature.py index 18f968a703..bf66b94e8b 100644 --- a/sentry_sdk/integrations/openfeature.py +++ b/sentry_sdk/integrations/openfeature.py @@ -2,7 +2,6 @@ import sentry_sdk from sentry_sdk.integrations import DidNotEnable, Integration -from sentry_sdk.flag_utils import flag_error_processor try: from openfeature import api @@ -21,9 +20,6 @@ class OpenFeatureIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - scope = sentry_sdk.get_current_scope() - scope.add_error_processor(flag_error_processor) - # Register the hook within the global openfeature hooks list. api.add_hooks(hooks=[OpenFeatureHook()]) diff --git a/sentry_sdk/integrations/unleash.py b/sentry_sdk/integrations/unleash.py index 33b0a4b9dc..442ec39d0f 100644 --- a/sentry_sdk/integrations/unleash.py +++ b/sentry_sdk/integrations/unleash.py @@ -2,7 +2,6 @@ from typing import Any import sentry_sdk -from sentry_sdk.flag_utils import flag_error_processor from sentry_sdk.integrations import Integration, DidNotEnable try: @@ -49,7 +48,3 @@ def sentry_get_variant(self, feature, *args, **kwargs): UnleashClient.is_enabled = sentry_is_enabled # type: ignore UnleashClient.get_variant = sentry_get_variant # type: ignore - - # Error processor - scope = sentry_sdk.get_current_scope() - scope.add_error_processor(flag_error_processor) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index cf72fabdd1..ab0f1f4156 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -11,7 +11,7 @@ from sentry_sdk.attachments import Attachment from sentry_sdk.consts import DEFAULT_MAX_BREADCRUMBS, FALSE_VALUES, INSTRUMENTER -from sentry_sdk.flag_utils import FlagBuffer, DEFAULT_FLAG_CAPACITY +from sentry_sdk.feature_flags import FlagBuffer, DEFAULT_FLAG_CAPACITY from sentry_sdk.profiler.continuous_profiler import try_autostart_continuous_profiler from sentry_sdk.profiler.transaction_profiler import Profile from sentry_sdk.session import Session @@ -1378,6 +1378,14 @@ def _apply_contexts_to_event(self, event, hint, options): else: contexts["trace"] = self.get_trace_context() + def _apply_flags_to_event(self, event, hint, options): + # type: (Event, Hint, Optional[Dict[str, Any]]) -> None + flags = self.flags.get() + if len(flags) > 0: + event.setdefault("contexts", {}).setdefault("flags", {}).update( + {"values": flags} + ) + def _drop(self, cause, ty): # type: (Any, str) -> Optional[Any] logger.info("%s (%s) dropped event", ty, cause) @@ -1476,6 +1484,7 @@ def apply_to_event( if not is_transaction and not is_check_in: self._apply_breadcrumbs_to_event(event, hint, options) + self._apply_flags_to_event(event, hint, options) event = self.run_error_processors(event, hint) if event is None: @@ -1518,6 +1527,12 @@ def update_from_scope(self, scope): self._propagation_context = scope._propagation_context if scope._session: self._session = scope._session + if scope._flags: + if not self._flags: + self._flags = deepcopy(scope._flags) + else: + for flag in scope._flags.get(): + self._flags.set(flag["flag"], flag["result"]) def update_from_kwargs( self, diff --git a/tests/integrations/feature_flags/__init__.py b/tests/integrations/feature_flags/__init__.py deleted file mode 100644 index e69de29bb2..0000000000 diff --git a/tests/integrations/feature_flags/test_feature_flags.py b/tests/test_feature_flags.py similarity index 75% rename from tests/integrations/feature_flags/test_feature_flags.py rename to tests/test_feature_flags.py index ca6ac16949..14d74cb04b 100644 --- a/tests/integrations/feature_flags/test_feature_flags.py +++ b/tests/test_feature_flags.py @@ -4,15 +4,11 @@ import pytest import sentry_sdk -from sentry_sdk.integrations.feature_flags import ( - FeatureFlagsIntegration, - add_feature_flag, -) +from sentry_sdk.feature_flags import add_feature_flag, FlagBuffer def test_featureflags_integration(sentry_init, capture_events, uninstall_integration): - uninstall_integration(FeatureFlagsIntegration.identifier) - sentry_init(integrations=[FeatureFlagsIntegration()]) + sentry_init() add_feature_flag("hello", False) add_feature_flag("world", True) @@ -34,8 +30,7 @@ def test_featureflags_integration(sentry_init, capture_events, uninstall_integra def test_featureflags_integration_threaded( sentry_init, capture_events, uninstall_integration ): - uninstall_integration(FeatureFlagsIntegration.identifier) - sentry_init(integrations=[FeatureFlagsIntegration()]) + sentry_init() events = capture_events() # Capture an eval before we split isolation scopes. @@ -86,8 +81,7 @@ def test_featureflags_integration_asyncio( ): asyncio = pytest.importorskip("asyncio") - uninstall_integration(FeatureFlagsIntegration.identifier) - sentry_init(integrations=[FeatureFlagsIntegration()]) + sentry_init() events = capture_events() # Capture an eval before we split isolation scopes. @@ -131,3 +125,45 @@ async def runner(): {"flag": "world", "result": False}, ] } + + +def test_flag_tracking(): + """Assert the ring buffer works.""" + buffer = FlagBuffer(capacity=3) + buffer.set("a", True) + flags = buffer.get() + assert len(flags) == 1 + assert flags == [{"flag": "a", "result": True}] + + buffer.set("b", True) + flags = buffer.get() + assert len(flags) == 2 + assert flags == [{"flag": "a", "result": True}, {"flag": "b", "result": True}] + + buffer.set("c", True) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "a", "result": True}, + {"flag": "b", "result": True}, + {"flag": "c", "result": True}, + ] + + buffer.set("d", False) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "b", "result": True}, + {"flag": "c", "result": True}, + {"flag": "d", "result": False}, + ] + + buffer.set("e", False) + buffer.set("f", False) + flags = buffer.get() + assert len(flags) == 3 + assert flags == [ + {"flag": "d", "result": False}, + {"flag": "e", "result": False}, + {"flag": "f", "result": False}, + ] diff --git a/tests/test_flag_utils.py b/tests/test_flag_utils.py deleted file mode 100644 index 3fa4f3abfe..0000000000 --- a/tests/test_flag_utils.py +++ /dev/null @@ -1,43 +0,0 @@ -from sentry_sdk.flag_utils import FlagBuffer - - -def test_flag_tracking(): - """Assert the ring buffer works.""" - buffer = FlagBuffer(capacity=3) - buffer.set("a", True) - flags = buffer.get() - assert len(flags) == 1 - assert flags == [{"flag": "a", "result": True}] - - buffer.set("b", True) - flags = buffer.get() - assert len(flags) == 2 - assert flags == [{"flag": "a", "result": True}, {"flag": "b", "result": True}] - - buffer.set("c", True) - flags = buffer.get() - assert len(flags) == 3 - assert flags == [ - {"flag": "a", "result": True}, - {"flag": "b", "result": True}, - {"flag": "c", "result": True}, - ] - - buffer.set("d", False) - flags = buffer.get() - assert len(flags) == 3 - assert flags == [ - {"flag": "b", "result": True}, - {"flag": "c", "result": True}, - {"flag": "d", "result": False}, - ] - - buffer.set("e", False) - buffer.set("f", False) - flags = buffer.get() - assert len(flags) == 3 - assert flags == [ - {"flag": "d", "result": False}, - {"flag": "e", "result": False}, - {"flag": "f", "result": False}, - ] From 2ee194c0d4fac809b40ef81d90ae859998962afa Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Tue, 14 Jan 2025 00:00:55 -0800 Subject: [PATCH 093/591] feat(flags): remove Unleash get_variant patching code (#3914) Follow-up to https://github.com/getsentry/sentry-python/pull/3888 The original PR patched 2 methods used for evaluating feature flags, `is_enabled` (simple toggle on/off) and `get_variant` (returns a dict of metadata, see https://docs.getunleash.io/reference/sdks/python#getting-a-variant). We want to remove all `get_variant` code since we only support boolean flag evals atm. It seems like the main usecase for variants is reading payloads (non-bool) for A/B/multivariate testing. This could lead to a lot of extraneous flags, so until it is requested and/or we support non-bool values, let's not patch this method. --- sentry_sdk/integrations/unleash.py | 16 --- tests/integrations/unleash/test_unleash.py | 156 +-------------------- tests/integrations/unleash/testutils.py | 36 +---- 3 files changed, 9 insertions(+), 199 deletions(-) diff --git a/sentry_sdk/integrations/unleash.py b/sentry_sdk/integrations/unleash.py index 442ec39d0f..c7108394d0 100644 --- a/sentry_sdk/integrations/unleash.py +++ b/sentry_sdk/integrations/unleash.py @@ -18,7 +18,6 @@ def setup_once(): # type: () -> None # Wrap and patch evaluation methods (instance methods) old_is_enabled = UnleashClient.is_enabled - old_get_variant = UnleashClient.get_variant @wraps(old_is_enabled) def sentry_is_enabled(self, feature, *args, **kwargs): @@ -32,19 +31,4 @@ def sentry_is_enabled(self, feature, *args, **kwargs): return enabled - @wraps(old_get_variant) - def sentry_get_variant(self, feature, *args, **kwargs): - # type: (UnleashClient, str, *Any, **Any) -> Any - variant = old_get_variant(self, feature, *args, **kwargs) - enabled = variant.get("enabled", False) - - # Payloads are not always used as the feature's value for application logic. They - # may be used for metrics or debugging context instead. Therefore, we treat every - # variant as a boolean toggle, using the `enabled` field. - flags = sentry_sdk.get_current_scope().flags - flags.set(feature, enabled) - - return variant - UnleashClient.is_enabled = sentry_is_enabled # type: ignore - UnleashClient.get_variant = sentry_get_variant # type: ignore diff --git a/tests/integrations/unleash/test_unleash.py b/tests/integrations/unleash/test_unleash.py index 9a7a3f57bd..379abba8f6 100644 --- a/tests/integrations/unleash/test_unleash.py +++ b/tests/integrations/unleash/test_unleash.py @@ -15,7 +15,7 @@ def test_is_enabled(sentry_init, capture_events, uninstall_integration): uninstall_integration(UnleashIntegration.identifier) with mock_unleash_client(): - client = UnleashClient() + client = UnleashClient() # type: ignore[arg-type] sentry_init(integrations=[UnleashIntegration()]) client.is_enabled("hello") client.is_enabled("world") @@ -34,41 +34,12 @@ def test_is_enabled(sentry_init, capture_events, uninstall_integration): } -def test_get_variant(sentry_init, capture_events, uninstall_integration): - uninstall_integration(UnleashIntegration.identifier) - - with mock_unleash_client(): - client = UnleashClient() - sentry_init(integrations=[UnleashIntegration()]) # type: ignore - client.get_variant("no_payload_feature") - client.get_variant("string_feature") - client.get_variant("json_feature") - client.get_variant("csv_feature") - client.get_variant("number_feature") - client.get_variant("unknown_feature") - - events = capture_events() - sentry_sdk.capture_exception(Exception("something wrong!")) - - assert len(events) == 1 - assert events[0]["contexts"]["flags"] == { - "values": [ - {"flag": "no_payload_feature", "result": True}, - {"flag": "string_feature", "result": True}, - {"flag": "json_feature", "result": True}, - {"flag": "csv_feature", "result": True}, - {"flag": "number_feature", "result": True}, - {"flag": "unknown_feature", "result": False}, - ] - } - - def test_is_enabled_threaded(sentry_init, capture_events, uninstall_integration): uninstall_integration(UnleashIntegration.identifier) with mock_unleash_client(): - client = UnleashClient() - sentry_init(integrations=[UnleashIntegration()]) # type: ignore + client = UnleashClient() # type: ignore[arg-type] + sentry_init(integrations=[UnleashIntegration()]) events = capture_events() def task(flag_key): @@ -112,63 +83,14 @@ def task(flag_key): } -def test_get_variant_threaded(sentry_init, capture_events, uninstall_integration): - uninstall_integration(UnleashIntegration.identifier) - - with mock_unleash_client(): - client = UnleashClient() - sentry_init(integrations=[UnleashIntegration()]) # type: ignore - events = capture_events() - - def task(flag_key): - # Creates a new isolation scope for the thread. - # This means the evaluations in each task are captured separately. - with sentry_sdk.isolation_scope(): - client.get_variant(flag_key) - # use a tag to identify to identify events later on - sentry_sdk.set_tag("task_id", flag_key) - sentry_sdk.capture_exception(Exception("something wrong!")) - - # Capture an eval before we split isolation scopes. - client.get_variant("hello") - - with cf.ThreadPoolExecutor(max_workers=2) as pool: - pool.map(task, ["no_payload_feature", "other"]) - - # Capture error in original scope - sentry_sdk.set_tag("task_id", "0") - sentry_sdk.capture_exception(Exception("something wrong!")) - - assert len(events) == 3 - events.sort(key=lambda e: e["tags"]["task_id"]) - - assert events[0]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - ] - } - assert events[1]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - {"flag": "no_payload_feature", "result": True}, - ] - } - assert events[2]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - {"flag": "other", "result": False}, - ] - } - - @pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") def test_is_enabled_asyncio(sentry_init, capture_events, uninstall_integration): asyncio = pytest.importorskip("asyncio") uninstall_integration(UnleashIntegration.identifier) with mock_unleash_client(): - client = UnleashClient() - sentry_init(integrations=[UnleashIntegration()]) # type: ignore + client = UnleashClient() # type: ignore[arg-type] + sentry_init(integrations=[UnleashIntegration()]) events = capture_events() async def task(flag_key): @@ -212,66 +134,12 @@ async def runner(): } -@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") -def test_get_variant_asyncio(sentry_init, capture_events, uninstall_integration): - asyncio = pytest.importorskip("asyncio") - - uninstall_integration(UnleashIntegration.identifier) - - with mock_unleash_client(): - client = UnleashClient() - sentry_init(integrations=[UnleashIntegration()]) # type: ignore - events = capture_events() - - async def task(flag_key): - with sentry_sdk.isolation_scope(): - client.get_variant(flag_key) - # use a tag to identify to identify events later on - sentry_sdk.set_tag("task_id", flag_key) - sentry_sdk.capture_exception(Exception("something wrong!")) - - async def runner(): - return asyncio.gather(task("no_payload_feature"), task("other")) - - # Capture an eval before we split isolation scopes. - client.get_variant("hello") - - asyncio.run(runner()) - - # Capture error in original scope - sentry_sdk.set_tag("task_id", "0") - sentry_sdk.capture_exception(Exception("something wrong!")) - - assert len(events) == 3 - events.sort(key=lambda e: e["tags"]["task_id"]) - - assert events[0]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - ] - } - assert events[1]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - {"flag": "no_payload_feature", "result": True}, - ] - } - assert events[2]["contexts"]["flags"] == { - "values": [ - {"flag": "hello", "result": False}, - {"flag": "other", "result": False}, - ] - } - - def test_wraps_original(sentry_init, uninstall_integration): with mock_unleash_client(): - client = UnleashClient() + client = UnleashClient() # type: ignore[arg-type] mock_is_enabled = mock.Mock(return_value=random() < 0.5) - mock_get_variant = mock.Mock(return_value={"enabled": random() < 0.5}) client.is_enabled = mock_is_enabled - client.get_variant = mock_get_variant uninstall_integration(UnleashIntegration.identifier) sentry_init(integrations=[UnleashIntegration()]) # type: ignore @@ -283,20 +151,12 @@ def test_wraps_original(sentry_init, uninstall_integration): {"kwarg": 1}, ) - res = client.get_variant("test-flag", "arg", kwarg=1) - assert res == mock_get_variant.return_value - assert mock_get_variant.call_args == ( - ("test-flag", "arg"), - {"kwarg": 1}, - ) - def test_wrapper_attributes(sentry_init, uninstall_integration): with mock_unleash_client(): - client = UnleashClient() # <- Returns a MockUnleashClient + client = UnleashClient() # type: ignore[arg-type] original_is_enabled = client.is_enabled - original_get_variant = client.get_variant uninstall_integration(UnleashIntegration.identifier) sentry_init(integrations=[UnleashIntegration()]) # type: ignore @@ -304,5 +164,3 @@ def test_wrapper_attributes(sentry_init, uninstall_integration): # Mock clients methods have not lost their qualified names after decoration. assert client.is_enabled.__name__ == "is_enabled" assert client.is_enabled.__qualname__ == original_is_enabled.__qualname__ - assert client.get_variant.__name__ == "get_variant" - assert client.get_variant.__qualname__ == original_get_variant.__qualname__ diff --git a/tests/integrations/unleash/testutils.py b/tests/integrations/unleash/testutils.py index c424b34c3a..07b065e2f0 100644 --- a/tests/integrations/unleash/testutils.py +++ b/tests/integrations/unleash/testutils.py @@ -8,8 +8,8 @@ def mock_unleash_client(): Temporarily replaces UnleashClient's methods with mock implementations for testing. - This context manager swaps out UnleashClient's __init__, is_enabled, - and get_variant methods with mock versions from MockUnleashClient. + This context manager swaps out UnleashClient's __init__ and is_enabled, + methods with mock versions from MockUnleashClient. Original methods are restored when exiting the context. After mocking the client class the integration can be initialized. @@ -23,17 +23,14 @@ def mock_unleash_client(): """ old_init = UnleashClient.__init__ old_is_enabled = UnleashClient.is_enabled - old_get_variant = UnleashClient.get_variant UnleashClient.__init__ = MockUnleashClient.__init__ UnleashClient.is_enabled = MockUnleashClient.is_enabled - UnleashClient.get_variant = MockUnleashClient.get_variant yield UnleashClient.__init__ = old_init UnleashClient.is_enabled = old_is_enabled - UnleashClient.get_variant = old_get_variant class MockUnleashClient: @@ -44,34 +41,5 @@ def __init__(self, *a, **kw): "world": False, } - self.feature_to_variant = { - "string_feature": { - "name": "variant1", - "enabled": True, - "payload": {"type": "string", "value": "val1"}, - }, - "json_feature": { - "name": "variant1", - "enabled": True, - "payload": {"type": "json", "value": '{"key1": 0.53}'}, - }, - "number_feature": { - "name": "variant1", - "enabled": True, - "payload": {"type": "number", "value": "134.5"}, - }, - "csv_feature": { - "name": "variant1", - "enabled": True, - "payload": {"type": "csv", "value": "abc 123\ncsbq 94"}, - }, - "no_payload_feature": {"name": "variant1", "enabled": True}, - } - - self.disabled_variant = {"name": "disabled", "enabled": False} - def is_enabled(self, feature, *a, **kw): return self.features.get(feature, False) - - def get_variant(self, feature, *a, **kw): - return self.feature_to_variant.get(feature, self.disabled_variant) From ca68a7f3fb8e1cb6e1c58432211422b4c2bc4530 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 14 Jan 2025 08:33:39 +0000 Subject: [PATCH 094/591] release: 2.20.0 --- CHANGELOG.md | 31 +++++++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 34 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index af4eb04fef..57df5a9035 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,36 @@ # Changelog +## 2.20.0 + +### Various fixes & improvements + +- feat(flags): remove Unleash get_variant patching code (#3914) by @aliu39 +- Moved adding of `flags` context into Scope (#3917) by @antonpirker +- tests: Create a separate group for feature flag suites (#3911) by @sentrivana +- Treat potel-base as release branch in CI (#3912) by @sentrivana +- Centralize minimum version checking (#3910) by @sentrivana +- Small contribution docs update (#3909) by @antonpirker +- feat(flags): add Unleash feature flagging integration (#3888) by @aliu39 +- ref(flags): Beter naming for featureflags module and identifier (#3902) by @aliu39 +- Revert "ref(flags): register LD hook in setup instead of init, and don't chec…" (#3900) by @cmanallen +- Update test matrix for Sanic (#3904) by @antonpirker +- fix: preserve ARQ enqueue_job __kwdefaults__ after patching (#3903) by @danmr +- fix(flags): fix/refactor flaky launchdarkly tests (#3896) by @aliu39 +- Fix cache pollution from mutable reference (#3887) by @cmanallen +- ref(flags): register LD hook in setup instead of init, and don't check for initialization (#3890) by @aliu39 +- build(deps): bump actions/create-github-app-token from 1.11.0 to 1.11.1 (#3893) by @dependabot +- build(deps): bump codecov/codecov-action from 5.1.1 to 5.1.2 (#3892) by @dependabot +- Fix lru cache copying (#3883) by @ffelixg +- Rename scripts (#3885) by @sentrivana +- Support SparkIntegration activation after SparkContext created (#3411) by @seyoon-lim +- build(deps): bump codecov/codecov-action from 5.0.7 to 5.1.1 (#3867) by @dependabot +- Add github workflow to comment on issues when a fix was released (#3866) by @antonpirker +- feat(flags): Add integration for custom tracking of flag evaluations (#3860) by @aliu39 +- ✨ Add Typer integration (#3869) by @patrick91 +- Fix CI (#3878) by @sentrivana + +_Plus 3 more_ + ## 2.19.2 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 3ecdbe2e68..1d58274beb 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.19.2" +release = "2.20.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 0bb71cb98d..23f79ebd63 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -581,4 +581,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.19.2" +VERSION = "2.20.0" diff --git a/setup.py b/setup.py index 9e24d59d21..1bfbb6f7e4 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.19.2", + version="2.20.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 4e0505ea5c58943f31de35f03d834daa18e7f7ed Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 14 Jan 2025 10:22:18 +0100 Subject: [PATCH 095/591] Updated changelog --- CHANGELOG.md | 52 +++++++++++++++++++++++++++++----------------------- 1 file changed, 29 insertions(+), 23 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 57df5a9035..abbb5d5627 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,34 +2,32 @@ ## 2.20.0 -### Various fixes & improvements +- **New integration:** Add [Typer](https://typer.tiangolo.com/) integration (#3869) by @patrick91 + + For more information, see the documentation for the [TyperIntegration](https://docs.sentry.io/platforms/python/integrations/typer/). + +- **New integration:** Add [Unleash](https://www.getunleash.io/) feature flagging integration (#3888) by @aliu39 + + For more information, see the documentation for the [UnleashIntegration](https://docs.sentry.io/platforms/python/integrations/unleash/). -- feat(flags): remove Unleash get_variant patching code (#3914) by @aliu39 -- Moved adding of `flags` context into Scope (#3917) by @antonpirker -- tests: Create a separate group for feature flag suites (#3911) by @sentrivana -- Treat potel-base as release branch in CI (#3912) by @sentrivana +- Add custom tracking of feature flag evaluations (#3860) by @aliu39 +- Feature Flags: Register LD hook in setup instead of init, and don't check for initialization (#3890) by @aliu39 +- Feature Flags: Moved adding of `flags` context into Scope (#3917) by @antonpirker +- Create a separate group for feature flag test suites (#3911) by @sentrivana +- Fix flaky LaunchDarkly tests (#3896) by @aliu39 +- Fix LRU cache copying (#3883) by @ffelixg +- Fix cache pollution from mutable reference (#3887) by @cmanallen - Centralize minimum version checking (#3910) by @sentrivana -- Small contribution docs update (#3909) by @antonpirker -- feat(flags): add Unleash feature flagging integration (#3888) by @aliu39 -- ref(flags): Beter naming for featureflags module and identifier (#3902) by @aliu39 -- Revert "ref(flags): register LD hook in setup instead of init, and don't chec…" (#3900) by @cmanallen +- Support SparkIntegration activation after SparkContext created (#3411) by @seyoon-lim +- Preserve ARQ enqueue_job __kwdefaults__ after patching (#3903) by @danmr +- Add Github workflow to comment on issues when a fix was released (#3866) by @antonpirker - Update test matrix for Sanic (#3904) by @antonpirker -- fix: preserve ARQ enqueue_job __kwdefaults__ after patching (#3903) by @danmr -- fix(flags): fix/refactor flaky launchdarkly tests (#3896) by @aliu39 -- Fix cache pollution from mutable reference (#3887) by @cmanallen -- ref(flags): register LD hook in setup instead of init, and don't check for initialization (#3890) by @aliu39 -- build(deps): bump actions/create-github-app-token from 1.11.0 to 1.11.1 (#3893) by @dependabot -- build(deps): bump codecov/codecov-action from 5.1.1 to 5.1.2 (#3892) by @dependabot -- Fix lru cache copying (#3883) by @ffelixg - Rename scripts (#3885) by @sentrivana -- Support SparkIntegration activation after SparkContext created (#3411) by @seyoon-lim -- build(deps): bump codecov/codecov-action from 5.0.7 to 5.1.1 (#3867) by @dependabot -- Add github workflow to comment on issues when a fix was released (#3866) by @antonpirker -- feat(flags): Add integration for custom tracking of flag evaluations (#3860) by @aliu39 -- ✨ Add Typer integration (#3869) by @patrick91 - Fix CI (#3878) by @sentrivana - -_Plus 3 more_ +- Treat `potel-base` as release branch in CI (#3912) by @sentrivana +- build(deps): bump actions/create-github-app-token from 1.11.0 to 1.11.1 (#3893) by @dependabot +- build(deps): bump codecov/codecov-action from 5.0.7 to 5.1.1 (#3867) by @dependabot +- build(deps): bump codecov/codecov-action from 5.1.1 to 5.1.2 (#3892) by @dependabot ## 2.19.2 @@ -86,6 +84,14 @@ _Plus 3 more_ ### Various fixes & improvements +- **New integration:** Add [LaunchDarkly](https://launchdarkly.com/) integration (#3648) by @cmanallen + + For more information, see the documentation for the [LaunchDarklyIntegration](https://docs.sentry.io/platforms/python/integrations/launchdarkly/). + +- **New integration:** Add [OpenFeature](https://openfeature.dev/) feature flagging integration (#3648) by @cmanallen + + For more information, see the documentation for the [OpenFeatureIntegration](https://docs.sentry.io/platforms/python/integrations/opoenfeature/). + - Add LaunchDarkly and OpenFeature integration (#3648) by @cmanallen - Correct typo in a comment (#3726) by @szokeasaurusrex - End `http.client` span on timeout (#3723) by @Zylphrex From 98d0415cc354f76949add22136a9ae5af7db2089 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 14 Jan 2025 13:55:28 +0100 Subject: [PATCH 096/591] Typo (#3923) --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. The AWS Lambda tests additionally require a maintainer to add a special label, and they will fail until this label is added. --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index abbb5d5627..80ff6c2796 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -90,7 +90,7 @@ - **New integration:** Add [OpenFeature](https://openfeature.dev/) feature flagging integration (#3648) by @cmanallen - For more information, see the documentation for the [OpenFeatureIntegration](https://docs.sentry.io/platforms/python/integrations/opoenfeature/). + For more information, see the documentation for the [OpenFeatureIntegration](https://docs.sentry.io/platforms/python/integrations/openfeature/). - Add LaunchDarkly and OpenFeature integration (#3648) by @cmanallen - Correct typo in a comment (#3726) by @szokeasaurusrex From 9ff100a981e11c8f9bebd1ff51aee59864d693d4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 14 Jan 2025 14:26:35 +0100 Subject: [PATCH 097/591] Handle `None` lineno in `get_source_context` (#3925) Be more defensive in `get_source_context`. The current check makes no sense as we first try to decrement `tb_lineno` and then check the result against `None`: ```python lineno = tb_lineno - 1 if lineno is not None and abs_path: ``` So it looks like this was an oversight/got broken at some point. Closes https://github.com/getsentry/sentry-python/issues/3924 --- sentry_sdk/utils.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index ae6e7538ac..7a8917fecc 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -571,7 +571,7 @@ def get_lines_from_file( def get_source_context( frame, # type: FrameType - tb_lineno, # type: int + tb_lineno, # type: Optional[int] max_value_length=None, # type: Optional[int] ): # type: (...) -> Tuple[List[Annotated[str]], Optional[Annotated[str]], List[Annotated[str]]] @@ -587,11 +587,13 @@ def get_source_context( loader = frame.f_globals["__loader__"] except Exception: loader = None - lineno = tb_lineno - 1 - if lineno is not None and abs_path: + + if tb_lineno is not None and abs_path: + lineno = tb_lineno - 1 return get_lines_from_file( abs_path, lineno, max_value_length, loader=loader, module=module ) + return [], None, [] From 8a70b76f69789585efbd39fcef087005b765a346 Mon Sep 17 00:00:00 2001 From: Marcelo Galigniana Date: Tue, 14 Jan 2025 11:34:38 -0300 Subject: [PATCH 098/591] feat(tracing): Add `propagate_traces` deprecation warning (#3899) Fixes GH-3106 Co-authored-by: Anton Pirker --- sentry_sdk/integrations/celery/__init__.py | 7 +++++++ sentry_sdk/scope.py | 5 +++++ tests/integrations/celery/test_celery.py | 11 +++++++---- tests/tracing/test_integration_tests.py | 14 ++++++++++++++ 4 files changed, 33 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index dc48aac0e6..80decb6064 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -1,4 +1,6 @@ import sys +import warnings + from collections.abc import Mapping from functools import wraps @@ -68,6 +70,11 @@ def __init__( exclude_beat_tasks=None, ): # type: (bool, bool, Optional[List[str]]) -> None + warnings.warn( + "The `propagate_traces` parameter is deprecated. Please use `trace_propagation_targets` instead.", + DeprecationWarning, + stacklevel=2, + ) self.propagate_traces = propagate_traces self.monitor_beat_tasks = monitor_beat_tasks self.exclude_beat_tasks = exclude_beat_tasks diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index ab0f1f4156..c22cdfb030 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -621,6 +621,11 @@ def iter_trace_propagation_headers(self, *args, **kwargs): """ client = self.get_client() if not client.options.get("propagate_traces"): + warnings.warn( + "The `propagate_traces` parameter is deprecated. Please use `trace_propagation_targets` instead.", + DeprecationWarning, + stacklevel=2, + ) return span = kwargs.pop("span", None) diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index e51341599f..f8d118e7e9 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -268,7 +268,9 @@ def dummy_task(): def test_simple_no_propagation(capture_events, init_celery): - celery = init_celery(propagate_traces=False) + with pytest.warns(DeprecationWarning): + celery = init_celery(propagate_traces=False) + events = capture_events() @celery.task(name="dummy_task") @@ -532,9 +534,10 @@ def test_sentry_propagate_traces_override(init_celery): Test if the `sentry-propagate-traces` header given to `apply_async` overrides the `propagate_traces` parameter in the integration constructor. """ - celery = init_celery( - propagate_traces=True, traces_sample_rate=1.0, release="abcdef" - ) + with pytest.warns(DeprecationWarning): + celery = init_celery( + propagate_traces=True, traces_sample_rate=1.0, release="abcdef" + ) @celery.task(name="dummy_task", bind=True) def dummy_task(self, message): diff --git a/tests/tracing/test_integration_tests.py b/tests/tracing/test_integration_tests.py index e27dbea901..da3efef9eb 100644 --- a/tests/tracing/test_integration_tests.py +++ b/tests/tracing/test_integration_tests.py @@ -138,6 +138,20 @@ def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_r assert message_payload["message"] == "hello" +@pytest.mark.parametrize("sample_rate", [0.0, 1.0]) +def test_propagate_traces_deprecation_warning(sentry_init, sample_rate): + sentry_init(traces_sample_rate=sample_rate, propagate_traces=False) + + with start_transaction(name="hi"): + with start_span() as old_span: + with pytest.warns(DeprecationWarning): + dict( + sentry_sdk.get_current_scope().iter_trace_propagation_headers( + old_span + ) + ) + + @pytest.mark.parametrize("sample_rate", [0.5, 1.0]) def test_dynamic_sampling_head_sdk_creates_dsc( sentry_init, capture_envelopes, sample_rate, monkeypatch From 3f57299d1addff54a2d218c069e466a371edc8c4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 16 Jan 2025 14:19:13 +0100 Subject: [PATCH 099/591] Test Celery's latest RC (#3938) --- tox.ini | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tox.ini b/tox.ini index 95c09a573e..3cab20a1f1 100644 --- a/tox.ini +++ b/tox.ini @@ -391,7 +391,7 @@ deps = celery-v5.3: Celery~=5.3.0 celery-v5.4: Celery~=5.4.0 # TODO: update when stable is out - celery-v5.5: Celery==5.5.0rc3 + celery-v5.5: Celery==5.5.0rc4 celery-latest: Celery celery: newrelic From a85f0fb8ba6f235d0ca21760dbe3ab64cb46ea7d Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 20 Jan 2025 12:53:53 +0100 Subject: [PATCH 100/591] fix(utils): Check that `__module__` is `str` (#3942) Fixes #3939 --- sentry_sdk/utils.py | 2 +- tests/test_utils.py | 20 ++++++++++++++++++++ 2 files changed, 21 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 7a8917fecc..0fead48377 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1501,7 +1501,7 @@ def qualname_from_function(func): # Python 3: methods, functions, classes if func_qualname is not None: - if hasattr(func, "__module__"): + if hasattr(func, "__module__") and isinstance(func.__module__, str): func_qualname = func.__module__ + "." + func_qualname func_qualname = prefix + func_qualname + suffix diff --git a/tests/test_utils.py b/tests/test_utils.py index 6e01bb4f3a..894638bf4d 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -951,3 +951,23 @@ def test_format_timestamp_naive(): # Ensure that some timestamp is returned, without error. We currently treat these as local time, but this is an # implementation detail which we should not assert here. assert re.fullmatch(timestamp_regex, format_timestamp(datetime_object)) + + +def test_qualname_from_function_inner_function(): + def test_function(): ... + + assert ( + sentry_sdk.utils.qualname_from_function(test_function) + == "tests.test_utils.test_qualname_from_function_inner_function..test_function" + ) + + +def test_qualname_from_function_none_name(): + def test_function(): ... + + test_function.__module__ = None + + assert ( + sentry_sdk.utils.qualname_from_function(test_function) + == "test_qualname_from_function_none_name..test_function" + ) From 4ae94a5c1265218bc48ae1d38dec76f7e24b3df9 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 20 Jan 2025 14:08:09 +0100 Subject: [PATCH 101/591] Use httpx_mock in test_httpx (#3967) Co-authored-by: Neel Shah --- tests/integrations/httpx/test_httpx.py | 35 +++++++++++++++++--------- 1 file changed, 23 insertions(+), 12 deletions(-) diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index 17bf7017a5..107f873a3c 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -3,7 +3,6 @@ import httpx import pytest -import responses import sentry_sdk from sentry_sdk import capture_message, start_transaction @@ -16,7 +15,9 @@ "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_crumb_capture_and_hint(sentry_init, capture_events, httpx_client): +def test_crumb_capture_and_hint(sentry_init, capture_events, httpx_client, httpx_mock): + httpx_mock.add_response() + def before_breadcrumb(crumb, hint): crumb["data"]["extra"] = "foo" return crumb @@ -24,7 +25,6 @@ def before_breadcrumb(crumb, hint): sentry_init(integrations=[HttpxIntegration()], before_breadcrumb=before_breadcrumb) url = "http://example.com/" - responses.add(responses.GET, url, status=200) with start_transaction(): events = capture_events() @@ -61,11 +61,15 @@ def before_breadcrumb(crumb, hint): "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_outgoing_trace_headers(sentry_init, httpx_client): - sentry_init(traces_sample_rate=1.0, integrations=[HttpxIntegration()]) +def test_outgoing_trace_headers(sentry_init, httpx_client, httpx_mock): + httpx_mock.add_response() + + sentry_init( + traces_sample_rate=1.0, + integrations=[HttpxIntegration()], + ) url = "http://example.com/" - responses.add(responses.GET, url, status=200) with start_transaction( name="/interactions/other-dogs/new-dog", @@ -93,7 +97,13 @@ def test_outgoing_trace_headers(sentry_init, httpx_client): "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_outgoing_trace_headers_append_to_baggage(sentry_init, httpx_client): +def test_outgoing_trace_headers_append_to_baggage( + sentry_init, + httpx_client, + httpx_mock, +): + httpx_mock.add_response() + sentry_init( traces_sample_rate=1.0, integrations=[HttpxIntegration()], @@ -101,7 +111,6 @@ def test_outgoing_trace_headers_append_to_baggage(sentry_init, httpx_client): ) url = "http://example.com/" - responses.add(responses.GET, url, status=200) with start_transaction( name="/interactions/other-dogs/new-dog", @@ -290,12 +299,13 @@ def test_do_not_propagate_outside_transaction(sentry_init, httpx_mock): @pytest.mark.tests_internal_exceptions -def test_omit_url_data_if_parsing_fails(sentry_init, capture_events): +def test_omit_url_data_if_parsing_fails(sentry_init, capture_events, httpx_mock): + httpx_mock.add_response() + sentry_init(integrations=[HttpxIntegration()]) httpx_client = httpx.Client() url = "http://example.com/" - responses.add(responses.GET, url, status=200) events = capture_events() with mock.patch( @@ -326,7 +336,9 @@ def test_omit_url_data_if_parsing_fails(sentry_init, capture_events): "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_span_origin(sentry_init, capture_events, httpx_client): +def test_span_origin(sentry_init, capture_events, httpx_client, httpx_mock): + httpx_mock.add_response() + sentry_init( integrations=[HttpxIntegration()], traces_sample_rate=1.0, @@ -335,7 +347,6 @@ def test_span_origin(sentry_init, capture_events, httpx_client): events = capture_events() url = "http://example.com/" - responses.add(responses.GET, url, status=200) with start_transaction(name="test_transaction"): if asyncio.iscoroutinefunction(httpx_client.get): From d2ccac0addbe3591a887d19aa21ab69245296241 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 20 Jan 2025 15:50:57 +0100 Subject: [PATCH 102/591] Add support for Python 3.12 and 3.13 to AWS Lambda integration. (#3965) Its time to add support for newer versions of Python to our AWS Lambda integration. Fixes #3946 --- .craft.yml | 2 ++ tests/integrations/aws_lambda/test_aws.py | 3 +-- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/.craft.yml b/.craft.yml index 70875d5404..665f06834a 100644 --- a/.craft.yml +++ b/.craft.yml @@ -25,6 +25,8 @@ targets: - python3.9 - python3.10 - python3.11 + - python3.12 + - python3.13 license: MIT - name: sentry-pypi internalPypiRepo: getsentry/pypi diff --git a/tests/integrations/aws_lambda/test_aws.py b/tests/integrations/aws_lambda/test_aws.py index e229812336..f60bedc846 100644 --- a/tests/integrations/aws_lambda/test_aws.py +++ b/tests/integrations/aws_lambda/test_aws.py @@ -38,10 +38,9 @@ RUNTIMES_TO_TEST = [ "python3.8", - "python3.9", "python3.10", - "python3.11", "python3.12", + "python3.13", ] LAMBDA_PRELUDE = """ From 48d63683e675800edc079435bd4a63bed66e1e60 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 21 Jan 2025 09:55:09 +0100 Subject: [PATCH 103/591] Split gevent tests off (#3964) Same as https://github.com/getsentry/sentry-python/pull/3962, but for master --- .../workflows/test-integrations-gevent.yml | 91 +++++++++++++++++++ .../workflows/test-integrations-network.yml | 8 -- .../split_tox_gh_actions.py | 4 +- 3 files changed, 94 insertions(+), 9 deletions(-) create mode 100644 .github/workflows/test-integrations-gevent.yml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml new file mode 100644 index 0000000000..088f952ea3 --- /dev/null +++ b/.github/workflows/test-integrations-gevent.yml @@ -0,0 +1,91 @@ +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja +name: Test Gevent +on: + push: + branches: + - master + - release/** + - potel-base + pull_request: +# Cancel in progress workflows on pull_requests. +# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value +concurrency: + group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} + cancel-in-progress: true +permissions: + contents: read +env: + BUILD_CACHE_KEY: ${{ github.sha }} + CACHED_BUILD_PATHS: | + ${{ github.workspace }}/dist-serverless +jobs: + test-gevent-pinned: + name: Gevent (pinned) + timeout-minutes: 30 + runs-on: ${{ matrix.os }} + strategy: + fail-fast: false + matrix: + python-version: ["3.6","3.8","3.10","3.11","3.12"] + # python3.6 reached EOL and is no longer being supported on + # new versions of hosted runners on Github Actions + # ubuntu-20.04 is the last version that supported python3.6 + # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 + os: [ubuntu-20.04] + steps: + - uses: actions/checkout@v4.2.2 + - uses: actions/setup-python@v5 + with: + python-version: ${{ matrix.python-version }} + allow-prereleases: true + - name: Setup Test Env + run: | + pip install "coverage[toml]" tox + - name: Erase coverage + run: | + coverage erase + - name: Test gevent pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-gevent" + - name: Generate coverage XML (Python 3.6) + if: ${{ !cancelled() && matrix.python-version == '3.6' }} + run: | + export COVERAGE_RCFILE=.coveragerc36 + coverage combine .coverage-sentry-* + coverage xml --ignore-errors + - name: Generate coverage XML + if: ${{ !cancelled() && matrix.python-version != '3.6' }} + run: | + coverage combine .coverage-sentry-* + coverage xml + - name: Upload coverage to Codecov + if: ${{ !cancelled() }} + uses: codecov/codecov-action@v5.1.2 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: coverage.xml + # make sure no plugins alter our coverage reports + plugin: noop + verbose: true + - name: Upload test results to Codecov + if: ${{ !cancelled() }} + uses: codecov/test-results-action@v1 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: .junitxml + verbose: true + check_required_tests: + name: All pinned Gevent tests passed + needs: test-gevent-pinned + # Always run this, even if a dependent job failed + if: always() + runs-on: ubuntu-20.04 + steps: + - name: Check for failures + if: contains(needs.test-gevent-pinned.result, 'failure') || contains(needs.test-gevent-pinned.result, 'skipped') + run: | + echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index ab1c5b0658..b5593a58fd 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -47,10 +47,6 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test gevent latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-gevent-latest" - name: Test grpc latest run: | set -x # print commands that are executed @@ -115,10 +111,6 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test gevent pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-gevent" - name: Test grpc pinned run: | set -x # print commands that are executed diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 1537ad8389..43307c3093 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -89,6 +89,9 @@ "openfeature", "unleash", ], + "Gevent": [ + "gevent", + ], "GraphQL": [ "ariadne", "gql", @@ -96,7 +99,6 @@ "strawberry", ], "Network": [ - "gevent", "grpc", "httpx", "requests", From 7c757c221cb42cc5213b90a85d8bceff4ce67dc9 Mon Sep 17 00:00:00 2001 From: Philipp Hofmann Date: Wed, 22 Jan 2025 16:43:11 +0100 Subject: [PATCH 104/591] chore: Increase date range for MIT licence (#3990) It's 2025 now. --- LICENSE | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/LICENSE b/LICENSE index c4c8162f13..4477bfef36 100644 --- a/LICENSE +++ b/LICENSE @@ -1,6 +1,6 @@ MIT License -Copyright (c) 2018-2024 Functional Software, Inc. dba Sentry +Copyright (c) 2018-2025 Functional Software, Inc. dba Sentry Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal From 968b3620623bd1a6c90eb71682876e4f93e5c125 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 23 Jan 2025 17:32:14 +0100 Subject: [PATCH 105/591] Deprecate `enable_tracing` option (#3935) The option `enable_tracing` is deprecated in favor of using `traces_sample_rate`. Fixes #3918 --- sentry_sdk/client.py | 8 ++++++++ tests/test_client.py | 6 ++++++ 2 files changed, 14 insertions(+) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index cf345c41f9..cace8cc224 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -6,6 +6,7 @@ from datetime import datetime, timezone from importlib import import_module from typing import cast, overload +import warnings from sentry_sdk._compat import PY37, check_uwsgi_thread_support from sentry_sdk.utils import ( @@ -140,6 +141,13 @@ def _get_options(*args, **kwargs): ) rv["socket_options"] = None + if rv["enable_tracing"] is not None: + warnings.warn( + "The `enable_tracing` parameter is deprecated. Please use `traces_sample_rate` instead.", + DeprecationWarning, + stacklevel=2, + ) + return rv diff --git a/tests/test_client.py b/tests/test_client.py index 450e19603f..67f53d989a 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -1490,3 +1490,9 @@ def run(self, sentry_init, capture_record_lost_event_calls): ) def test_dropped_transaction(sentry_init, capture_record_lost_event_calls, test_config): test_config.run(sentry_init, capture_record_lost_event_calls) + + +@pytest.mark.parametrize("enable_tracing", [True, False]) +def test_enable_tracing_deprecated(sentry_init, enable_tracing): + with pytest.warns(DeprecationWarning): + sentry_init(enable_tracing=enable_tracing) From 7473afb77d7f0ba534bf5fdcd22622b06a5f7e62 Mon Sep 17 00:00:00 2001 From: Philipp Hofmann Date: Tue, 28 Jan 2025 11:23:08 +0100 Subject: [PATCH 106/591] Remove date range for LICENSE (#3991) While updating the date ranges for multiple PRs, Michi pointed out that we don't need date ranges for our licenses. I'm sorry about the fuzz. In our internal [Open Source Legal Policy](https://www.notion.so/sentry/Open-Source-Legal-Policy-ac4885d265cb4d7898a01c060b061e42), we decided that licenses don't require a data range. This also has the advantage of not updating the date range yearly. --- LICENSE | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/LICENSE b/LICENSE index 4477bfef36..016323bd8d 100644 --- a/LICENSE +++ b/LICENSE @@ -1,6 +1,6 @@ MIT License -Copyright (c) 2018-2025 Functional Software, Inc. dba Sentry +Copyright (c) 2018 Functional Software, Inc. dba Sentry Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal From 5a2750215f4c48fa98dfec01ae5bd2261ec0c2f4 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 28 Jan 2025 10:29:12 +0000 Subject: [PATCH 107/591] build(deps): bump codecov/codecov-action from 5.1.2 to 5.3.1 (#3995) --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-aws.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 4 ++-- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 4 ++-- .github/workflows/test-integrations-misc.yml | 4 ++-- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 4 ++-- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 14 files changed, 24 insertions(+), 24 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 6e06e6067c..b9ade22f08 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -80,7 +80,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -152,7 +152,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml index eae488776a..21171f7843 100644 --- a/.github/workflows/test-integrations-aws.yml +++ b/.github/workflows/test-integrations-aws.yml @@ -99,7 +99,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index af089caede..b929b8d899 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -144,7 +144,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index d9e08bbeb8..11506d0f0f 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -64,7 +64,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index f612b8fb14..0f5c37306a 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 0460868473..096da8d672 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -72,7 +72,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -136,7 +136,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 088f952ea3..2729c3e701 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -64,7 +64,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index d239b2ed6c..d7cf8d80c1 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -144,7 +144,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 9461ea506c..82577c7be6 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -84,7 +84,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -160,7 +160,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index b5593a58fd..56f4bcfd57 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -72,7 +72,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -136,7 +136,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 8ecc7ab598..31e6f3c97a 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 2dc5f361de..9b3a2f06ec 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 2b3204ae80..3c010fc0bd 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -100,7 +100,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -192,7 +192,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 186d70c9fd..66e346511d 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -92,7 +92,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.1.2 + uses: codecov/codecov-action@v5.3.1 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From 8c25c73ef8693a0d75d05e8278ee70dae7846fe7 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Fri, 31 Jan 2025 00:43:38 -0800 Subject: [PATCH 108/591] fix(ci): Various errors on master (#4009) - `black==25.1.0` changed some default styles - `pytest-aiohttp==1.1.0` removed the `loop` fixture - `huggingface-hub==0.28.0` deprecated `InferenceClient.post` to `InferenceClient._inner_post` - `pymongo==4.11.0` required `maxWireVersion` to be `7` --- sentry_sdk/_queue.py | 2 ++ tests/integrations/aiohttp/test_aiohttp.py | 11 +++++++- .../huggingface_hub/test_huggingface_hub.py | 26 +++++++++++++++---- tests/integrations/pymongo/test_pymongo.py | 2 +- 4 files changed, 34 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/_queue.py b/sentry_sdk/_queue.py index c0410d1f92..a21c86ec0a 100644 --- a/sentry_sdk/_queue.py +++ b/sentry_sdk/_queue.py @@ -86,11 +86,13 @@ class EmptyError(Exception): "Exception raised by Queue.get(block=0)/get_nowait()." + pass class FullError(Exception): "Exception raised by Queue.put(block=0)/put_nowait()." + pass diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index cd65e7cdd5..b689e3af17 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -1,5 +1,6 @@ import asyncio import json +import sys from contextlib import suppress from unittest import mock @@ -473,9 +474,17 @@ async def hello(request): assert error_event["contexts"]["trace"]["trace_id"] == trace_id +if sys.version_info < (3, 12): + # `loop` was deprecated in `pytest-aiohttp` + # in favor of `event_loop` from `pytest-asyncio` + @pytest.fixture + def event_loop(loop): + yield loop + + @pytest.mark.asyncio async def test_crumb_capture( - sentry_init, aiohttp_raw_server, aiohttp_client, loop, capture_events + sentry_init, aiohttp_raw_server, aiohttp_client, event_loop, capture_events ): def before_breadcrumb(crumb, hint): crumb["data"]["extra"] = "foo" diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index f43159d80e..e017ce2449 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -12,6 +12,13 @@ from unittest import mock # python 3.3 and above +def mock_client_post(client, post_mock): + # huggingface-hub==0.28.0 deprecates the `post` method + # so patch `_inner_post` instead + client.post = post_mock + client._inner_post = post_mock + + @pytest.mark.parametrize( "send_default_pii, include_prompts, details_arg", itertools.product([True, False], repeat=3), @@ -28,7 +35,7 @@ def test_nonstreaming_chat_completion( client = InferenceClient("some-model") if details_arg: - client.post = mock.Mock( + post_mock = mock.Mock( return_value=b"""[{ "generated_text": "the model response", "details": { @@ -40,9 +47,11 @@ def test_nonstreaming_chat_completion( }]""" ) else: - client.post = mock.Mock( + post_mock = mock.Mock( return_value=b'[{"generated_text": "the model response"}]' ) + mock_client_post(client, post_mock) + with start_transaction(name="huggingface_hub tx"): response = client.text_generation( prompt="hello", @@ -84,7 +93,8 @@ def test_streaming_chat_completion( events = capture_events() client = InferenceClient("some-model") - client.post = mock.Mock( + + post_mock = mock.Mock( return_value=[ b"""data:{ "token":{"id":1, "special": false, "text": "the model "} @@ -95,6 +105,8 @@ def test_streaming_chat_completion( }""", ] ) + mock_client_post(client, post_mock) + with start_transaction(name="huggingface_hub tx"): response = list( client.text_generation( @@ -131,7 +143,9 @@ def test_bad_chat_completion(sentry_init, capture_events): events = capture_events() client = InferenceClient("some-model") - client.post = mock.Mock(side_effect=OverloadedError("The server is overloaded")) + post_mock = mock.Mock(side_effect=OverloadedError("The server is overloaded")) + mock_client_post(client, post_mock) + with pytest.raises(OverloadedError): client.text_generation(prompt="hello") @@ -147,13 +161,15 @@ def test_span_origin(sentry_init, capture_events): events = capture_events() client = InferenceClient("some-model") - client.post = mock.Mock( + post_mock = mock.Mock( return_value=[ b"""data:{ "token":{"id":1, "special": false, "text": "the model "} }""", ] ) + mock_client_post(client, post_mock) + with start_transaction(name="huggingface_hub tx"): list( client.text_generation( diff --git a/tests/integrations/pymongo/test_pymongo.py b/tests/integrations/pymongo/test_pymongo.py index 80fe40fdcf..10f1c9fba9 100644 --- a/tests/integrations/pymongo/test_pymongo.py +++ b/tests/integrations/pymongo/test_pymongo.py @@ -10,7 +10,7 @@ @pytest.fixture(scope="session") def mongo_server(): server = MockupDB(verbose=True) - server.autoresponds("ismaster", maxWireVersion=6) + server.autoresponds("ismaster", maxWireVersion=7) server.run() server.autoresponds( {"find": "test_collection"}, cursor={"id": 123, "firstBatch": []} From 91bf3222740cfdf0d035fefc4c7073fb87e29937 Mon Sep 17 00:00:00 2001 From: Orhan Hirsch Date: Fri, 31 Jan 2025 13:48:59 +0100 Subject: [PATCH 109/591] Handle MultiPartParserError to avoid internal sentry crash (#4001) Handles an internal error in sentry_sdk if there is an issue with parsing request.POST. It would be better to handle this exception without request data instead of crashing and not reporting anything. --- Co-authored-by: Ivana Kellyer --- sentry_sdk/integrations/_wsgi_common.py | 11 +++++++++-- 1 file changed, 9 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/_wsgi_common.py b/sentry_sdk/integrations/_wsgi_common.py index 7266a91f56..48bc432887 100644 --- a/sentry_sdk/integrations/_wsgi_common.py +++ b/sentry_sdk/integrations/_wsgi_common.py @@ -149,8 +149,15 @@ def form(self): def parsed_body(self): # type: () -> Optional[Dict[str, Any]] - form = self.form() - files = self.files() + try: + form = self.form() + except Exception: + form = None + try: + files = self.files() + except Exception: + files = None + if form or files: data = {} if form: From 2724d65aa6739e391bfc19e689b0c7f0f403b4aa Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Mon, 3 Feb 2025 10:40:12 -0800 Subject: [PATCH 110/591] chore(profiling): Change continuous profile buffer size (#3987) This ~lowers the sampling frequency of continuous profiles to 21Hz and~ increases the buffer size to 1 minute to match the desired settings for continuous profiling. --- sentry_sdk/profiler/continuous_profiler.py | 2 +- sentry_sdk/transport.py | 2 +- tests/profiler/test_continuous_profiler.py | 6 +++--- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 5d64896b93..5a76a0696c 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -407,7 +407,7 @@ def teardown(self): self.buffer = None -PROFILE_BUFFER_SECONDS = 10 +PROFILE_BUFFER_SECONDS = 60 class ProfileBuffer: diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index 8798115898..3329b201b1 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -720,7 +720,7 @@ def _request( try: import httpcore - import h2 # type: ignore # noqa: F401 + import h2 # noqa: F401 except ImportError: # Sorry, no Http2Transport for you class Http2Transport(HttpTransport): diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 1b96f27036..32d0e8d0b0 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -200,7 +200,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.05) + time.sleep(0.1) assert_single_transaction_with_profile_chunks(envelopes, thread) @@ -211,7 +211,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.05) + time.sleep(0.1) assert_single_transaction_without_profile_chunks(envelopes) @@ -221,7 +221,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.05) + time.sleep(0.1) assert_single_transaction_with_profile_chunks(envelopes, thread) From bba389e5e52be1b0699f118c8aa60a08bcf00075 Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Tue, 4 Feb 2025 13:49:56 +0100 Subject: [PATCH 111/591] feat(spans): track and report spans that were dropped (#4005) `_SpanRecorder` now keeps track of `dropped_spans`, i.e. when above `max_spans`. When spans were dropped, the `"spans"` property will be wrapped in an `AnnotatedValue`, reporting the mutation. --- sentry_sdk/_types.py | 84 ++++++++++++++++++++++++++++++++++++-- sentry_sdk/client.py | 26 ++++++++---- sentry_sdk/scrubber.py | 5 +-- sentry_sdk/tracing.py | 10 ++++- sentry_sdk/transport.py | 8 ++-- sentry_sdk/utils.py | 81 +----------------------------------- tests/tracing/test_misc.py | 28 +++++++++++++ 7 files changed, 143 insertions(+), 99 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 4e3c195cc6..883b4cbc81 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -1,10 +1,88 @@ -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, TypeVar, Union # Re-exported for compat, since code out there in the wild might use this variable. MYPY = TYPE_CHECKING +SENSITIVE_DATA_SUBSTITUTE = "[Filtered]" + + +class AnnotatedValue: + """ + Meta information for a data field in the event payload. + This is to tell Relay that we have tampered with the fields value. + See: + https://github.com/getsentry/relay/blob/be12cd49a0f06ea932ed9b9f93a655de5d6ad6d1/relay-general/src/types/meta.rs#L407-L423 + """ + + __slots__ = ("value", "metadata") + + def __init__(self, value, metadata): + # type: (Optional[Any], Dict[str, Any]) -> None + self.value = value + self.metadata = metadata + + def __eq__(self, other): + # type: (Any) -> bool + if not isinstance(other, AnnotatedValue): + return False + + return self.value == other.value and self.metadata == other.metadata + + @classmethod + def removed_because_raw_data(cls): + # type: () -> AnnotatedValue + """The value was removed because it could not be parsed. This is done for request body values that are not json nor a form.""" + return AnnotatedValue( + value="", + metadata={ + "rem": [ # Remark + [ + "!raw", # Unparsable raw data + "x", # The fields original value was removed + ] + ] + }, + ) + + @classmethod + def removed_because_over_size_limit(cls): + # type: () -> AnnotatedValue + """The actual value was removed because the size of the field exceeded the configured maximum size (specified with the max_request_body_size sdk option)""" + return AnnotatedValue( + value="", + metadata={ + "rem": [ # Remark + [ + "!config", # Because of configured maximum size + "x", # The fields original value was removed + ] + ] + }, + ) + + @classmethod + def substituted_because_contains_sensitive_data(cls): + # type: () -> AnnotatedValue + """The actual value was removed because it contained sensitive information.""" + return AnnotatedValue( + value=SENSITIVE_DATA_SUBSTITUTE, + metadata={ + "rem": [ # Remark + [ + "!config", # Because of SDK configuration (in this case the config is the hard coded removal of certain django cookies) + "s", # The fields original value was substituted + ] + ] + }, + ) + + +T = TypeVar("T") +Annotated = Union[AnnotatedValue, T] + + if TYPE_CHECKING: from collections.abc import Container, MutableMapping, Sequence @@ -19,7 +97,6 @@ from typing import Optional from typing import Tuple from typing import Type - from typing import Union from typing_extensions import Literal, TypedDict class SDKInfo(TypedDict): @@ -101,7 +178,7 @@ class SDKInfo(TypedDict): "request": dict[str, object], "sdk": Mapping[str, object], "server_name": str, - "spans": list[dict[str, object]], + "spans": Annotated[list[dict[str, object]]], "stacktrace": dict[ str, object ], # We access this key in the code, but I am unsure whether we ever set it @@ -118,6 +195,7 @@ class SDKInfo(TypedDict): "transaction_info": Mapping[str, Any], # TODO: We can expand on this type "type": Literal["check_in", "transaction"], "user": dict[str, object], + "_dropped_spans": int, "_metrics_summary": dict[str, object], }, total=False, diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index cace8cc224..4f5c1566b3 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -5,11 +5,12 @@ from collections.abc import Mapping from datetime import datetime, timezone from importlib import import_module -from typing import cast, overload +from typing import TYPE_CHECKING, List, Dict, cast, overload import warnings from sentry_sdk._compat import PY37, check_uwsgi_thread_support from sentry_sdk.utils import ( + AnnotatedValue, ContextVar, capture_internal_exceptions, current_stacktrace, @@ -45,12 +46,9 @@ from sentry_sdk.monitor import Monitor from sentry_sdk.spotlight import setup_spotlight -from typing import TYPE_CHECKING - if TYPE_CHECKING: from typing import Any from typing import Callable - from typing import Dict from typing import Optional from typing import Sequence from typing import Type @@ -483,12 +481,14 @@ def _prepare_event( ): # type: (...) -> Optional[Event] + previous_total_spans = None # type: Optional[int] + if event.get("timestamp") is None: event["timestamp"] = datetime.now(timezone.utc) if scope is not None: is_transaction = event.get("type") == "transaction" - spans_before = len(event.get("spans", [])) + spans_before = len(cast(List[Dict[str, object]], event.get("spans", []))) event_ = scope.apply_to_event(event, hint, self.options) # one of the event/error processors returned None @@ -507,13 +507,18 @@ def _prepare_event( return None event = event_ - - spans_delta = spans_before - len(event.get("spans", [])) + spans_delta = spans_before - len( + cast(List[Dict[str, object]], event.get("spans", [])) + ) if is_transaction and spans_delta > 0 and self.transport is not None: self.transport.record_lost_event( "event_processor", data_category="span", quantity=spans_delta ) + dropped_spans = event.pop("_dropped_spans", 0) + spans_delta # type: int + if dropped_spans > 0: + previous_total_spans = spans_before + dropped_spans + if ( self.options["attach_stacktrace"] and "exception" not in event @@ -561,6 +566,11 @@ def _prepare_event( if event_scrubber: event_scrubber.scrub_event(event) + if previous_total_spans is not None: + event["spans"] = AnnotatedValue( + event.get("spans", []), {"len": previous_total_spans} + ) + # Postprocess the event here so that annotated types do # generally not surface in before_send if event is not None: @@ -598,7 +608,7 @@ def _prepare_event( and event.get("type") == "transaction" ): new_event = None - spans_before = len(event.get("spans", [])) + spans_before = len(cast(List[Dict[str, object]], event.get("spans", []))) with capture_internal_exceptions(): new_event = before_send_transaction(event, hint or {}) if new_event is None: diff --git a/sentry_sdk/scrubber.py b/sentry_sdk/scrubber.py index f4755ea93b..1df5573798 100644 --- a/sentry_sdk/scrubber.py +++ b/sentry_sdk/scrubber.py @@ -4,11 +4,10 @@ iter_event_frames, ) -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, cast, List, Dict if TYPE_CHECKING: from sentry_sdk._types import Event - from typing import List from typing import Optional @@ -161,7 +160,7 @@ def scrub_spans(self, event): # type: (Event) -> None with capture_internal_exceptions(): if "spans" in event: - for span in event["spans"]: + for span in cast(List[Dict[str, object]], event["spans"]): if "data" in span: self.scrub_dict(span["data"]) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 3868b2e6c8..86456b8964 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -193,7 +193,7 @@ def get_span_status_from_http_code(http_status_code): class _SpanRecorder: """Limits the number of spans recorded in a transaction.""" - __slots__ = ("maxlen", "spans") + __slots__ = ("maxlen", "spans", "dropped_spans") def __init__(self, maxlen): # type: (int) -> None @@ -204,11 +204,13 @@ def __init__(self, maxlen): # limits: either transaction+spans or only child spans. self.maxlen = maxlen - 1 self.spans = [] # type: List[Span] + self.dropped_spans = 0 # type: int def add(self, span): # type: (Span) -> None if len(self.spans) > self.maxlen: span._span_recorder = None + self.dropped_spans += 1 else: self.spans.append(span) @@ -972,6 +974,9 @@ def finish( if span.timestamp is not None ] + len_diff = len(self._span_recorder.spans) - len(finished_spans) + dropped_spans = len_diff + self._span_recorder.dropped_spans + # we do this to break the circular reference of transaction -> span # recorder -> span -> containing transaction (which is where we started) # before either the spans or the transaction goes out of scope and has @@ -996,6 +1001,9 @@ def finish( "spans": finished_spans, } # type: Event + if dropped_spans > 0: + event["_dropped_spans"] = dropped_spans + if self._profile is not None and self._profile.valid(): event["profile"] = self._profile self._profile = None diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index 3329b201b1..efc955ca7b 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -24,15 +24,13 @@ from sentry_sdk.worker import BackgroundWorker from sentry_sdk.envelope import Envelope, Item, PayloadRef -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, cast, List, Dict if TYPE_CHECKING: from typing import Any from typing import Callable - from typing import Dict from typing import DefaultDict from typing import Iterable - from typing import List from typing import Mapping from typing import Optional from typing import Self @@ -280,7 +278,9 @@ def record_lost_event( event = item.get_transaction_event() or {} # +1 for the transaction itself - span_count = len(event.get("spans") or []) + 1 + span_count = ( + len(cast(List[Dict[str, object]], event.get("spans") or [])) + 1 + ) self.record_lost_event(reason, "span", quantity=span_count) elif data_category == "attachment": diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 0fead48377..6a0e4579a1 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -32,6 +32,7 @@ DEFAULT_MAX_VALUE_LENGTH, EndpointType, ) +from sentry_sdk._types import Annotated, AnnotatedValue, SENSITIVE_DATA_SUBSTITUTE from typing import TYPE_CHECKING @@ -73,8 +74,6 @@ BASE64_ALPHABET = re.compile(r"^[a-zA-Z0-9/+=]*$") -SENSITIVE_DATA_SUBSTITUTE = "[Filtered]" - FALSY_ENV_VALUES = frozenset(("false", "f", "n", "no", "off", "0")) TRUTHY_ENV_VALUES = frozenset(("true", "t", "y", "yes", "on", "1")) @@ -404,84 +403,6 @@ def to_header(self): return "Sentry " + ", ".join("%s=%s" % (key, value) for key, value in rv) -class AnnotatedValue: - """ - Meta information for a data field in the event payload. - This is to tell Relay that we have tampered with the fields value. - See: - https://github.com/getsentry/relay/blob/be12cd49a0f06ea932ed9b9f93a655de5d6ad6d1/relay-general/src/types/meta.rs#L407-L423 - """ - - __slots__ = ("value", "metadata") - - def __init__(self, value, metadata): - # type: (Optional[Any], Dict[str, Any]) -> None - self.value = value - self.metadata = metadata - - def __eq__(self, other): - # type: (Any) -> bool - if not isinstance(other, AnnotatedValue): - return False - - return self.value == other.value and self.metadata == other.metadata - - @classmethod - def removed_because_raw_data(cls): - # type: () -> AnnotatedValue - """The value was removed because it could not be parsed. This is done for request body values that are not json nor a form.""" - return AnnotatedValue( - value="", - metadata={ - "rem": [ # Remark - [ - "!raw", # Unparsable raw data - "x", # The fields original value was removed - ] - ] - }, - ) - - @classmethod - def removed_because_over_size_limit(cls): - # type: () -> AnnotatedValue - """The actual value was removed because the size of the field exceeded the configured maximum size (specified with the max_request_body_size sdk option)""" - return AnnotatedValue( - value="", - metadata={ - "rem": [ # Remark - [ - "!config", # Because of configured maximum size - "x", # The fields original value was removed - ] - ] - }, - ) - - @classmethod - def substituted_because_contains_sensitive_data(cls): - # type: () -> AnnotatedValue - """The actual value was removed because it contained sensitive information.""" - return AnnotatedValue( - value=SENSITIVE_DATA_SUBSTITUTE, - metadata={ - "rem": [ # Remark - [ - "!config", # Because of SDK configuration (in this case the config is the hard coded removal of certain django cookies) - "s", # The fields original value was substituted - ] - ] - }, - ) - - -if TYPE_CHECKING: - from typing import TypeVar - - T = TypeVar("T") - Annotated = Union[AnnotatedValue, T] - - def get_type_name(cls): # type: (Optional[type]) -> Optional[str] return getattr(cls, "__qualname__", None) or getattr(cls, "__name__", None) diff --git a/tests/tracing/test_misc.py b/tests/tracing/test_misc.py index de2f782538..040fb24213 100644 --- a/tests/tracing/test_misc.py +++ b/tests/tracing/test_misc.py @@ -11,6 +11,7 @@ from sentry_sdk.tracing import Span, Transaction from sentry_sdk.tracing_utils import should_propagate_trace from sentry_sdk.utils import Dsn +from tests.conftest import ApproxDict def test_span_trimming(sentry_init, capture_events): @@ -31,6 +32,33 @@ def test_span_trimming(sentry_init, capture_events): assert span2["op"] == "foo1" assert span3["op"] == "foo2" + assert event["_meta"]["spans"][""]["len"] == 10 + assert "_dropped_spans" not in event + assert "dropped_spans" not in event + + +def test_span_data_scrubbing_and_trimming(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0, _experiments={"max_spans": 3}) + events = capture_events() + + with start_transaction(name="hi"): + with start_span(op="foo", name="bar") as span: + span.set_data("password", "secret") + span.set_data("datafoo", "databar") + + for i in range(10): + with start_span(op="foo{}".format(i)): + pass + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + {"password": "[Filtered]", "datafoo": "databar"} + ) + assert event["_meta"]["spans"] == { + "0": {"data": {"password": {"": {"rem": [["!config", "s"]]}}}}, + "": {"len": 11}, + } + def test_transaction_naming(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0) From 797e82ffb808cb0962c212b39b46204194aabdd9 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Tue, 4 Feb 2025 11:03:47 -0500 Subject: [PATCH 112/591] feat(profiling): Continuous profiling sample rate (#4002) This introduces a new top level setting for the continuous profiling session sample rate. The sample rate is evaluated once at the beginning and is used to determine whether or not the profiler will be run for the remainder of the process. --- sentry_sdk/consts.py | 1 + sentry_sdk/profiler/continuous_profiler.py | 81 +++++++++++------- tests/profiler/test_continuous_profiler.py | 95 ++++++++++++++++++---- 3 files changed, 134 insertions(+), 43 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 23f79ebd63..ce435de36b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -528,6 +528,7 @@ def __init__( profiles_sample_rate=None, # type: Optional[float] profiles_sampler=None, # type: Optional[TracesSampler] profiler_mode=None, # type: Optional[ProfilerMode] + profile_session_sample_rate=None, # type: Optional[float] auto_enabling_integrations=True, # type: bool disabled_integrations=None, # type: Optional[Sequence[sentry_sdk.integrations.Integration]] auto_session_tracking=True, # type: bool diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 5a76a0696c..b07fbec998 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -1,5 +1,6 @@ import atexit import os +import random import sys import threading import time @@ -83,11 +84,15 @@ def setup_continuous_profiler(options, sdk_info, capture_func): else: default_profiler_mode = ThreadContinuousScheduler.mode - experiments = options.get("_experiments", {}) + if options.get("profiler_mode") is not None: + profiler_mode = options["profiler_mode"] + else: + # TODO: deprecate this and just use the existing `profiler_mode` + experiments = options.get("_experiments", {}) - profiler_mode = ( - experiments.get("continuous_profiling_mode") or default_profiler_mode - ) + profiler_mode = ( + experiments.get("continuous_profiling_mode") or default_profiler_mode + ) frequency = DEFAULT_SAMPLING_FREQUENCY @@ -118,19 +123,10 @@ def try_autostart_continuous_profiler(): if _scheduler is None: return - # Ensure that the scheduler only autostarts once per process. - # This is necessary because many web servers use forks to spawn - # additional processes. And the profiler is only spawned on the - # master process, then it often only profiles the main process - # and not the ones where the requests are being handled. - # - # Additionally, we only want this autostart behaviour once per - # process. If the user explicitly calls `stop_profiler`, it should - # be respected and not start the profiler again. - if not _scheduler.should_autostart(): + if not _scheduler.is_auto_start_enabled(): return - _scheduler.ensure_running() + _scheduler.manual_start() def start_profiler(): @@ -138,7 +134,7 @@ def start_profiler(): if _scheduler is None: return - _scheduler.ensure_running() + _scheduler.manual_start() def stop_profiler(): @@ -146,7 +142,7 @@ def stop_profiler(): if _scheduler is None: return - _scheduler.teardown() + _scheduler.manual_stop() def teardown_continuous_profiler(): @@ -164,6 +160,16 @@ def get_profiler_id(): return _scheduler.profiler_id +def determine_profile_session_sampling_decision(sample_rate): + # type: (Union[float, None]) -> bool + + # `None` is treated as `0.0` + if not sample_rate: + return False + + return random.random() < float(sample_rate) + + class ContinuousScheduler: mode = "unknown" # type: ContinuousProfilerMode @@ -175,16 +181,43 @@ def __init__(self, frequency, options, sdk_info, capture_func): self.capture_func = capture_func self.sampler = self.make_sampler() self.buffer = None # type: Optional[ProfileBuffer] + self.pid = None # type: Optional[int] self.running = False - def should_autostart(self): + profile_session_sample_rate = self.options.get("profile_session_sample_rate") + self.sampled = determine_profile_session_sampling_decision( + profile_session_sample_rate + ) + + def is_auto_start_enabled(self): # type: () -> bool + + # Ensure that the scheduler only autostarts once per process. + # This is necessary because many web servers use forks to spawn + # additional processes. And the profiler is only spawned on the + # master process, then it often only profiles the main process + # and not the ones where the requests are being handled. + if self.pid == os.getpid(): + return False + experiments = self.options.get("_experiments") if not experiments: return False + return experiments.get("continuous_profiling_auto_start") + def manual_start(self): + # type: () -> None + if not self.sampled: + return + + self.ensure_running() + + def manual_stop(self): + # type: () -> None + self.teardown() + def ensure_running(self): # type: () -> None raise NotImplementedError @@ -277,15 +310,11 @@ def __init__(self, frequency, options, sdk_info, capture_func): super().__init__(frequency, options, sdk_info, capture_func) self.thread = None # type: Optional[threading.Thread] - self.pid = None # type: Optional[int] self.lock = threading.Lock() - def should_autostart(self): - # type: () -> bool - return super().should_autostart() and self.pid != os.getpid() - def ensure_running(self): # type: () -> None + pid = os.getpid() # is running on the right process @@ -356,13 +385,8 @@ def __init__(self, frequency, options, sdk_info, capture_func): super().__init__(frequency, options, sdk_info, capture_func) self.thread = None # type: Optional[_ThreadPool] - self.pid = None # type: Optional[int] self.lock = threading.Lock() - def should_autostart(self): - # type: () -> bool - return super().should_autostart() and self.pid != os.getpid() - def ensure_running(self): # type: () -> None pid = os.getpid() @@ -393,7 +417,6 @@ def ensure_running(self): # longer allows us to spawn a thread and we have to bail. self.running = False self.thread = None - return def teardown(self): # type: () -> None diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 32d0e8d0b0..6f4893e59d 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -23,13 +23,25 @@ requires_gevent = pytest.mark.skipif(gevent is None, reason="gevent not enabled") -def experimental_options(mode=None, auto_start=None): - return { - "_experiments": { - "continuous_profiling_auto_start": auto_start, - "continuous_profiling_mode": mode, +def get_client_options(use_top_level_profiler_mode): + def client_options(mode=None, auto_start=None, profile_session_sample_rate=1.0): + if use_top_level_profiler_mode: + return { + "profiler_mode": mode, + "profile_session_sample_rate": profile_session_sample_rate, + "_experiments": { + "continuous_profiling_auto_start": auto_start, + }, + } + return { + "profile_session_sample_rate": profile_session_sample_rate, + "_experiments": { + "continuous_profiling_auto_start": auto_start, + "continuous_profiling_mode": mode, + }, } - } + + return client_options mock_sdk_info = { @@ -42,7 +54,10 @@ def experimental_options(mode=None, auto_start=None): @pytest.mark.parametrize("mode", [pytest.param("foo")]) @pytest.mark.parametrize( "make_options", - [pytest.param(experimental_options, id="experiment")], + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], ) def test_continuous_profiler_invalid_mode(mode, make_options, teardown_profiling): with pytest.raises(ValueError): @@ -62,7 +77,10 @@ def test_continuous_profiler_invalid_mode(mode, make_options, teardown_profiling ) @pytest.mark.parametrize( "make_options", - [pytest.param(experimental_options, id="experiment")], + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], ) def test_continuous_profiler_valid_mode(mode, make_options, teardown_profiling): options = make_options(mode=mode) @@ -82,7 +100,10 @@ def test_continuous_profiler_valid_mode(mode, make_options, teardown_profiling): ) @pytest.mark.parametrize( "make_options", - [pytest.param(experimental_options, id="experiment")], + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], ) def test_continuous_profiler_setup_twice(mode, make_options, teardown_profiling): options = make_options(mode=mode) @@ -178,7 +199,10 @@ def assert_single_transaction_without_profile_chunks(envelopes): ) @pytest.mark.parametrize( "make_options", - [pytest.param(experimental_options, id="experiment")], + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], ) @mock.patch("sentry_sdk.profiler.continuous_profiler.PROFILE_BUFFER_SECONDS", 0.01) def test_continuous_profiler_auto_start_and_manual_stop( @@ -191,7 +215,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( options = make_options(mode=mode, auto_start=True) sentry_init( traces_sample_rate=1.0, - _experiments=options.get("_experiments", {}), + **options, ) envelopes = capture_envelopes() @@ -235,10 +259,13 @@ def test_continuous_profiler_auto_start_and_manual_stop( ) @pytest.mark.parametrize( "make_options", - [pytest.param(experimental_options, id="experiment")], + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], ) @mock.patch("sentry_sdk.profiler.continuous_profiler.PROFILE_BUFFER_SECONDS", 0.01) -def test_continuous_profiler_manual_start_and_stop( +def test_continuous_profiler_manual_start_and_stop_sampled( sentry_init, capture_envelopes, mode, @@ -248,7 +275,7 @@ def test_continuous_profiler_manual_start_and_stop( options = make_options(mode=mode) sentry_init( traces_sample_rate=1.0, - _experiments=options.get("_experiments", {}), + **options, ) envelopes = capture_envelopes() @@ -275,3 +302,43 @@ def test_continuous_profiler_manual_start_and_stop( time.sleep(0.05) assert_single_transaction_without_profile_chunks(envelopes) + + +@pytest.mark.parametrize( + "mode", + [ + pytest.param("thread"), + pytest.param("gevent", marks=requires_gevent), + ], +) +@pytest.mark.parametrize( + "make_options", + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], +) +def test_continuous_profiler_manual_start_and_stop_unsampled( + sentry_init, + capture_envelopes, + mode, + make_options, + teardown_profiling, +): + options = make_options(mode=mode, profile_session_sample_rate=0.0) + sentry_init( + traces_sample_rate=1.0, + **options, + ) + + envelopes = capture_envelopes() + + start_profiler() + + with sentry_sdk.start_transaction(name="profiling"): + with sentry_sdk.start_span(op="op"): + time.sleep(0.05) + + assert_single_transaction_without_profile_chunks(envelopes) + + stop_profiler() From 1fd2b86a6be0b637fce3a0dc0da3962b58f20cc6 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 6 Feb 2025 13:32:39 +0100 Subject: [PATCH 113/591] Fix mypy (#4019) mypy is unhappy in CI, fix it. --- sentry_sdk/integrations/grpc/__init__.py | 6 +++--- sentry_sdk/integrations/socket.py | 12 ++++++++---- sentry_sdk/integrations/tornado.py | 2 +- 3 files changed, 12 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/integrations/grpc/__init__.py b/sentry_sdk/integrations/grpc/__init__.py index 3d949091eb..d9dcdddb55 100644 --- a/sentry_sdk/integrations/grpc/__init__.py +++ b/sentry_sdk/integrations/grpc/__init__.py @@ -81,7 +81,7 @@ def _wrap_channel_async(func: Callable[P, AsyncChannel]) -> Callable[P, AsyncCha "Wrapper for asynchronous secure and insecure channel." @wraps(func) - def patched_channel( + def patched_channel( # type: ignore *args: P.args, interceptors: Optional[Sequence[grpc.aio.ClientInterceptor]] = None, **kwargs: P.kwargs, @@ -100,7 +100,7 @@ def _wrap_sync_server(func: Callable[P, Server]) -> Callable[P, Server]: """Wrapper for synchronous server.""" @wraps(func) - def patched_server( + def patched_server( # type: ignore *args: P.args, interceptors: Optional[Sequence[grpc.ServerInterceptor]] = None, **kwargs: P.kwargs, @@ -121,7 +121,7 @@ def _wrap_async_server(func: Callable[P, AsyncServer]) -> Callable[P, AsyncServe """Wrapper for asynchronous server.""" @wraps(func) - def patched_aio_server( + def patched_aio_server( # type: ignore *args: P.args, interceptors: Optional[Sequence[grpc.ServerInterceptor]] = None, **kwargs: P.kwargs, diff --git a/sentry_sdk/integrations/socket.py b/sentry_sdk/integrations/socket.py index 0866ceb608..babf61aa7a 100644 --- a/sentry_sdk/integrations/socket.py +++ b/sentry_sdk/integrations/socket.py @@ -27,15 +27,19 @@ def setup_once(): def _get_span_description(host, port): - # type: (Union[bytes, str, None], Union[str, int, None]) -> str + # type: (Union[bytes, str, None], Union[bytes, str, int, None]) -> str try: host = host.decode() # type: ignore except (UnicodeDecodeError, AttributeError): pass - description = "%s:%s" % (host, port) # type: ignore + try: + port = port.decode() # type: ignore + except (UnicodeDecodeError, AttributeError): + pass + description = "%s:%s" % (host, port) # type: ignore return description @@ -74,7 +78,7 @@ def _patch_getaddrinfo(): real_getaddrinfo = socket.getaddrinfo def getaddrinfo(host, port, family=0, type=0, proto=0, flags=0): - # type: (Union[bytes, str, None], Union[str, int, None], int, int, int, int) -> List[Tuple[AddressFamily, SocketKind, int, str, Union[Tuple[str, int], Tuple[str, int, int, int]]]] + # type: (Union[bytes, str, None], Union[bytes, str, int, None], int, int, int, int) -> List[Tuple[AddressFamily, SocketKind, int, str, Union[Tuple[str, int], Tuple[str, int, int, int], Tuple[int, bytes]]]] integration = sentry_sdk.get_client().get_integration(SocketIntegration) if integration is None: return real_getaddrinfo(host, port, family, type, proto, flags) @@ -89,4 +93,4 @@ def getaddrinfo(host, port, family=0, type=0, proto=0, flags=0): return real_getaddrinfo(host, port, family, type, proto, flags) - socket.getaddrinfo = getaddrinfo # type: ignore + socket.getaddrinfo = getaddrinfo diff --git a/sentry_sdk/integrations/tornado.py b/sentry_sdk/integrations/tornado.py index b9e465c7c7..0f0f64d1a1 100644 --- a/sentry_sdk/integrations/tornado.py +++ b/sentry_sdk/integrations/tornado.py @@ -79,7 +79,7 @@ async def sentry_execute_request_handler(self, *args, **kwargs): else: @coroutine # type: ignore - def sentry_execute_request_handler(self, *args, **kwargs): + def sentry_execute_request_handler(self, *args, **kwargs): # type: ignore # type: (RequestHandler, *Any, **Any) -> Any with _handle_request_impl(self): result = yield from old_execute(self, *args, **kwargs) From ab36fc41b80eaba821cf8be4017108462675bd69 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 6 Feb 2025 15:31:14 +0100 Subject: [PATCH 114/591] build(deps): bump actions/create-github-app-token from 1.11.1 to 1.11.2 (#4015) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.1 to 1.11.2.
Release notes

Sourced from actions/create-github-app-token's releases.

v1.11.2

1.11.2 (2025-01-30)

Bug Fixes

Commits
  • 136412a build(release): 1.11.2 [skip ci]
  • b4192a5 fix(deps): bump @​octokit/request from 9.1.3 to 9.1.4 in the production-depend...
  • 29aa051 fix(deps): bump undici from 6.19.8 to 7.2.0 (#198)
  • a5f8600 build(deps-dev): bump @​sinonjs/fake-timers from 13.0.2 to 14.0.0 (#199)
  • 0edddd7 build(deps-dev): bump the development-dependencies group with 2 updates (#197)
  • bb3ca76 docs(README): remove extra space in variable syntax in README example (#201)
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=1.11.1&new-version=1.11.2)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 6450150138..9886ee74e5 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@c1a285145b9d317df6ced56c09f525b5c2b6f755 # v1.11.1 + uses: actions/create-github-app-token@136412a57a7081aa63c935a2cc2918f76c34f514 # v1.11.2 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From bc72f78eea76a77bfd4b445a0424767223d76787 Mon Sep 17 00:00:00 2001 From: Lev Vereshchagin Date: Thu, 6 Feb 2025 18:12:32 +0300 Subject: [PATCH 115/591] feat(litestar): Add `failed_request_status_codes` (#4021) --- sentry_sdk/integrations/litestar.py | 22 ++++++++++++- tests/integrations/conftest.py | 21 +++++++++++++ tests/integrations/fastapi/test_fastapi.py | 3 +- tests/integrations/litestar/test_litestar.py | 31 +++++++++++++++++++ .../integrations/starlette/test_starlette.py | 23 ++------------ 5 files changed, 77 insertions(+), 23 deletions(-) diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 4b04dada8a..841c8a5cce 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -1,6 +1,11 @@ +from collections.abc import Set import sentry_sdk from sentry_sdk.consts import OP -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.integrations import ( + _DEFAULT_FAILED_REQUEST_STATUS_CODES, + DidNotEnable, + Integration, +) from sentry_sdk.integrations.asgi import SentryAsgiMiddleware from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii @@ -17,6 +22,7 @@ from litestar.middleware import DefineMiddleware # type: ignore from litestar.routes.http import HTTPRoute # type: ignore from litestar.data_extractors import ConnectionDataExtractor # type: ignore + from litestar.exceptions import HTTPException # type: ignore except ImportError: raise DidNotEnable("Litestar is not installed") @@ -45,6 +51,12 @@ class LitestarIntegration(Integration): identifier = "litestar" origin = f"auto.http.{identifier}" + def __init__( + self, + failed_request_status_codes=_DEFAULT_FAILED_REQUEST_STATUS_CODES, # type: Set[int] + ) -> None: + self.failed_request_status_codes = failed_request_status_codes + @staticmethod def setup_once(): # type: () -> None @@ -277,6 +289,14 @@ def exception_handler(exc, scope): sentry_scope = sentry_sdk.get_isolation_scope() sentry_scope.set_user(user_info) + if isinstance(exc, HTTPException): + integration = sentry_sdk.get_client().get_integration(LitestarIntegration) + if ( + integration is not None + and exc.status_code not in integration.failed_request_status_codes + ): + return + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, diff --git a/tests/integrations/conftest.py b/tests/integrations/conftest.py index 560155e2b5..7ac43b0efe 100644 --- a/tests/integrations/conftest.py +++ b/tests/integrations/conftest.py @@ -32,3 +32,24 @@ def capture_event_scope(self, event, hint=None, scope=None): return errors return inner + + +parametrize_test_configurable_status_codes = pytest.mark.parametrize( + ("failed_request_status_codes", "status_code", "expected_error"), + ( + (None, 500, True), + (None, 400, False), + ({500, 501}, 500, True), + ({500, 501}, 401, False), + ({*range(400, 500)}, 401, True), + ({*range(400, 500)}, 500, False), + ({*range(400, 600)}, 300, False), + ({*range(400, 600)}, 403, True), + ({*range(400, 600)}, 503, True), + ({*range(400, 403), 500, 501}, 401, True), + ({*range(400, 403), 500, 501}, 405, False), + ({*range(400, 403), 500, 501}, 501, True), + ({*range(400, 403), 500, 501}, 503, False), + (set(), 500, False), + ), +) diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index 97aea06344..f1c0a69305 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -19,6 +19,7 @@ FASTAPI_VERSION = parse_version(fastapi.__version__) +from tests.integrations.conftest import parametrize_test_configurable_status_codes from tests.integrations.starlette import test_starlette @@ -650,7 +651,7 @@ def test_transaction_http_method_custom(sentry_init, capture_events): assert event2["request"]["method"] == "HEAD" -@test_starlette.parametrize_test_configurable_status_codes +@parametrize_test_configurable_status_codes def test_configurable_status_codes( sentry_init, capture_events, diff --git a/tests/integrations/litestar/test_litestar.py b/tests/integrations/litestar/test_litestar.py index 90346537a7..4f642479e4 100644 --- a/tests/integrations/litestar/test_litestar.py +++ b/tests/integrations/litestar/test_litestar.py @@ -1,6 +1,7 @@ from __future__ import annotations import functools +from litestar.exceptions import HTTPException import pytest from sentry_sdk import capture_message @@ -16,6 +17,8 @@ from litestar.middleware.session.server_side import ServerSideSessionConfig from litestar.testing import TestClient +from tests.integrations.conftest import parametrize_test_configurable_status_codes + def litestar_app_factory(middleware=None, debug=True, exception_handlers=None): class MyController(Controller): @@ -396,3 +399,31 @@ async def __call__(self, scope, receive, send): } else: assert "user" not in event + + +@parametrize_test_configurable_status_codes +def test_configurable_status_codes( + sentry_init, + capture_events, + failed_request_status_codes, + status_code, + expected_error, +): + integration_kwargs = ( + {"failed_request_status_codes": failed_request_status_codes} + if failed_request_status_codes is not None + else {} + ) + sentry_init(integrations=[LitestarIntegration(**integration_kwargs)]) + + events = capture_events() + + @get("/error") + async def error() -> None: + raise HTTPException(status_code=status_code) + + app = Litestar([error]) + client = TestClient(app) + client.get("/error") + + assert len(events) == int(expected_error) diff --git a/tests/integrations/starlette/test_starlette.py b/tests/integrations/starlette/test_starlette.py index fd47895f5a..93da0420aa 100644 --- a/tests/integrations/starlette/test_starlette.py +++ b/tests/integrations/starlette/test_starlette.py @@ -32,6 +32,8 @@ from starlette.middleware.trustedhost import TrustedHostMiddleware from starlette.testclient import TestClient +from tests.integrations.conftest import parametrize_test_configurable_status_codes + STARLETTE_VERSION = parse_version(starlette.__version__) @@ -1298,27 +1300,6 @@ def test_transaction_http_method_custom(sentry_init, capture_events): assert event2["request"]["method"] == "HEAD" -parametrize_test_configurable_status_codes = pytest.mark.parametrize( - ("failed_request_status_codes", "status_code", "expected_error"), - ( - (None, 500, True), - (None, 400, False), - ({500, 501}, 500, True), - ({500, 501}, 401, False), - ({*range(400, 500)}, 401, True), - ({*range(400, 500)}, 500, False), - ({*range(400, 600)}, 300, False), - ({*range(400, 600)}, 403, True), - ({*range(400, 600)}, 503, True), - ({*range(400, 403), 500, 501}, 401, True), - ({*range(400, 403), 500, 501}, 405, False), - ({*range(400, 403), 500, 501}, 501, True), - ({*range(400, 403), 500, 501}, 503, False), - (set(), 500, False), - ), -) - - @parametrize_test_configurable_status_codes def test_configurable_status_codes( sentry_init, From d670a150c470ef551120d89ec205a4af9df8b4b6 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 10 Feb 2025 13:51:55 +0100 Subject: [PATCH 116/591] Don't set transaction status to error on sys.exit(0) (#4025) We set transaction status to `internal_error` if there is an exception exiting the `start_transaction` context manager. We don't check what kind of exception it was. Some exceptions aren't a sign of anything wrong, like `SystemExit` with a value of 0, so we shouldn't mark the transaction as failed in that case. Closes https://github.com/getsentry/sentry-python/issues/4024 --- sentry_sdk/tracing.py | 3 +- sentry_sdk/utils.py | 9 ++++ tests/tracing/test_integration_tests.py | 58 ++++++++++++++++++++++++- 3 files changed, 67 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 86456b8964..59473d752c 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -11,6 +11,7 @@ is_valid_sample_rate, logger, nanosecond_time, + should_be_treated_as_error, ) from typing import TYPE_CHECKING @@ -374,7 +375,7 @@ def __enter__(self): def __exit__(self, ty, value, tb): # type: (Optional[Any], Optional[Any], Optional[Any]) -> None - if value is not None: + if value is not None and should_be_treated_as_error(ty, value): self.set_status(SPANSTATUS.INTERNAL_ERROR) scope, old_span = self._context_manager_state diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 6a0e4579a1..f60c31e676 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1879,3 +1879,12 @@ def get_current_thread_meta(thread=None): # we've tried everything, time to give up return None, None + + +def should_be_treated_as_error(ty, value): + # type: (Any, Any) -> bool + if ty == SystemExit and hasattr(value, "code") and value.code in (0, None): + # https://docs.python.org/3/library/exceptions.html#SystemExit + return False + + return True diff --git a/tests/tracing/test_integration_tests.py b/tests/tracing/test_integration_tests.py index da3efef9eb..f269023f87 100644 --- a/tests/tracing/test_integration_tests.py +++ b/tests/tracing/test_integration_tests.py @@ -1,8 +1,10 @@ -import weakref import gc +import random import re +import sys +import weakref + import pytest -import random import sentry_sdk from sentry_sdk import ( @@ -297,3 +299,55 @@ def test_trace_propagation_meta_head_sdk(sentry_init): assert 'meta name="baggage"' in baggage baggage_content = re.findall('content="([^"]*)"', baggage)[0] assert baggage_content == transaction.get_baggage().serialize() + + +@pytest.mark.parametrize( + "exception_cls,exception_value", + [ + (SystemExit, 0), + ], +) +def test_non_error_exceptions( + sentry_init, capture_events, exception_cls, exception_value +): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="hi") as transaction: + transaction.set_status(SPANSTATUS.OK) + with pytest.raises(exception_cls): + with start_span(op="foo", name="foodesc"): + raise exception_cls(exception_value) + + assert len(events) == 1 + event = events[0] + + span = event["spans"][0] + assert "status" not in span.get("tags", {}) + assert "status" not in event["tags"] + assert event["contexts"]["trace"]["status"] == "ok" + + +@pytest.mark.parametrize("exception_value", [None, 0, False]) +def test_good_sysexit_doesnt_fail_transaction( + sentry_init, capture_events, exception_value +): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="hi") as transaction: + transaction.set_status(SPANSTATUS.OK) + with pytest.raises(SystemExit): + with start_span(op="foo", name="foodesc"): + if exception_value is not False: + sys.exit(exception_value) + else: + sys.exit() + + assert len(events) == 1 + event = events[0] + + span = event["spans"][0] + assert "status" not in span.get("tags", {}) + assert "status" not in event["tags"] + assert event["contexts"]["trace"]["status"] == "ok" From 5fb97a92b278477cfdb8049f9dc35af892cf1be5 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 11 Feb 2025 10:56:57 +0100 Subject: [PATCH 117/591] build(deps): bump actions/create-github-app-token from 1.11.2 to 1.11.3 (#4023) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.2 to 1.11.3.
Release notes

Sourced from actions/create-github-app-token's releases.

v1.11.3

1.11.3 (2025-02-04)

Bug Fixes

Commits
  • 67e27a7 build(release): 1.11.3 [skip ci]
  • 8e85a3c fix(deps): bump the production-dependencies group with 3 updates (#203)
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=1.11.2&new-version=1.11.3)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 9886ee74e5..ae9ae279c7 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@136412a57a7081aa63c935a2cc2918f76c34f514 # v1.11.2 + uses: actions/create-github-app-token@67e27a7eb7db372a1c61a7f9bdab8699e9ee57f7 # v1.11.3 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From c1cf0fef79db0d7ebe5c640ab0fe0f7ae06c9d21 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 11 Feb 2025 10:58:04 +0100 Subject: [PATCH 118/591] Set level based on status code for HTTP client breadcrumbs (#4004) - add logic to `maybe_create_breadcrumbs_from_span` to set the `level` of the breadcrumb to `warning` for the client error range (4xx) and to `error` for server errors (5xx) - add functionality to the simple HTTP server that we use in some tests to respond with a specific error code - we were (and are) still "using" `responses` in multiple places, but they're not actually active (the `activate` decorator is missing) and we're making actual requests outside -- we should clean this up - we also can't use `responses` for stdlib/requests tests since they patch something that we patch - add `httpx`, `stdlib`, `requests`, `aiohttp` tests for the new behavior - restrict the `requests` tests to 3.7+ since in 3.6, the span is finished before the HTTP status is set for some reason... Closes https://github.com/getsentry/sentry-python/issues/4000 --- sentry_sdk/tracing_utils.py | 18 +++++- tests/conftest.py | 10 +++- tests/integrations/aiohttp/test_aiohttp.py | 55 ++++++++++++++++++ tests/integrations/httpx/test_httpx.py | 58 +++++++++++++++++++ tests/integrations/requests/test_requests.py | 61 +++++++++++++++++--- tests/integrations/stdlib/test_httplib.py | 45 +++++++++++++++ 6 files changed, 235 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 0459563776..9ea2d9859a 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -156,13 +156,27 @@ def record_sql_queries( def maybe_create_breadcrumbs_from_span(scope, span): # type: (sentry_sdk.Scope, sentry_sdk.tracing.Span) -> None - if span.op == OP.DB_REDIS: scope.add_breadcrumb( message=span.description, type="redis", category="redis", data=span._tags ) + elif span.op == OP.HTTP_CLIENT: - scope.add_breadcrumb(type="http", category="httplib", data=span._data) + level = None + status_code = span._data.get(SPANDATA.HTTP_STATUS_CODE) + if status_code: + if 500 <= status_code <= 599: + level = "error" + elif 400 <= status_code <= 499: + level = "warning" + + if level: + scope.add_breadcrumb( + type="http", category="httplib", data=span._data, level=level + ) + else: + scope.add_breadcrumb(type="http", category="httplib", data=span._data) + elif span.op == "subprocess": scope.add_breadcrumb( type="subprocess", diff --git a/tests/conftest.py b/tests/conftest.py index b5ab7aa804..b5f3f8b00e 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -587,8 +587,14 @@ def suppress_deprecation_warnings(): class MockServerRequestHandler(BaseHTTPRequestHandler): def do_GET(self): # noqa: N802 - # Process an HTTP GET request and return a response with an HTTP 200 status. - self.send_response(200) + # Process an HTTP GET request and return a response. + # If the path ends with /status/, return status code . + # Otherwise return a 200 response. + code = 200 + if "/status/" in self.path: + code = int(self.path[-3:]) + + self.send_response(code) self.end_headers() return diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index b689e3af17..83dc021844 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -525,6 +525,61 @@ async def handler(request): ) +@pytest.mark.parametrize( + "status_code,level", + [ + (200, None), + (301, None), + (403, "warning"), + (405, "warning"), + (500, "error"), + ], +) +@pytest.mark.asyncio +async def test_crumb_capture_client_error( + sentry_init, + aiohttp_raw_server, + aiohttp_client, + event_loop, + capture_events, + status_code, + level, +): + sentry_init(integrations=[AioHttpIntegration()]) + + async def handler(request): + return web.Response(status=status_code) + + raw_server = await aiohttp_raw_server(handler) + + with start_transaction(): + events = capture_events() + + client = await aiohttp_client(raw_server) + resp = await client.get("/") + assert resp.status == status_code + capture_message("Testing!") + + (event,) = events + + crumb = event["breadcrumbs"]["values"][0] + assert crumb["type"] == "http" + if level is None: + assert "level" not in crumb + else: + assert crumb["level"] == level + assert crumb["category"] == "httplib" + assert crumb["data"] == ApproxDict( + { + "url": "http://127.0.0.1:{}/".format(raw_server.port), + "http.fragment": "", + "http.method": "GET", + "http.query": "", + "http.response.status_code": status_code, + } + ) + + @pytest.mark.asyncio async def test_outgoing_trace_headers(sentry_init, aiohttp_raw_server, aiohttp_client): sentry_init( diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index 107f873a3c..d37e1fddf2 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -57,6 +57,64 @@ def before_breadcrumb(crumb, hint): ) +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +@pytest.mark.parametrize( + "status_code,level", + [ + (200, None), + (301, None), + (403, "warning"), + (405, "warning"), + (500, "error"), + ], +) +def test_crumb_capture_client_error( + sentry_init, capture_events, httpx_client, httpx_mock, status_code, level +): + httpx_mock.add_response(status_code=status_code) + + sentry_init(integrations=[HttpxIntegration()]) + + url = "http://example.com/" + + with start_transaction(): + events = capture_events() + + if asyncio.iscoroutinefunction(httpx_client.get): + response = asyncio.get_event_loop().run_until_complete( + httpx_client.get(url) + ) + else: + response = httpx_client.get(url) + + assert response.status_code == status_code + capture_message("Testing!") + + (event,) = events + + crumb = event["breadcrumbs"]["values"][0] + assert crumb["type"] == "http" + assert crumb["category"] == "httplib" + + if level is None: + assert "level" not in crumb + else: + assert crumb["level"] == level + + assert crumb["data"] == ApproxDict( + { + "url": url, + SPANDATA.HTTP_METHOD: "GET", + SPANDATA.HTTP_FRAGMENT: "", + SPANDATA.HTTP_QUERY: "", + SPANDATA.HTTP_STATUS_CODE: status_code, + } + ) + + @pytest.mark.parametrize( "httpx_client", (httpx.Client(), httpx.AsyncClient()), diff --git a/tests/integrations/requests/test_requests.py b/tests/integrations/requests/test_requests.py index 42efbb5acc..8cfc0f932f 100644 --- a/tests/integrations/requests/test_requests.py +++ b/tests/integrations/requests/test_requests.py @@ -1,30 +1,77 @@ +import sys from unittest import mock import pytest import requests -import responses from sentry_sdk import capture_message from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.stdlib import StdlibIntegration -from tests.conftest import ApproxDict +from tests.conftest import ApproxDict, create_mock_http_server + +PORT = create_mock_http_server() def test_crumb_capture(sentry_init, capture_events): sentry_init(integrations=[StdlibIntegration()]) + events = capture_events() - url = "http://example.com/" - responses.add(responses.GET, url, status=200) + url = f"http://localhost:{PORT}/hello-world" # noqa:E231 + response = requests.get(url) + capture_message("Testing!") + + (event,) = events + (crumb,) = event["breadcrumbs"]["values"] + assert crumb["type"] == "http" + assert crumb["category"] == "httplib" + assert crumb["data"] == ApproxDict( + { + "url": url, + SPANDATA.HTTP_METHOD: "GET", + SPANDATA.HTTP_FRAGMENT: "", + SPANDATA.HTTP_QUERY: "", + SPANDATA.HTTP_STATUS_CODE: response.status_code, + "reason": response.reason, + } + ) + + +@pytest.mark.skipif( + sys.version_info < (3, 7), + reason="The response status is not set on the span early enough in 3.6", +) +@pytest.mark.parametrize( + "status_code,level", + [ + (200, None), + (301, None), + (403, "warning"), + (405, "warning"), + (500, "error"), + ], +) +def test_crumb_capture_client_error(sentry_init, capture_events, status_code, level): + sentry_init(integrations=[StdlibIntegration()]) events = capture_events() + url = f"http://localhost:{PORT}/status/{status_code}" # noqa:E231 response = requests.get(url) + + assert response.status_code == status_code + capture_message("Testing!") (event,) = events (crumb,) = event["breadcrumbs"]["values"] assert crumb["type"] == "http" assert crumb["category"] == "httplib" + + if level is None: + assert "level" not in crumb + else: + assert crumb["level"] == level + assert crumb["data"] == ApproxDict( { "url": url, @@ -41,11 +88,10 @@ def test_crumb_capture(sentry_init, capture_events): def test_omit_url_data_if_parsing_fails(sentry_init, capture_events): sentry_init(integrations=[StdlibIntegration()]) - url = "https://example.com/" - responses.add(responses.GET, url, status=200) - events = capture_events() + url = f"http://localhost:{PORT}/ok" # noqa:E231 + with mock.patch( "sentry_sdk.integrations.stdlib.parse_url", side_effect=ValueError, @@ -63,7 +109,6 @@ def test_omit_url_data_if_parsing_fails(sentry_init, capture_events): # no url related data } ) - assert "url" not in event["breadcrumbs"]["values"][0]["data"] assert SPANDATA.HTTP_FRAGMENT not in event["breadcrumbs"]["values"][0]["data"] assert SPANDATA.HTTP_QUERY not in event["breadcrumbs"]["values"][0]["data"] diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 200b282f53..7f2c5d68b2 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -1,6 +1,7 @@ import random from http.client import HTTPConnection, HTTPSConnection from socket import SocketIO +from urllib.error import HTTPError from urllib.request import urlopen from unittest import mock @@ -42,6 +43,50 @@ def test_crumb_capture(sentry_init, capture_events): ) +@pytest.mark.parametrize( + "status_code,level", + [ + (200, None), + (301, None), + (403, "warning"), + (405, "warning"), + (500, "error"), + ], +) +def test_crumb_capture_client_error(sentry_init, capture_events, status_code, level): + sentry_init(integrations=[StdlibIntegration()]) + events = capture_events() + + url = f"http://localhost:{PORT}/status/{status_code}" # noqa:E231 + try: + urlopen(url) + except HTTPError: + pass + + capture_message("Testing!") + + (event,) = events + (crumb,) = event["breadcrumbs"]["values"] + + assert crumb["type"] == "http" + assert crumb["category"] == "httplib" + + if level is None: + assert "level" not in crumb + else: + assert crumb["level"] == level + + assert crumb["data"] == ApproxDict( + { + "url": url, + SPANDATA.HTTP_METHOD: "GET", + SPANDATA.HTTP_STATUS_CODE: status_code, + SPANDATA.HTTP_FRAGMENT: "", + SPANDATA.HTTP_QUERY: "", + } + ) + + def test_crumb_capture_hint(sentry_init, capture_events): def before_breadcrumb(crumb, hint): crumb["data"]["extra"] = "foo" From f995d8c191667380c41f339544b40443c0ee4453 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 11 Feb 2025 11:13:18 +0100 Subject: [PATCH 119/591] [1] Add tox generation script, but don't use it yet (#3971) Add: * tox generation script * tox template * script for generating tox and CI yamls in one go * readme for the script In this PR, the script is set to ignore all integrations, so no tox configuration is actually added. However, it's still the script actually generating the real `tox.ini` from the `tox.jinja` template. See follow-up PRs for more. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- scripts/generate-test-files.sh | 17 + scripts/populate_tox/README.md | 159 +++++ scripts/populate_tox/config.py | 8 + scripts/populate_tox/populate_tox.py | 548 ++++++++++++++++ scripts/populate_tox/requirements.txt | 3 + scripts/populate_tox/tox.jinja | 899 ++++++++++++++++++++++++++ tox.ini | 17 + 7 files changed, 1651 insertions(+) create mode 100755 scripts/generate-test-files.sh create mode 100644 scripts/populate_tox/README.md create mode 100644 scripts/populate_tox/config.py create mode 100644 scripts/populate_tox/populate_tox.py create mode 100644 scripts/populate_tox/requirements.txt create mode 100644 scripts/populate_tox/tox.jinja diff --git a/scripts/generate-test-files.sh b/scripts/generate-test-files.sh new file mode 100755 index 0000000000..40e279cdf4 --- /dev/null +++ b/scripts/generate-test-files.sh @@ -0,0 +1,17 @@ +#!/bin/sh + +# This script generates tox.ini and CI YAML files in one go. + +set -xe + +cd "$(dirname "$0")" + +python -m venv toxgen.venv +. toxgen.venv/bin/activate + +pip install -e .. +pip install -r populate_tox/requirements.txt +pip install -r split_tox_gh_actions/requirements.txt + +python populate_tox/populate_tox.py +python split_tox_gh_actions/split_tox_gh_actions.py diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md new file mode 100644 index 0000000000..aa9884387e --- /dev/null +++ b/scripts/populate_tox/README.md @@ -0,0 +1,159 @@ +# Populate Tox + +We integrate with a number of frameworks and libraries and have a test suite for +each. The tests run against different versions of the framework/library to make +sure we support everything we claim to. + +This `populate_tox.py` script is responsible for picking reasonable versions to +test automatically and generating parts of `tox.ini` to capture this. + +## How it works + +There is a template in this directory called `tox.jinja` which contains a +combination of hardcoded and generated entries. + +The `populate_tox.py` script fills out the auto-generated part of that template. +It does this by querying PyPI for each framework's package and its metadata and +then determining which versions make sense to test to get good coverage. + +The lowest supported and latest version of a framework are always tested, with +a number of releases in between: +- If the package has majors, we pick the highest version of each major. For the + latest major, we also pick the lowest version in that major. +- If the package doesn't have multiple majors, we pick two versions in between + lowest and highest. + +#### Caveats + +- Make sure the integration name is the same everywhere. If it consists of + multiple words, use an underscore instead of a hyphen. + +## Defining constraints + +The `TEST_SUITE_CONFIG` dictionary defines, for each integration test suite, +the main package (framework, library) to test with; any additional test +dependencies, optionally gated behind specific conditions; and optionally +the Python versions to test on. + +Constraints are defined using the format specified below. The following sections describe each key. + +``` +integration_name: { + "package": name_of_main_package_on_pypi, + "deps": { + rule1: [package1, package2, ...], + rule2: [package3, package4, ...], + }, + "python": python_version_specifier, +} +``` + +### `package` + +The name of the third party package as it's listed on PyPI. The script will +be picking different versions of this package to test. + +This key is mandatory. + +### `deps` + +The test dependencies of the test suite. They're defined as a dictionary of +`rule: [package1, package2, ...]` key-value pairs. All packages +in the package list of a rule will be installed as long as the rule applies. + +`rule`s are predefined. Each `rule` must be one of the following: + - `*`: packages will be always installed + - a version specifier on the main package (e.g. `<=0.32`): packages will only + be installed if the main package falls into the version bounds specified + - specific Python version(s) in the form `py3.8,py3.9`: packages will only be + installed if the Python version matches one from the list + +Rules can be used to specify version bounds on older versions of the main +package's dependencies, for example. If e.g. Flask tests generally need +Werkzeug and don't care about its version, but Flask older than 3.0 needs +a specific Werkzeug version to work, you can say: + +```python +"flask": { + "deps": { + "*": ["Werkzeug"], + "<3.0": ["Werkzeug<2.1.0"], + }, + ... +} +``` + +If you need to install a specific version of a secondary dependency on specific +Python versions, you can say: + +```python +"celery": { + "deps": { + "*": ["newrelic", "redis"], + "py3.7": ["importlib-metadata<5.0"], + }, + ... +} +``` +This key is optional. + +### `python` + +Sometimes, the whole test suite should only run on specific Python versions. +This can be achieved via the `python` key, which expects a version specifier. + +For example, if you want AIOHTTP tests to only run on Python 3.7+, you can say: + +```python +"aiohttp": { + "python": ">=3.7", + ... +} +``` + +The `python` key is optional, and when possible, it should be omitted. The script +should automatically detect which Python versions the package supports. +However, if a package has broken +metadata or the SDK is explicitly not supporting some packages on specific +Python versions (because of, for example, broken context vars), the `python` +key can be used. + + +## How-Tos + +### Add a new test suite + +1. Add the minimum supported version of the framework/library to `_MIN_VERSIONS` + in `integrations/__init__.py`. This should be the lowest version of the + framework that we can guarantee works with the SDK. If you've just added the + integration, you should generally set this to the latest version of the framework + at the time. +2. Add the integration and any constraints to `TEST_SUITE_CONFIG`. See the + "Defining constraints" section for the format. +3. Add the integration to one of the groups in the `GROUPS` dictionary in + `scripts/split_tox_gh_actions/split_tox_gh_actions.py`. +4. Add the `TESTPATH` for the test suite in `tox.jinja`'s `setenv` section. +5. Run `scripts/generate-test-files.sh` and commit the changes. + +### Migrate a test suite to populate_tox.py + +A handful of integration test suites are still hardcoded. The goal is to migrate +them all to `populate_tox.py` over time. + +1. Remove the integration from the `IGNORE` list in `populate_tox.py`. +2. Remove the hardcoded entries for the integration from the `envlist` and `deps` sections of `tox.jinja`. +3. Run `scripts/generate-test-files.sh`. +4. Run the test suite, either locally or by creating a PR. +5. Address any test failures that happen. + +You might have to introduce additional version bounds on the dependencies of the +package. Try to determine the source of the failure and address it. + +Common scenarios: +- An old version of the tested package installs a dependency without defining + an upper version bound on it. A new version of the dependency is installed that + is incompatible with the package. In this case you need to determine which + versions of the dependency don't contain the breaking change and restrict this + in `TEST_SUITE_CONFIG`. +- Tests are failing on an old Python version. In this case first double-check + whether we were even testing them on that version in the original `tox.ini`. diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py new file mode 100644 index 0000000000..9e1366c25b --- /dev/null +++ b/scripts/populate_tox/config.py @@ -0,0 +1,8 @@ +# The TEST_SUITE_CONFIG dictionary defines, for each integration test suite, +# the main package (framework, library) to test with; any additional test +# dependencies, optionally gated behind specific conditions; and optionally +# the Python versions to test on. +# +# See scripts/populate_tox/README.md for more info on the format and examples. + +TEST_SUITE_CONFIG = {} diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py new file mode 100644 index 0000000000..83db87bd35 --- /dev/null +++ b/scripts/populate_tox/populate_tox.py @@ -0,0 +1,548 @@ +""" +This script populates tox.ini automatically using release data from PYPI. +""" + +import functools +import os +import sys +import time +from bisect import bisect_left +from collections import defaultdict +from datetime import datetime, timedelta +from importlib.metadata import metadata +from packaging.specifiers import SpecifierSet +from packaging.version import Version +from pathlib import Path +from typing import Optional, Union + +# Adding the scripts directory to PATH. This is necessary in order to be able +# to import stuff from the split_tox_gh_actions script +sys.path.append(os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))) + +import requests +from jinja2 import Environment, FileSystemLoader +from sentry_sdk.integrations import _MIN_VERSIONS + +from config import TEST_SUITE_CONFIG +from split_tox_gh_actions.split_tox_gh_actions import GROUPS + + +# Only consider package versions going back this far +CUTOFF = datetime.now() - timedelta(days=365 * 5) + +TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" +ENV = Environment( + loader=FileSystemLoader(Path(__file__).resolve().parent), + trim_blocks=True, + lstrip_blocks=True, +) + +PYPI_PROJECT_URL = "https://pypi.python.org/pypi/%7Bproject%7D/json" +PYPI_VERSION_URL = "https://pypi.python.org/pypi/%7Bproject%7D/%7Bversion%7D/json" +CLASSIFIER_PREFIX = "Programming Language :: Python :: " + + +IGNORE = { + # Do not try auto-generating the tox entries for these. They will be + # hardcoded in tox.ini. + # + # This set should be getting smaller over time as we migrate more test + # suites over to this script. Some entries will probably stay forever + # as they don't fit the mold (e.g. common, asgi, which don't have a 3rd party + # pypi package to install in different versions). + "common", + "gevent", + "opentelemetry", + "potel", + "aiohttp", + "anthropic", + "ariadne", + "arq", + "asgi", + "asyncpg", + "aws_lambda", + "beam", + "boto3", + "bottle", + "celery", + "chalice", + "clickhouse_driver", + "cohere", + "cloud_resource_context", + "cohere", + "django", + "dramatiq", + "falcon", + "fastapi", + "flask", + "gcp", + "gql", + "graphene", + "grpc", + "httpx", + "huey", + "huggingface_hub", + "langchain", + "langchain_notiktoken", + "launchdarkly", + "litestar", + "loguru", + "openai", + "openai_notiktoken", + "openfeature", + "pure_eval", + "pymongo", + "pyramid", + "quart", + "ray", + "redis", + "redis_py_cluster_legacy", + "requests", + "rq", + "sanic", + "spark", + "starlette", + "starlite", + "sqlalchemy", + "strawberry", + "tornado", + "trytond", + "typer", + "unleash", +} + + +@functools.cache +def fetch_package(package: str) -> dict: + """Fetch package metadata from PyPI.""" + url = PYPI_PROJECT_URL.format(project=package) + pypi_data = requests.get(url) + + if pypi_data.status_code != 200: + print(f"{package} not found") + + return pypi_data.json() + + +@functools.cache +def fetch_release(package: str, version: Version) -> dict: + url = PYPI_VERSION_URL.format(project=package, version=version) + pypi_data = requests.get(url) + + if pypi_data.status_code != 200: + print(f"{package} not found") + + return pypi_data.json() + + +def _prefilter_releases(integration: str, releases: dict[str, dict]) -> list[Version]: + """ + Filter `releases`, removing releases that are for sure unsupported. + + This function doesn't guarantee that all releases it returns are supported -- + there are further criteria that will be checked later in the pipeline because + they require additional API calls to be made. The purpose of this function is + to slim down the list so that we don't have to make more API calls than + necessary for releases that are for sure not supported. + """ + min_supported = _MIN_VERSIONS.get(integration) + if min_supported is not None: + min_supported = Version(".".join(map(str, min_supported))) + else: + print( + f" {integration} doesn't have a minimum version defined in sentry_sdk/integrations/__init__.py. Consider defining one" + ) + + filtered_releases = [] + + for release, data in releases.items(): + if not data: + continue + + meta = data[0] + if datetime.fromisoformat(meta["upload_time"]) < CUTOFF: + continue + + if meta["yanked"]: + continue + + version = Version(release) + + if min_supported and version < min_supported: + continue + + if version.is_prerelease or version.is_postrelease: + # TODO: consider the newest prerelease unless obsolete + # https://github.com/getsentry/sentry-python/issues/4030 + continue + + for i, saved_version in enumerate(filtered_releases): + if ( + version.major == saved_version.major + and version.minor == saved_version.minor + and version.micro > saved_version.micro + ): + # Don't save all patch versions of a release, just the newest one + filtered_releases[i] = version + break + else: + filtered_releases.append(version) + + return sorted(filtered_releases) + + +def get_supported_releases(integration: str, pypi_data: dict) -> list[Version]: + """ + Get a list of releases that are currently supported by the SDK. + + This takes into account a handful of parameters (Python support, the lowest + version we've defined for the framework, the date of the release). + """ + package = pypi_data["info"]["name"] + + # Get a consolidated list without taking into account Python support yet + # (because that might require an additional API call for some + # of the releases) + releases = _prefilter_releases(integration, pypi_data["releases"]) + + # Determine Python support + expected_python_versions = TEST_SUITE_CONFIG[integration].get("python") + if expected_python_versions: + expected_python_versions = SpecifierSet(expected_python_versions) + else: + expected_python_versions = SpecifierSet(f">={MIN_PYTHON_VERSION}") + + def _supports_lowest(release: Version) -> bool: + time.sleep(0.1) # don't DoS PYPI + py_versions = determine_python_versions(fetch_release(package, release)) + target_python_versions = TEST_SUITE_CONFIG[integration].get("python") + if target_python_versions: + target_python_versions = SpecifierSet(target_python_versions) + return bool(supported_python_versions(py_versions, target_python_versions)) + + if not _supports_lowest(releases[0]): + i = bisect_left(releases, True, key=_supports_lowest) + if i != len(releases) and _supports_lowest(releases[i]): + # we found the lowest version that supports at least some Python + # version(s) that we do, cut off the rest + releases = releases[i:] + + return releases + + +def pick_releases_to_test(releases: list[Version]) -> list[Version]: + """Pick a handful of releases to test from a sorted list of supported releases.""" + # If the package has majors (or major-like releases, even if they don't do + # semver), we want to make sure we're testing them all. If not, we just pick + # the oldest, the newest, and a couple in between. + has_majors = len(set([v.major for v in releases])) > 1 + filtered_releases = set() + + if has_majors: + # Always check the very first supported release + filtered_releases.add(releases[0]) + + # Find out the min and max release by each major + releases_by_major = {} + for release in releases: + if release.major not in releases_by_major: + releases_by_major[release.major] = [release, release] + if release < releases_by_major[release.major][0]: + releases_by_major[release.major][0] = release + if release > releases_by_major[release.major][1]: + releases_by_major[release.major][1] = release + + for i, (min_version, max_version) in enumerate(releases_by_major.values()): + filtered_releases.add(max_version) + if i == len(releases_by_major) - 1: + # If this is the latest major release, also check the lowest + # version of this version + filtered_releases.add(min_version) + + else: + filtered_releases = { + releases[0], # oldest version supported + releases[len(releases) // 3], + releases[ + len(releases) // 3 * 2 + ], # two releases in between, roughly evenly spaced + releases[-1], # latest + } + + return sorted(filtered_releases) + + +def supported_python_versions( + package_python_versions: Union[SpecifierSet, list[Version]], + custom_supported_versions: Optional[SpecifierSet] = None, +) -> list[Version]: + """ + Get the intersection of Python versions supported by the package and the SDK. + + Optionally, if `custom_supported_versions` is provided, the function will + return the intersection of Python versions supported by the package, the SDK, + and `custom_supported_versions`. This is used when a test suite definition + in `TEST_SUITE_CONFIG` contains a range of Python versions to run the tests + on. + + Examples: + - The Python SDK supports Python 3.6-3.13. The package supports 3.5-3.8. This + function will return [3.6, 3.7, 3.8] as the Python versions supported + by both. + - The Python SDK supports Python 3.6-3.13. The package supports 3.5-3.8. We + have an additional test limitation in place to only test this framework + on Python 3.7, so we can provide this as `custom_supported_versions`. The + result of this function will then by the intersection of all three, i.e., + [3.7]. + """ + supported = [] + + # Iterate through Python versions from MIN_PYTHON_VERSION to MAX_PYTHON_VERSION + curr = MIN_PYTHON_VERSION + while curr <= MAX_PYTHON_VERSION: + if curr in package_python_versions: + if not custom_supported_versions or curr in custom_supported_versions: + supported.append(curr) + + # Construct the next Python version (i.e., bump the minor) + next = [int(v) for v in str(curr).split(".")] + next[1] += 1 + curr = Version(".".join(map(str, next))) + + return supported + + +def pick_python_versions_to_test(python_versions: list[Version]) -> list[Version]: + """ + Given a list of Python versions, pick those that make sense to test on. + + Currently, this is the oldest, the newest, and the second newest Python + version. + """ + filtered_python_versions = { + python_versions[0], + } + + filtered_python_versions.add(python_versions[-1]) + try: + filtered_python_versions.add(python_versions[-2]) + except IndexError: + pass + + return sorted(filtered_python_versions) + + +def _parse_python_versions_from_classifiers(classifiers: list[str]) -> list[Version]: + python_versions = [] + for classifier in classifiers: + if classifier.startswith(CLASSIFIER_PREFIX): + python_version = classifier[len(CLASSIFIER_PREFIX) :] + if "." in python_version: + # We don't care about stuff like + # Programming Language :: Python :: 3 :: Only, + # Programming Language :: Python :: 3, + # etc., we're only interested in specific versions, like 3.13 + python_versions.append(Version(python_version)) + + if python_versions: + python_versions.sort() + return python_versions + + +def determine_python_versions(pypi_data: dict) -> Union[SpecifierSet, list[Version]]: + """ + Given data from PyPI's release endpoint, determine the Python versions supported by the package + from the Python version classifiers, when present, or from `requires_python` if there are no classifiers. + """ + try: + classifiers = pypi_data["info"]["classifiers"] + except (AttributeError, KeyError): + # This function assumes `pypi_data` contains classifiers. This is the case + # for the most recent release in the /{project} endpoint or for any release + # fetched via the /{project}/{version} endpoint. + return [] + + # Try parsing classifiers + python_versions = _parse_python_versions_from_classifiers(classifiers) + if python_versions: + return python_versions + + # We only use `requires_python` if there are no classifiers. This is because + # `requires_python` doesn't tell us anything about the upper bound, which + # depends on when the release first came out + try: + requires_python = pypi_data["info"]["requires_python"] + except (AttributeError, KeyError): + pass + + if requires_python: + return SpecifierSet(requires_python) + + return [] + + +def _render_python_versions(python_versions: list[Version]) -> str: + return ( + "{" + + ",".join(f"py{version.major}.{version.minor}" for version in python_versions) + + "}" + ) + + +def _render_dependencies(integration: str, releases: list[Version]) -> list[str]: + rendered = [] + + if TEST_SUITE_CONFIG[integration].get("deps") is None: + return rendered + + for constraint, deps in TEST_SUITE_CONFIG[integration]["deps"].items(): + if constraint == "*": + for dep in deps: + rendered.append(f"{integration}: {dep}") + elif constraint.startswith("py3"): + for dep in deps: + rendered.append(f"{constraint}-{integration}: {dep}") + else: + restriction = SpecifierSet(constraint) + for release in releases: + if release in restriction: + for dep in deps: + rendered.append(f"{integration}-v{release}: {dep}") + + return rendered + + +def write_tox_file(packages: dict) -> None: + template = ENV.get_template("tox.jinja") + + context = {"groups": {}} + for group, integrations in packages.items(): + context["groups"][group] = [] + for integration in integrations: + context["groups"][group].append( + { + "name": integration["name"], + "package": integration["package"], + "extra": integration["extra"], + "releases": integration["releases"], + "dependencies": _render_dependencies( + integration["name"], integration["releases"] + ), + } + ) + + rendered = template.render(context) + + with open(TOX_FILE, "w") as file: + file.write(rendered) + file.write("\n") + + +def _get_package_name(integration: str) -> tuple[str, Optional[str]]: + package = TEST_SUITE_CONFIG[integration]["package"] + extra = None + if "[" in package: + extra = package[package.find("[") + 1 : package.find("]")] + package = package[: package.find("[")] + + return package, extra + + +def _compare_min_version_with_defined( + integration: str, releases: list[Version] +) -> None: + defined_min_version = _MIN_VERSIONS.get(integration) + if defined_min_version: + defined_min_version = Version(".".join([str(v) for v in defined_min_version])) + if ( + defined_min_version.major != releases[0].major + or defined_min_version.minor != releases[0].minor + ): + print( + f" Integration defines {defined_min_version} as minimum " + f"version, but the effective minimum version is {releases[0]}." + ) + + +def _add_python_versions_to_release(integration: str, package: str, release: Version): + release_pypi_data = fetch_release(package, release) + time.sleep(0.1) # give PYPI some breathing room + + target_python_versions = TEST_SUITE_CONFIG[integration].get("python") + if target_python_versions: + target_python_versions = SpecifierSet(target_python_versions) + + release.python_versions = pick_python_versions_to_test( + supported_python_versions( + determine_python_versions(release_pypi_data), + target_python_versions, + ) + ) + + release.rendered_python_versions = _render_python_versions(release.python_versions) + + +def main() -> None: + global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION + sdk_python_versions = _parse_python_versions_from_classifiers( + metadata("sentry-sdk").get_all("Classifier") + ) + MIN_PYTHON_VERSION = sdk_python_versions[0] + MAX_PYTHON_VERSION = sdk_python_versions[-1] + print( + f"The SDK supports Python versions {MIN_PYTHON_VERSION} - {MAX_PYTHON_VERSION}." + ) + + packages = defaultdict(list) + + for group, integrations in GROUPS.items(): + for integration in integrations: + if integration in IGNORE: + continue + + print(f"Processing {integration}...") + + # Figure out the actual main package + package, extra = _get_package_name(integration) + + # Fetch data for the main package + pypi_data = fetch_package(package) + + # Get the list of all supported releases + releases = get_supported_releases(integration, pypi_data) + if not releases: + print(" Found no supported releases.") + continue + + _compare_min_version_with_defined(integration, releases) + + # Pick a handful of the supported releases to actually test against + # and fetch the PYPI data for each to determine which Python versions + # to test it on + test_releases = pick_releases_to_test(releases) + + for release in test_releases: + py_versions = _add_python_versions_to_release( + integration, package, release + ) + if not py_versions: + print(f" Release {release} has no Python versions, skipping.") + + test_releases = [ + release for release in test_releases if release.python_versions + ] + if test_releases: + packages[group].append( + { + "name": integration, + "package": package, + "extra": extra, + "releases": test_releases, + } + ) + + write_tox_file(packages) + + +if __name__ == "__main__": + main() diff --git a/scripts/populate_tox/requirements.txt b/scripts/populate_tox/requirements.txt new file mode 100644 index 0000000000..0402fac5ab --- /dev/null +++ b/scripts/populate_tox/requirements.txt @@ -0,0 +1,3 @@ +jinja2 +packaging +requests diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja new file mode 100644 index 0000000000..b60c6f137a --- /dev/null +++ b/scripts/populate_tox/tox.jinja @@ -0,0 +1,899 @@ +# Tox (http://codespeak.net/~hpk/tox/) is a tool for running tests +# in multiple virtualenvs. This configuration file will run the +# test suite on all supported python versions. To use it, "pip install tox" +# and then run "tox" from this directory. +# +# This file has been generated from a template +# by "scripts/populate_tox/populate_tox.py". Any changes to the file should +# be made in the template (if you want to change a hardcoded part of the file) +# or in the script (if you want to change the auto-generated part). +# The file (and all resulting CI YAMLs) then need to be regenerated via +# "scripts/generate-test-files.sh". + +[tox] +requires = + # This version introduced using pip 24.1 which does not work with older Celery and HTTPX versions. + virtualenv<20.26.3 +envlist = + # === Common === + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common + + # === Gevent === + {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent + + # === Integrations === + # General format is {pythonversion}-{integrationname}-v{frameworkversion} + # 1 blank line between different integrations + # Each framework version should only be mentioned once. I.e: + # {py3.7,py3.10}-django-v{3.2} + # {py3.10}-django-v{4.0} + # instead of: + # {py3.7}-django-v{3.2} + # {py3.7,py3.10}-django-v{3.2,4.0} + # + # At a minimum, we should test against at least the lowest + # and the latest supported version of a framework. + + # AIOHTTP + {py3.7}-aiohttp-v{3.4} + {py3.7,py3.9,py3.11}-aiohttp-v{3.8} + {py3.8,py3.12,py3.13}-aiohttp-latest + + # Anthropic + {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} + {py3.7,py3.11,py3.12}-anthropic-latest + + # Ariadne + {py3.8,py3.11}-ariadne-v{0.20} + {py3.8,py3.12,py3.13}-ariadne-latest + + # Arq + {py3.7,py3.11}-arq-v{0.23} + {py3.7,py3.12,py3.13}-arq-latest + + # Asgi + {py3.7,py3.12,py3.13}-asgi + + # asyncpg + {py3.7,py3.10}-asyncpg-v{0.23} + {py3.8,py3.11,py3.12}-asyncpg-latest + + # AWS Lambda + # The aws_lambda tests deploy to the real AWS and have their own + # matrix of Python versions to run the test lambda function in. + # see `lambda_runtime` fixture in tests/integrations/aws_lambda.py + {py3.9}-aws_lambda + + # Beam + {py3.7}-beam-v{2.12} + {py3.8,py3.11}-beam-latest + + # Boto3 + {py3.6,py3.7}-boto3-v{1.12} + {py3.7,py3.11,py3.12}-boto3-v{1.23} + {py3.11,py3.12}-boto3-v{1.34} + {py3.11,py3.12,py3.13}-boto3-latest + + # Bottle + {py3.6,py3.9}-bottle-v{0.12} + {py3.6,py3.12,py3.13}-bottle-latest + + # Celery + {py3.6,py3.8}-celery-v{4} + {py3.6,py3.8}-celery-v{5.0} + {py3.7,py3.10}-celery-v{5.1,5.2} + {py3.8,py3.11,py3.12}-celery-v{5.3,5.4,5.5} + {py3.8,py3.12,py3.13}-celery-latest + + # Chalice + {py3.6,py3.9}-chalice-v{1.16} + {py3.8,py3.12,py3.13}-chalice-latest + + # Clickhouse Driver + {py3.8,py3.11}-clickhouse_driver-v{0.2.0} + {py3.8,py3.12,py3.13}-clickhouse_driver-latest + + # Cloud Resource Context + {py3.6,py3.12,py3.13}-cloud_resource_context + + # Cohere + {py3.9,py3.11,py3.12}-cohere-v5 + {py3.9,py3.11,py3.12}-cohere-latest + + # Django + # - Django 1.x + {py3.6,py3.7}-django-v{1.11} + # - Django 2.x + {py3.6,py3.7}-django-v{2.0} + {py3.6,py3.9}-django-v{2.2} + # - Django 3.x + {py3.6,py3.9}-django-v{3.0} + {py3.6,py3.9,py3.11}-django-v{3.2} + # - Django 4.x + {py3.8,py3.11,py3.12}-django-v{4.0,4.1,4.2} + # - Django 5.x + {py3.10,py3.11,py3.12}-django-v{5.0,5.1} + {py3.10,py3.12,py3.13}-django-latest + + # dramatiq + {py3.6,py3.9}-dramatiq-v{1.13} + {py3.7,py3.10,py3.11}-dramatiq-v{1.15} + {py3.8,py3.11,py3.12}-dramatiq-v{1.17} + {py3.8,py3.11,py3.12}-dramatiq-latest + + # Falcon + {py3.6,py3.7}-falcon-v{1,1.4,2} + {py3.6,py3.11,py3.12}-falcon-v{3} + {py3.8,py3.11,py3.12}-falcon-v{4} + {py3.7,py3.11,py3.12}-falcon-latest + + # FastAPI + {py3.7,py3.10}-fastapi-v{0.79} + {py3.8,py3.12,py3.13}-fastapi-latest + + # Flask + {py3.6,py3.8}-flask-v{1} + {py3.8,py3.11,py3.12}-flask-v{2} + {py3.10,py3.11,py3.12}-flask-v{3} + {py3.10,py3.12,py3.13}-flask-latest + + # GCP + {py3.7}-gcp + + # GQL + {py3.7,py3.11}-gql-v{3.4} + {py3.7,py3.12,py3.13}-gql-latest + + # Graphene + {py3.7,py3.11}-graphene-v{3.3} + {py3.7,py3.12,py3.13}-graphene-latest + + # gRPC + {py3.7,py3.9}-grpc-v{1.39} + {py3.7,py3.10}-grpc-v{1.49} + {py3.7,py3.11}-grpc-v{1.59} + {py3.8,py3.11,py3.12}-grpc-latest + + # HTTPX + {py3.6,py3.9}-httpx-v{0.16,0.18} + {py3.6,py3.10}-httpx-v{0.20,0.22} + {py3.7,py3.11,py3.12}-httpx-v{0.23,0.24} + {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} + {py3.9,py3.12,py3.13}-httpx-latest + + # Huey + {py3.6,py3.11,py3.12}-huey-v{2.0} + {py3.6,py3.12,py3.13}-huey-latest + + # Huggingface Hub + {py3.9,py3.12,py3.13}-huggingface_hub-{v0.22} + {py3.9,py3.12,py3.13}-huggingface_hub-latest + + # Langchain + {py3.9,py3.11,py3.12}-langchain-v0.1 + {py3.9,py3.11,py3.12}-langchain-v0.3 + {py3.9,py3.11,py3.12}-langchain-latest + {py3.9,py3.11,py3.12}-langchain-notiktoken + + # LaunchDarkly + {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 + {py3.8,py3.12,py3.13}-launchdarkly-latest + + # Litestar + {py3.8,py3.11}-litestar-v{2.0} + {py3.8,py3.11,py3.12}-litestar-v{2.6} + {py3.8,py3.11,py3.12}-litestar-v{2.12} + {py3.8,py3.11,py3.12}-litestar-latest + + # Loguru + {py3.6,py3.11,py3.12}-loguru-v{0.5} + {py3.6,py3.12,py3.13}-loguru-latest + + # OpenAI + {py3.9,py3.11,py3.12}-openai-v1.0 + {py3.9,py3.11,py3.12}-openai-v1.22 + {py3.9,py3.11,py3.12}-openai-v1.55 + {py3.9,py3.11,py3.12}-openai-latest + {py3.9,py3.11,py3.12}-openai-notiktoken + + # OpenFeature + {py3.8,py3.12,py3.13}-openfeature-v0.7 + {py3.8,py3.12,py3.13}-openfeature-latest + + # OpenTelemetry (OTel) + {py3.7,py3.9,py3.12,py3.13}-opentelemetry + + # OpenTelemetry Experimental (POTel) + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel + + # pure_eval + {py3.6,py3.12,py3.13}-pure_eval + + # PyMongo (Mongo DB) + {py3.6}-pymongo-v{3.1} + {py3.6,py3.9}-pymongo-v{3.12} + {py3.6,py3.11}-pymongo-v{4.0} + {py3.7,py3.11,py3.12}-pymongo-v{4.3,4.7} + {py3.7,py3.12,py3.13}-pymongo-latest + + # Pyramid + {py3.6,py3.11}-pyramid-v{1.6} + {py3.6,py3.11,py3.12}-pyramid-v{1.10} + {py3.6,py3.11,py3.12}-pyramid-v{2.0} + {py3.6,py3.11,py3.12}-pyramid-latest + + # Quart + {py3.7,py3.11}-quart-v{0.16} + {py3.8,py3.11,py3.12}-quart-v{0.19} + {py3.8,py3.12,py3.13}-quart-latest + + # Ray + {py3.10,py3.11}-ray-v{2.34} + {py3.10,py3.11}-ray-latest + + # Redis + {py3.6,py3.8}-redis-v{3} + {py3.7,py3.8,py3.11}-redis-v{4} + {py3.7,py3.11,py3.12}-redis-v{5} + {py3.7,py3.12,py3.13}-redis-latest + + # Redis Cluster + {py3.6,py3.8}-redis_py_cluster_legacy-v{1,2} + # no -latest, not developed anymore + + # Requests + {py3.6,py3.8,py3.12,py3.13}-requests + + # RQ (Redis Queue) + {py3.6}-rq-v{0.6} + {py3.6,py3.9}-rq-v{0.13,1.0} + {py3.6,py3.11}-rq-v{1.5,1.10} + {py3.7,py3.11,py3.12}-rq-v{1.15,1.16} + {py3.7,py3.12,py3.13}-rq-latest + + # Sanic + {py3.6,py3.7}-sanic-v{0.8} + {py3.6,py3.8}-sanic-v{20} + {py3.8,py3.11,py3.12}-sanic-v{24.6} + {py3.9,py3.12,py3.13}-sanic-latest + + # Spark + {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} + {py3.8,py3.10,py3.11,py3.12}-spark-latest + + # Starlette + {py3.7,py3.10}-starlette-v{0.19} + {py3.7,py3.11}-starlette-v{0.24,0.28} + {py3.8,py3.11,py3.12}-starlette-v{0.32,0.36,0.40} + {py3.8,py3.12,py3.13}-starlette-latest + + # Starlite + {py3.8,py3.11}-starlite-v{1.48,1.51} + # 1.51.14 is the last starlite version; the project continues as litestar + + # SQL Alchemy + {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} + {py3.7,py3.11}-sqlalchemy-v{2.0} + {py3.7,py3.12,py3.13}-sqlalchemy-latest + + # Strawberry + {py3.8,py3.11}-strawberry-v{0.209} + {py3.8,py3.11,py3.12}-strawberry-v{0.222} + {py3.8,py3.12,py3.13}-strawberry-latest + + # Tornado + {py3.8,py3.11,py3.12}-tornado-v{6.0} + {py3.8,py3.11,py3.12}-tornado-v{6.2} + {py3.8,py3.11,py3.12}-tornado-latest + + # Trytond + {py3.6}-trytond-v{4} + {py3.6,py3.8}-trytond-v{5} + {py3.6,py3.11}-trytond-v{6} + {py3.8,py3.11,py3.12}-trytond-v{7} + {py3.8,py3.12,py3.13}-trytond-latest + + # Typer + {py3.7,py3.12,py3.13}-typer-v{0.15} + {py3.7,py3.12,py3.13}-typer-latest + + # Unleash + {py3.8,py3.12,py3.13}-unleash-v6.0.1 + {py3.8,py3.12,py3.13}-unleash-latest + + # === Integrations - Auto-generated === + # These come from the populate_tox.py script. Eventually we should move all + # integration tests there. + + {% for group, integrations in groups.items() %} + # ~~~ {{ group }} ~~~ + {% for integration in integrations %} + {% for release in integration.releases %} + {{ release.rendered_python_versions }}-{{ integration.name }}-v{{ release }} + {% endfor %} + + {% endfor %} + + {% endfor %} + +[testenv] +deps = + # if you change requirements-testing.txt and your change is not being reflected + # in what's installed by tox (when running tox locally), try running tox + # with the -r flag + -r requirements-testing.txt + + linters: -r requirements-linting.txt + linters: werkzeug<2.3.0 + + # === Common === + py3.8-common: hypothesis + common: pytest-asyncio + # See https://github.com/pytest-dev/pytest/issues/9621 + # and https://github.com/pytest-dev/pytest-forked/issues/67 + # for justification of the upper bound on pytest + {py3.6,py3.7}-common: pytest<7.0.0 + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common: pytest + + # === Gevent === + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 + {py3.12}-gevent: gevent + # See https://github.com/pytest-dev/pytest/issues/9621 + # and https://github.com/pytest-dev/pytest-forked/issues/67 + # for justification of the upper bound on pytest + {py3.6,py3.7}-gevent: pytest<7.0.0 + {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest + + # === Integrations === + + # AIOHTTP + aiohttp-v3.4: aiohttp~=3.4.0 + aiohttp-v3.8: aiohttp~=3.8.0 + aiohttp-latest: aiohttp + aiohttp: pytest-aiohttp + aiohttp-v3.8: pytest-asyncio + aiohttp-latest: pytest-asyncio + + # Anthropic + anthropic: pytest-asyncio + anthropic-v{0.16,0.28}: httpx<0.28.0 + anthropic-v0.16: anthropic~=0.16.0 + anthropic-v0.28: anthropic~=0.28.0 + anthropic-v0.40: anthropic~=0.40.0 + anthropic-latest: anthropic + + # Ariadne + ariadne-v0.20: ariadne~=0.20.0 + ariadne-latest: ariadne + ariadne: fastapi + ariadne: flask + ariadne: httpx + + # Arq + arq-v0.23: arq~=0.23.0 + arq-v0.23: pydantic<2 + arq-latest: arq + arq: fakeredis>=2.2.0,<2.8 + arq: pytest-asyncio + arq: async-timeout + + # Asgi + asgi: pytest-asyncio + asgi: async-asgi-testclient + + # Asyncpg + asyncpg-v0.23: asyncpg~=0.23.0 + asyncpg-latest: asyncpg + asyncpg: pytest-asyncio + + # AWS Lambda + aws_lambda: boto3 + + # Beam + beam-v2.12: apache-beam~=2.12.0 + beam-latest: apache-beam + + # Boto3 + boto3-v1.12: boto3~=1.12.0 + boto3-v1.23: boto3~=1.23.0 + boto3-v1.34: boto3~=1.34.0 + boto3-latest: boto3 + + # Bottle + bottle: Werkzeug<2.1.0 + bottle-v0.12: bottle~=0.12.0 + bottle-latest: bottle + + # Celery + celery: redis + celery-v4: Celery~=4.0 + celery-v5.0: Celery~=5.0.0 + celery-v5.1: Celery~=5.1.0 + celery-v5.2: Celery~=5.2.0 + celery-v5.3: Celery~=5.3.0 + celery-v5.4: Celery~=5.4.0 + # TODO: update when stable is out + celery-v5.5: Celery==5.5.0rc4 + celery-latest: Celery + + celery: newrelic + {py3.7}-celery: importlib-metadata<5.0 + + # Chalice + chalice: pytest-chalice==0.0.5 + chalice-v1.16: chalice~=1.16.0 + chalice-latest: chalice + + # Clickhouse Driver + clickhouse_driver-v0.2.0: clickhouse_driver~=0.2.0 + clickhouse_driver-latest: clickhouse_driver + + # Cohere + cohere-v5: cohere~=5.3.3 + cohere-latest: cohere + + # Django + django: psycopg2-binary + django-v{1.11,2.0,2.1,2.2,3.0,3.1,3.2}: djangorestframework>=3.0.0,<4.0.0 + django-v{2.0,2.2,3.0,3.2,4.0,4.1,4.2,5.0,5.1}: channels[daphne] + django-v{2.2,3.0}: six + django-v{1.11,2.0,2.2,3.0,3.2}: Werkzeug<2.1.0 + django-v{1.11,2.0,2.2,3.0}: pytest-django<4.0 + django-v{3.2,4.0,4.1,4.2,5.0,5.1}: pytest-django + django-v{4.0,4.1,4.2,5.0,5.1}: djangorestframework + django-v{4.0,4.1,4.2,5.0,5.1}: pytest-asyncio + django-v{4.0,4.1,4.2,5.0,5.1}: Werkzeug + django-latest: djangorestframework + django-latest: pytest-asyncio + django-latest: pytest-django + django-latest: Werkzeug + django-latest: channels[daphne] + + django-v1.11: Django~=1.11.0 + django-v2.0: Django~=2.0.0 + django-v2.2: Django~=2.2.0 + django-v3.0: Django~=3.0.0 + django-v3.2: Django~=3.2.0 + django-v4.0: Django~=4.0.0 + django-v4.1: Django~=4.1.0 + django-v4.2: Django~=4.2.0 + django-v5.0: Django~=5.0.0 + django-v5.1: Django==5.1rc1 + django-latest: Django + + # dramatiq + dramatiq-v1.13: dramatiq>=1.13,<1.14 + dramatiq-v1.15: dramatiq>=1.15,<1.16 + dramatiq-v1.17: dramatiq>=1.17,<1.18 + dramatiq-latest: dramatiq + + # Falcon + falcon-v1.4: falcon~=1.4.0 + falcon-v1: falcon~=1.0 + falcon-v2: falcon~=2.0 + falcon-v3: falcon~=3.0 + falcon-v4: falcon~=4.0 + falcon-latest: falcon + + # FastAPI + fastapi: httpx + # (this is a dependency of httpx) + fastapi: anyio<4.0.0 + fastapi: pytest-asyncio + fastapi: python-multipart + fastapi: requests + fastapi-v{0.79}: fastapi~=0.79.0 + fastapi-latest: fastapi + + # Flask + flask: flask-login + flask-v{1,2.0}: Werkzeug<2.1.0 + flask-v{1,2.0}: markupsafe<2.1.0 + flask-v{3}: Werkzeug + flask-v1: Flask~=1.0 + flask-v2: Flask~=2.0 + flask-v3: Flask~=3.0 + flask-latest: Flask + + # GQL + gql-v{3.4}: gql[all]~=3.4.0 + gql-latest: gql[all] + + # Graphene + graphene: blinker + graphene: fastapi + graphene: flask + graphene: httpx + graphene-v{3.3}: graphene~=3.3.0 + graphene-latest: graphene + + # gRPC + grpc: protobuf + grpc: mypy-protobuf + grpc: types-protobuf + grpc: pytest-asyncio + grpc-v1.39: grpcio~=1.39.0 + grpc-v1.49: grpcio~=1.49.1 + grpc-v1.59: grpcio~=1.59.0 + grpc-latest: grpcio + + # HTTPX + httpx-v0.16: pytest-httpx==0.10.0 + httpx-v0.18: pytest-httpx==0.12.0 + httpx-v0.20: pytest-httpx==0.14.0 + httpx-v0.22: pytest-httpx==0.19.0 + httpx-v0.23: pytest-httpx==0.21.0 + httpx-v0.24: pytest-httpx==0.22.0 + httpx-v0.25: pytest-httpx==0.25.0 + httpx: pytest-httpx + # anyio is a dep of httpx + httpx: anyio<4.0.0 + httpx-v0.16: httpx~=0.16.0 + httpx-v0.18: httpx~=0.18.0 + httpx-v0.20: httpx~=0.20.0 + httpx-v0.22: httpx~=0.22.0 + httpx-v0.23: httpx~=0.23.0 + httpx-v0.24: httpx~=0.24.0 + httpx-v0.25: httpx~=0.25.0 + httpx-v0.27: httpx~=0.27.0 + httpx-latest: httpx + + # Huey + huey-v2.0: huey~=2.0.0 + huey-latest: huey + + # Huggingface Hub + huggingface_hub-v0.22: huggingface_hub~=0.22.2 + huggingface_hub-latest: huggingface_hub + + # Langchain + langchain-v0.1: openai~=1.0.0 + langchain-v0.1: langchain~=0.1.11 + langchain-v0.1: tiktoken~=0.6.0 + langchain-v0.1: httpx<0.28.0 + langchain-v0.3: langchain~=0.3.0 + langchain-v0.3: langchain-community + langchain-v0.3: tiktoken + langchain-v0.3: openai + langchain-{latest,notiktoken}: langchain + langchain-{latest,notiktoken}: langchain-openai + langchain-{latest,notiktoken}: openai>=1.6.1 + langchain-latest: tiktoken~=0.6.0 + + # Litestar + litestar: pytest-asyncio + litestar: python-multipart + litestar: requests + litestar: cryptography + litestar-v{2.0,2.6}: httpx<0.28 + litestar-v2.0: litestar~=2.0.0 + litestar-v2.6: litestar~=2.6.0 + litestar-v2.12: litestar~=2.12.0 + litestar-latest: litestar + + # Loguru + loguru-v0.5: loguru~=0.5.0 + loguru-latest: loguru + + # OpenAI + openai: pytest-asyncio + openai-v1.0: openai~=1.0.0 + openai-v1.0: tiktoken + openai-v1.0: httpx<0.28.0 + openai-v1.22: openai~=1.22.0 + openai-v1.22: tiktoken + openai-v1.22: httpx<0.28.0 + openai-v1.55: openai~=1.55.0 + openai-v1.55: tiktoken + openai-latest: openai + openai-latest: tiktoken~=0.6.0 + openai-notiktoken: openai + + # OpenFeature + openfeature-v0.7: openfeature-sdk~=0.7.1 + openfeature-latest: openfeature-sdk + + # LaunchDarkly + launchdarkly-v9.8.0: launchdarkly-server-sdk~=9.8.0 + launchdarkly-latest: launchdarkly-server-sdk + + # Unleash + unleash-v6.0.1: UnleashClient~=6.0.1 + unleash-latest: UnleashClient + + # OpenTelemetry (OTel) + opentelemetry: opentelemetry-distro + + # OpenTelemetry Experimental (POTel) + potel: -e .[opentelemetry-experimental] + + # pure_eval + pure_eval: pure_eval + + # PyMongo (MongoDB) + pymongo: mockupdb + pymongo-v3.1: pymongo~=3.1.0 + pymongo-v3.13: pymongo~=3.13.0 + pymongo-v4.0: pymongo~=4.0.0 + pymongo-v4.3: pymongo~=4.3.0 + pymongo-v4.7: pymongo~=4.7.0 + pymongo-latest: pymongo + + # Pyramid + pyramid: Werkzeug<2.1.0 + pyramid-v1.6: pyramid~=1.6.0 + pyramid-v1.10: pyramid~=1.10.0 + pyramid-v2.0: pyramid~=2.0.0 + pyramid-latest: pyramid + + # Quart + quart: quart-auth + quart: pytest-asyncio + quart-v0.16: blinker<1.6 + quart-v0.16: jinja2<3.1.0 + quart-v0.16: Werkzeug<2.1.0 + quart-v0.16: hypercorn<0.15.0 + quart-v0.16: quart~=0.16.0 + quart-v0.19: Werkzeug>=3.0.0 + quart-v0.19: quart~=0.19.0 + {py3.8}-quart: taskgroup==0.0.0a4 + quart-latest: quart + + # Ray + ray-v2.34: ray~=2.34.0 + ray-latest: ray + + # Redis + redis: fakeredis!=1.7.4 + redis: pytest<8.0.0 + {py3.6,py3.7}-redis: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 + {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio + redis-v3: redis~=3.0 + redis-v4: redis~=4.0 + redis-v5: redis~=5.0 + redis-latest: redis + + # Redis Cluster + redis_py_cluster_legacy-v1: redis-py-cluster~=1.0 + redis_py_cluster_legacy-v2: redis-py-cluster~=2.0 + + # Requests + requests: requests>=2.0 + + # RQ (Redis Queue) + # https://github.com/jamesls/fakeredis/issues/245 + rq-v{0.6}: fakeredis<1.0 + rq-v{0.6}: redis<3.2.2 + rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 + rq-v{1.15,1.16}: fakeredis + {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 + rq-latest: fakeredis + {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 + rq-v0.6: rq~=0.6.0 + rq-v0.13: rq~=0.13.0 + rq-v1.0: rq~=1.0.0 + rq-v1.5: rq~=1.5.0 + rq-v1.10: rq~=1.10.0 + rq-v1.15: rq~=1.15.0 + rq-v1.16: rq~=1.16.0 + rq-latest: rq + + # Sanic + sanic: websockets<11.0 + sanic: aiohttp + sanic-v{24.6}: sanic_testing + sanic-latest: sanic_testing + {py3.6}-sanic: aiocontextvars==0.2.1 + sanic-v0.8: sanic~=0.8.0 + sanic-v20: sanic~=20.0 + sanic-v24.6: sanic~=24.6.0 + sanic-latest: sanic + + # Spark + spark-v3.1: pyspark~=3.1.0 + spark-v3.3: pyspark~=3.3.0 + spark-v3.5: pyspark~=3.5.0 + # TODO: update to ~=4.0.0 once stable is out + spark-v4.0: pyspark==4.0.0.dev2 + spark-latest: pyspark + + # Starlette + starlette: pytest-asyncio + starlette: python-multipart + starlette: requests + # (this is a dependency of httpx) + starlette: anyio<4.0.0 + starlette: jinja2 + starlette-v{0.19,0.24,0.28,0.32,0.36}: httpx<0.28.0 + starlette-v0.40: httpx + starlette-latest: httpx + starlette-v0.19: starlette~=0.19.0 + starlette-v0.24: starlette~=0.24.0 + starlette-v0.28: starlette~=0.28.0 + starlette-v0.32: starlette~=0.32.0 + starlette-v0.36: starlette~=0.36.0 + starlette-v0.40: starlette~=0.40.0 + starlette-latest: starlette + + # Starlite + starlite: pytest-asyncio + starlite: python-multipart + starlite: requests + starlite: cryptography + starlite: pydantic<2.0.0 + starlite: httpx<0.28 + starlite-v{1.48}: starlite~=1.48.0 + starlite-v{1.51}: starlite~=1.51.0 + + # SQLAlchemy + sqlalchemy-v1.2: sqlalchemy~=1.2.0 + sqlalchemy-v1.4: sqlalchemy~=1.4.0 + sqlalchemy-v2.0: sqlalchemy~=2.0.0 + sqlalchemy-latest: sqlalchemy + + # Strawberry + strawberry: fastapi + strawberry: flask + strawberry: httpx + strawberry-v0.209: strawberry-graphql[fastapi,flask]~=0.209.0 + strawberry-v0.222: strawberry-graphql[fastapi,flask]~=0.222.0 + strawberry-latest: strawberry-graphql[fastapi,flask] + + # Tornado + # Tornado <6.4.1 is incompatible with Pytest ≥8.2 + # See https://github.com/tornadoweb/tornado/pull/3382. + tornado-{v6.0,v6.2}: pytest<8.2 + tornado-v6.0: tornado~=6.0.0 + tornado-v6.2: tornado~=6.2.0 + tornado-latest: tornado + + # Trytond + trytond: werkzeug + trytond-v4: werkzeug<1.0 + trytond-v4: trytond~=4.0 + trytond-v5: trytond~=5.0 + trytond-v6: trytond~=6.0 + trytond-v7: trytond~=7.0 + trytond-latest: trytond + + # Typer + typer-v0.15: typer~=0.15.0 + typer-latest: typer + + # === Integrations - Auto-generated === + # These come from the populate_tox.py script. Eventually we should move all + # integration tests there. + + {% for group, integrations in groups.items() %} + # ~~~ {{ group }} ~~~ + {% for integration in integrations %} + {% for release in integration.releases %} + {% if integration.extra %} + {{ integration.name }}-v{{ release }}: {{ integration.package }}[{{ integration.extra }}]=={{ release }} + {% else %} + {{ integration.name }}-v{{ release }}: {{ integration.package }}=={{ release }} + {% endif %} + {% endfor %} + {% for dep in integration.dependencies %} + {{ dep }} + {% endfor %} + + {% endfor %} + + {% endfor %} + +setenv = + PYTHONDONTWRITEBYTECODE=1 + OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES + COVERAGE_FILE=.coverage-sentry-{envname} + py3.6: COVERAGE_RCFILE=.coveragerc36 + + django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings + + common: TESTPATH=tests + gevent: TESTPATH=tests + aiohttp: TESTPATH=tests/integrations/aiohttp + anthropic: TESTPATH=tests/integrations/anthropic + ariadne: TESTPATH=tests/integrations/ariadne + arq: TESTPATH=tests/integrations/arq + asgi: TESTPATH=tests/integrations/asgi + asyncpg: TESTPATH=tests/integrations/asyncpg + aws_lambda: TESTPATH=tests/integrations/aws_lambda + beam: TESTPATH=tests/integrations/beam + boto3: TESTPATH=tests/integrations/boto3 + bottle: TESTPATH=tests/integrations/bottle + celery: TESTPATH=tests/integrations/celery + chalice: TESTPATH=tests/integrations/chalice + clickhouse_driver: TESTPATH=tests/integrations/clickhouse_driver + cohere: TESTPATH=tests/integrations/cohere + cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context + django: TESTPATH=tests/integrations/django + dramatiq: TESTPATH=tests/integrations/dramatiq + falcon: TESTPATH=tests/integrations/falcon + fastapi: TESTPATH=tests/integrations/fastapi + flask: TESTPATH=tests/integrations/flask + gcp: TESTPATH=tests/integrations/gcp + gql: TESTPATH=tests/integrations/gql + graphene: TESTPATH=tests/integrations/graphene + grpc: TESTPATH=tests/integrations/grpc + httpx: TESTPATH=tests/integrations/httpx + huey: TESTPATH=tests/integrations/huey + huggingface_hub: TESTPATH=tests/integrations/huggingface_hub + langchain: TESTPATH=tests/integrations/langchain + launchdarkly: TESTPATH=tests/integrations/launchdarkly + litestar: TESTPATH=tests/integrations/litestar + loguru: TESTPATH=tests/integrations/loguru + openai: TESTPATH=tests/integrations/openai + openfeature: TESTPATH=tests/integrations/openfeature + opentelemetry: TESTPATH=tests/integrations/opentelemetry + potel: TESTPATH=tests/integrations/opentelemetry + pure_eval: TESTPATH=tests/integrations/pure_eval + pymongo: TESTPATH=tests/integrations/pymongo + pyramid: TESTPATH=tests/integrations/pyramid + quart: TESTPATH=tests/integrations/quart + ray: TESTPATH=tests/integrations/ray + redis: TESTPATH=tests/integrations/redis + redis_py_cluster_legacy: TESTPATH=tests/integrations/redis_py_cluster_legacy + requests: TESTPATH=tests/integrations/requests + rq: TESTPATH=tests/integrations/rq + sanic: TESTPATH=tests/integrations/sanic + spark: TESTPATH=tests/integrations/spark + starlette: TESTPATH=tests/integrations/starlette + starlite: TESTPATH=tests/integrations/starlite + sqlalchemy: TESTPATH=tests/integrations/sqlalchemy + strawberry: TESTPATH=tests/integrations/strawberry + tornado: TESTPATH=tests/integrations/tornado + trytond: TESTPATH=tests/integrations/trytond + typer: TESTPATH=tests/integrations/typer + unleash: TESTPATH=tests/integrations/unleash + socket: TESTPATH=tests/integrations/socket + +passenv = + SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID + SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY + SENTRY_PYTHON_TEST_POSTGRES_HOST + SENTRY_PYTHON_TEST_POSTGRES_USER + SENTRY_PYTHON_TEST_POSTGRES_PASSWORD + SENTRY_PYTHON_TEST_POSTGRES_NAME + +usedevelop = True + +extras = + bottle: bottle + falcon: falcon + flask: flask + pymongo: pymongo + +basepython = + py3.6: python3.6 + py3.7: python3.7 + py3.8: python3.8 + py3.9: python3.9 + py3.10: python3.10 + py3.11: python3.11 + py3.12: python3.12 + py3.13: python3.13 + + # Python version is pinned here because flake8 actually behaves differently + # depending on which version is used. You can patch this out to point to + # some random Python 3 binary, but then you get guaranteed mismatches with + # CI. Other tools such as mypy and black have options that pin the Python + # version. + linters: python3.12 + +commands = + {py3.7,py3.8}-boto3: pip install urllib3<2.0.0 + + ; https://github.com/pallets/flask/issues/4455 + {py3.7,py3.8,py3.9,py3.10,py3.11}-flask-v{1}: pip install "itsdangerous>=0.24,<2.0" "markupsafe<2.0.0" "jinja2<3.1.1" + + ; Running `pytest` as an executable suffers from an import error + ; when loading tests in scenarios. In particular, django fails to + ; load the settings from the test module. + python -m pytest {env:TESTPATH} -o junit_suite_name={envname} {posargs} + +[testenv:linters] +commands = + flake8 tests sentry_sdk + black --check tests sentry_sdk + mypy sentry_sdk diff --git a/tox.ini b/tox.ini index 3cab20a1f1..c82d7d9159 100644 --- a/tox.ini +++ b/tox.ini @@ -2,6 +2,13 @@ # in multiple virtualenvs. This configuration file will run the # test suite on all supported python versions. To use it, "pip install tox" # and then run "tox" from this directory. +# +# This file has been generated from a template +# by "scripts/populate_tox/populate_tox.py". Any changes to the file should +# be made in the template (if you want to change a hardcoded part of the file) +# or in the script (if you want to change the auto-generated part). +# The file (and all resulting CI YAMLs) then need to be regenerated via +# "scripts/generate-test-files.sh". [tox] requires = @@ -294,6 +301,11 @@ envlist = {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-latest + # === Integrations - Auto-generated === + # These come from the populate_tox.py script. Eventually we should move all + # integration tests there. + + [testenv] deps = # if you change requirements-testing.txt and your change is not being reflected @@ -738,6 +750,11 @@ deps = typer-v0.15: typer~=0.15.0 typer-latest: typer + # === Integrations - Auto-generated === + # These come from the populate_tox.py script. Eventually we should move all + # integration tests there. + + setenv = PYTHONDONTWRITEBYTECODE=1 OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES From 2ebaa7cebf37c72caca10c24d2dd6f16c6a9e1ec Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 11 Feb 2025 11:26:25 +0100 Subject: [PATCH 120/591] ref(integrations): Add more min versions of frameworks (#3973) These mostly come from our existing `tox.ini`. They're used by the `populate_tox.py` script to filter out unsupported releases. They are not actually checked in the integrations. Since they were more of a suggestion before than a hard requirement, we don't want an integration to suddenly stop working for someone who is on an older version. We can consider actually checking them in a new major. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/integrations/__init__.py | 17 ++++++++++++++++- 1 file changed, 16 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 683382bb9a..45235a41c4 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -125,21 +125,36 @@ def iter_default_integrations(with_auto_enabling_integrations): "ariadne": (0, 20), "arq": (0, 23), "asyncpg": (0, 23), - "boto3": (1, 12), # this is actually the botocore version + "beam": (2, 12), + "boto3": (1, 12), # botocore "bottle": (0, 12), "celery": (4, 4, 7), + "chalice": (1, 16, 0), "clickhouse_driver": (0, 2, 0), "django": (1, 8), + "dramatiq": (1, 9), "falcon": (1, 4), + "fastapi": (0, 79, 0), "flask": (0, 10), "gql": (3, 4, 1), "graphene": (3, 3), + "grpc": (1, 32, 0), # grpcio + "huggingface_hub": (0, 22), + "langchain": (0, 0, 210), + "launchdarkly": (9, 8, 0), + "openai": (1, 0, 0), + "openfeature": (0, 7, 1), + "quart": (0, 16, 0), "ray": (2, 7, 0), + "requests": (2, 0, 0), "rq": (0, 6), "sanic": (0, 8), "sqlalchemy": (1, 2), + "starlite": (1, 48), "strawberry": (0, 209, 5), "tornado": (6, 0), + "typer": (0, 15), + "unleash": (6, 0, 1), } From 0cda7d9c5bfaa21e1b4a0c0b0c7cf194d17a8f4d Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 11 Feb 2025 14:24:09 +0100 Subject: [PATCH 121/591] test: Fix typo in test name (#4036) --- tests/test_propagationcontext.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/test_propagationcontext.py b/tests/test_propagationcontext.py index c650071511..85f82913f8 100644 --- a/tests/test_propagationcontext.py +++ b/tests/test_propagationcontext.py @@ -35,7 +35,7 @@ def test_context_with_values(): } -def test_lacy_uuids(): +def test_lazy_uuids(): ctx = PropagationContext() assert ctx._trace_id is None assert ctx._span_id is None From 3217ccab1497d695a563019167d3878d6cd13f7c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 11 Feb 2025 15:47:57 +0100 Subject: [PATCH 122/591] fix(integrations): Do not patch `execute` (#4026) New Strawberry version removes the `execute` and `execute_sync` functions that we were monkeypatching in favor of integrating the code directly in `Schema.execute` and `Schema.execute_sync`. We were previously patching `execute` instead of `Schema.execute` that's calling it because that way we had access to a populated `execution_context` which contains data that we wanted to put on the event via an event processor. We have access to the `execution_context` directly in the extension hooks Strawberry provides, so we now add the event processor there instead of monkeypatching anything. This should also work for older Strawberry versions, so shouldn't be necessary to keep the old implementation around for compat. Closes https://github.com/getsentry/sentry-python/issues/4037 --- requirements-linting.txt | 1 + sentry_sdk/integrations/ariadne.py | 2 +- sentry_sdk/integrations/gql.py | 7 ++- sentry_sdk/integrations/graphene.py | 4 +- sentry_sdk/integrations/strawberry.py | 70 ++++++++------------------- 5 files changed, 29 insertions(+), 55 deletions(-) diff --git a/requirements-linting.txt b/requirements-linting.txt index 4227acc26a..014e177793 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -19,3 +19,4 @@ openfeature-sdk launchdarkly-server-sdk UnleashClient typer +strawberry-graphql diff --git a/sentry_sdk/integrations/ariadne.py b/sentry_sdk/integrations/ariadne.py index 0336140441..1a95bc0145 100644 --- a/sentry_sdk/integrations/ariadne.py +++ b/sentry_sdk/integrations/ariadne.py @@ -25,7 +25,7 @@ if TYPE_CHECKING: from typing import Any, Dict, List, Optional from ariadne.types import GraphQLError, GraphQLResult, GraphQLSchema, QueryParser # type: ignore - from graphql.language.ast import DocumentNode # type: ignore + from graphql.language.ast import DocumentNode from sentry_sdk._types import Event, EventProcessor diff --git a/sentry_sdk/integrations/gql.py b/sentry_sdk/integrations/gql.py index d5341d2cf6..5f4436f5b2 100644 --- a/sentry_sdk/integrations/gql.py +++ b/sentry_sdk/integrations/gql.py @@ -10,7 +10,12 @@ try: import gql # type: ignore[import-not-found] - from graphql import print_ast, get_operation_ast, DocumentNode, VariableDefinitionNode # type: ignore[import-not-found] + from graphql import ( + print_ast, + get_operation_ast, + DocumentNode, + VariableDefinitionNode, + ) from gql.transport import Transport, AsyncTransport # type: ignore[import-not-found] from gql.transport.exceptions import TransportQueryError # type: ignore[import-not-found] except ImportError: diff --git a/sentry_sdk/integrations/graphene.py b/sentry_sdk/integrations/graphene.py index 198aea50d2..00a8d155d4 100644 --- a/sentry_sdk/integrations/graphene.py +++ b/sentry_sdk/integrations/graphene.py @@ -22,8 +22,8 @@ from collections.abc import Generator from typing import Any, Dict, Union from graphene.language.source import Source # type: ignore - from graphql.execution import ExecutionResult # type: ignore - from graphql.type import GraphQLSchema # type: ignore + from graphql.execution import ExecutionResult + from graphql.type import GraphQLSchema from sentry_sdk._types import Event diff --git a/sentry_sdk/integrations/strawberry.py b/sentry_sdk/integrations/strawberry.py index d27e0eaf1c..f12019cd60 100644 --- a/sentry_sdk/integrations/strawberry.py +++ b/sentry_sdk/integrations/strawberry.py @@ -27,16 +27,17 @@ raise DidNotEnable("strawberry-graphql integration requires Python 3.8 or newer") try: - import strawberry.schema.schema as strawberry_schema # type: ignore from strawberry import Schema - from strawberry.extensions import SchemaExtension # type: ignore - from strawberry.extensions.tracing.utils import should_skip_tracing as strawberry_should_skip_tracing # type: ignore - from strawberry.http import async_base_view, sync_base_view # type: ignore + from strawberry.extensions import SchemaExtension + from strawberry.extensions.tracing.utils import ( + should_skip_tracing as strawberry_should_skip_tracing, + ) + from strawberry.http import async_base_view, sync_base_view except ImportError: raise DidNotEnable("strawberry-graphql is not installed") try: - from strawberry.extensions.tracing import ( # type: ignore + from strawberry.extensions.tracing import ( SentryTracingExtension as StrawberrySentryAsyncExtension, SentryTracingExtensionSync as StrawberrySentrySyncExtension, ) @@ -47,10 +48,10 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Any, Callable, Generator, List, Optional, Union - from graphql import GraphQLError, GraphQLResolveInfo # type: ignore + from typing import Any, Callable, Generator, List, Optional + from graphql import GraphQLError, GraphQLResolveInfo from strawberry.http import GraphQLHTTPResponse - from strawberry.types import ExecutionContext, ExecutionResult, SubscriptionExecutionResult # type: ignore + from strawberry.types import ExecutionContext from sentry_sdk._types import Event, EventProcessor @@ -78,7 +79,6 @@ def setup_once(): _check_minimum_version(StrawberryIntegration, version, "strawberry-graphql") _patch_schema_init() - _patch_execute() _patch_views() @@ -124,10 +124,10 @@ def _sentry_patched_schema_init(self, *args, **kwargs): return old_schema_init(self, *args, **kwargs) - Schema.__init__ = _sentry_patched_schema_init + Schema.__init__ = _sentry_patched_schema_init # type: ignore[method-assign] -class SentryAsyncExtension(SchemaExtension): # type: ignore +class SentryAsyncExtension(SchemaExtension): def __init__( self, *, @@ -140,7 +140,7 @@ def __init__( @cached_property def _resource_name(self): # type: () -> str - query_hash = self.hash_query(self.execution_context.query) + query_hash = self.hash_query(self.execution_context.query) # type: ignore if self.execution_context.operation_name: return "{}:{}".format(self.execution_context.operation_name, query_hash) @@ -180,6 +180,10 @@ def on_operation(self): }, ) + scope = sentry_sdk.get_isolation_scope() + event_processor = _make_request_event_processor(self.execution_context) + scope.add_event_processor(event_processor) + span = sentry_sdk.get_current_span() if span: self.graphql_span = span.start_child( @@ -287,41 +291,6 @@ def resolve(self, _next, root, info, *args, **kwargs): return _next(root, info, *args, **kwargs) -def _patch_execute(): - # type: () -> None - old_execute_async = strawberry_schema.execute - old_execute_sync = strawberry_schema.execute_sync - - async def _sentry_patched_execute_async(*args, **kwargs): - # type: (Any, Any) -> Union[ExecutionResult, SubscriptionExecutionResult] - result = await old_execute_async(*args, **kwargs) - - if sentry_sdk.get_client().get_integration(StrawberryIntegration) is None: - return result - - if "execution_context" in kwargs: - scope = sentry_sdk.get_isolation_scope() - event_processor = _make_request_event_processor(kwargs["execution_context"]) - scope.add_event_processor(event_processor) - - return result - - @ensure_integration_enabled(StrawberryIntegration, old_execute_sync) - def _sentry_patched_execute_sync(*args, **kwargs): - # type: (Any, Any) -> ExecutionResult - result = old_execute_sync(*args, **kwargs) - - if "execution_context" in kwargs: - scope = sentry_sdk.get_isolation_scope() - event_processor = _make_request_event_processor(kwargs["execution_context"]) - scope.add_event_processor(event_processor) - - return result - - strawberry_schema.execute = _sentry_patched_execute_async - strawberry_schema.execute_sync = _sentry_patched_execute_sync - - def _patch_views(): # type: () -> None old_async_view_handle_errors = async_base_view.AsyncBaseHTTPView._handle_errors @@ -359,10 +328,10 @@ def _sentry_patched_handle_errors(self, errors, response_data): ) sentry_sdk.capture_event(event, hint=hint) - async_base_view.AsyncBaseHTTPView._handle_errors = ( + async_base_view.AsyncBaseHTTPView._handle_errors = ( # type: ignore[method-assign] _sentry_patched_async_view_handle_errors ) - sync_base_view.SyncBaseHTTPView._handle_errors = ( + sync_base_view.SyncBaseHTTPView._handle_errors = ( # type: ignore[method-assign] _sentry_patched_sync_view_handle_errors ) @@ -378,8 +347,7 @@ def inner(event, hint): request_data["api_target"] = "graphql" if not request_data.get("data"): - data = {"query": execution_context.query} - + data = {"query": execution_context.query} # type: dict[str, Any] if execution_context.variables: data["variables"] = execution_context.variables if execution_context.operation_name: From d9372724a0a9addde5d5f864160868719142ac69 Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Tue, 11 Feb 2025 09:36:00 -0600 Subject: [PATCH 123/591] fix(flags): Fix bug where concurrent accesses to the flags property could raise a RunTime error (#4034) On error the SDK deep copies the flag buffer. If the SDK is receiving flags at the same time, the buffer copy can potentially raise a RunTime error. To fix this we guard the FlagBuffer with a lock. Fixes: https://sentry.sentry.io/issues/6286673308/?project=1 --- sentry_sdk/feature_flags.py | 36 +++++++++++++++++++++++++++++++----- tests/test_feature_flags.py | 34 ++++++++++++++++++++++++++++++++++ 2 files changed, 65 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/feature_flags.py b/sentry_sdk/feature_flags.py index 1187c2fa12..a0b1338356 100644 --- a/sentry_sdk/feature_flags.py +++ b/sentry_sdk/feature_flags.py @@ -1,7 +1,9 @@ +import copy import sentry_sdk from sentry_sdk._lru_cache import LRUCache +from threading import Lock -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, Any if TYPE_CHECKING: from typing import TypedDict @@ -16,20 +18,44 @@ class FlagBuffer: def __init__(self, capacity): # type: (int) -> None - self.buffer = LRUCache(capacity) self.capacity = capacity + self.lock = Lock() + + # Buffer is private. The name is mangled to discourage use. If you use this attribute + # directly you're on your own! + self.__buffer = LRUCache(capacity) def clear(self): # type: () -> None - self.buffer = LRUCache(self.capacity) + self.__buffer = LRUCache(self.capacity) + + def __deepcopy__(self, memo): + # type: (dict[int, Any]) -> FlagBuffer + with self.lock: + buffer = FlagBuffer(self.capacity) + buffer.__buffer = copy.deepcopy(self.__buffer, memo) + return buffer def get(self): # type: () -> list[FlagData] - return [{"flag": key, "result": value} for key, value in self.buffer.get_all()] + with self.lock: + return [ + {"flag": key, "result": value} for key, value in self.__buffer.get_all() + ] def set(self, flag, result): # type: (str, bool) -> None - self.buffer.set(flag, result) + if isinstance(result, FlagBuffer): + # If someone were to insert `self` into `self` this would create a circular dependency + # on the lock. This is of course a deadlock. However, this is far outside the expected + # usage of this class. We guard against it here for completeness and to document this + # expected failure mode. + raise ValueError( + "FlagBuffer instances can not be inserted into the dictionary." + ) + + with self.lock: + self.__buffer.set(flag, result) def add_feature_flag(flag, result): diff --git a/tests/test_feature_flags.py b/tests/test_feature_flags.py index 14d74cb04b..4469b5c2ca 100644 --- a/tests/test_feature_flags.py +++ b/tests/test_feature_flags.py @@ -1,5 +1,7 @@ import concurrent.futures as cf import sys +import copy +import threading import pytest @@ -167,3 +169,35 @@ def test_flag_tracking(): {"flag": "e", "result": False}, {"flag": "f", "result": False}, ] + + +def test_flag_buffer_concurrent_access(): + buffer = FlagBuffer(capacity=100) + error_occurred = False + + def writer(): + for i in range(1_000_000): + buffer.set(f"key_{i}", True) + + def reader(): + nonlocal error_occurred + + try: + for _ in range(1000): + copy.deepcopy(buffer) + except RuntimeError: + error_occurred = True + + writer_thread = threading.Thread(target=writer) + reader_thread = threading.Thread(target=reader) + + writer_thread.start() + reader_thread.start() + + writer_thread.join(timeout=5) + reader_thread.join(timeout=5) + + # This should always be false. If this ever fails we know we have concurrent access to a + # shared resource. When deepcopying we should have exclusive access to the underlying + # memory. + assert error_occurred is False From c227e11460a9cde0562ea660fdd6de8942485e83 Mon Sep 17 00:00:00 2001 From: Matt Purnell <65473602+mpurnell1@users.noreply.github.com> Date: Tue, 11 Feb 2025 09:44:28 -0600 Subject: [PATCH 124/591] ref(utils): Explicitly use None default when checking metadata (#4039) Fixes #4035 As described in the above issue, starting in Python 3.14 importlib_metadata 8 provides the desired behavior, raising KeyError on a missing key. In preparation for this change, and to remove a DeprecationWarning, we should explicitly default to None when getting metadata. --- sentry_sdk/utils.py | 2 +- tests/test_utils.py | 8 ++++---- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index f60c31e676..b2a39b7af1 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1665,7 +1665,7 @@ def _generate_installed_modules(): yielded = set() for dist in metadata.distributions(): - name = dist.metadata["Name"] + name = dist.metadata.get("Name", None) # type: ignore[attr-defined] # `metadata` values may be `None`, see: # https://github.com/python/cpython/issues/91216 # and diff --git a/tests/test_utils.py b/tests/test_utils.py index 894638bf4d..6083ad7ad2 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -650,12 +650,12 @@ def test_installed_modules(): if importlib_available: importlib_distributions = { - _normalize_distribution_name(dist.metadata["Name"]): version( - dist.metadata["Name"] + _normalize_distribution_name(dist.metadata.get("Name", None)): version( + dist.metadata.get("Name", None) ) for dist in distributions() - if dist.metadata["Name"] is not None - and version(dist.metadata["Name"]) is not None + if dist.metadata.get("Name", None) is not None + and version(dist.metadata.get("Name", None)) is not None } assert installed_distributions == importlib_distributions From 2f51db730f6e2297bf1c9c891d05c6b8ee8db8b6 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Wed, 12 Feb 2025 10:13:08 +0100 Subject: [PATCH 125/591] feat(tracing): Add `__repr__` to `Baggage` (#4043) The default `__repr__` does not show what is in the `Baggage`, making it extremely difficult to debug code involving `Baggage` objects. Add a `__repr__` which includes the serialized `Baggage` to improve debuggability. --- sentry_sdk/tracing_utils.py | 4 ++++ tests/test_tracing_utils.py | 31 +++++++++++++++++++++++++++++++ 2 files changed, 35 insertions(+) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 9ea2d9859a..a1cfd729c2 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -638,6 +638,10 @@ def strip_sentry_baggage(header): ) ) + def __repr__(self): + # type: () -> str + return f'' + def should_propagate_trace(client, url): # type: (sentry_sdk.client.BaseClient, str) -> bool diff --git a/tests/test_tracing_utils.py b/tests/test_tracing_utils.py index 5c1f70516d..2b2c62a6f9 100644 --- a/tests/test_tracing_utils.py +++ b/tests/test_tracing_utils.py @@ -115,3 +115,34 @@ def test_should_be_included(test_case, expected): ) def test_strip_sentry_baggage(header, expected): assert Baggage.strip_sentry_baggage(header) == expected + + +@pytest.mark.parametrize( + ("baggage", "expected_repr"), + ( + (Baggage(sentry_items={}), ''), + (Baggage(sentry_items={}, mutable=False), ''), + ( + Baggage(sentry_items={"foo": "bar"}), + '', + ), + ( + Baggage(sentry_items={"foo": "bar"}, mutable=False), + '', + ), + ( + Baggage(sentry_items={"foo": "bar"}, third_party_items="asdf=1234,"), + '', + ), + ( + Baggage( + sentry_items={"foo": "bar"}, + third_party_items="asdf=1234,", + mutable=False, + ), + '', + ), + ), +) +def test_baggage_repr(baggage, expected_repr): + assert repr(baggage) == expected_repr From d7dff6d8f8d794bfb7d7ee36bab56515e338017d Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Wed, 12 Feb 2025 09:54:41 +0000 Subject: [PATCH 126/591] release: 2.21.0 --- CHANGELOG.md | 31 +++++++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 34 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 80ff6c2796..8402a18f81 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,36 @@ # Changelog +## 2.21.0 + +### Various fixes & improvements + +- feat(tracing): Add `__repr__` to `Baggage` (#4043) by @szokeasaurusrex +- ref(utils): Explicitly use None default when checking metadata (#4039) by @mpurnell1 +- fix(flags): Fix bug where concurrent accesses to the flags property could raise a RunTime error (#4034) by @cmanallen +- fix(integrations): Do not patch `execute` (#4026) by @sentrivana +- test: Fix typo in test name (#4036) by @szokeasaurusrex +- ref(integrations): Add more min versions of frameworks (#3973) by @sentrivana +- [1] Add tox generation script, but don't use it yet (#3971) by @sentrivana +- Set level based on status code for HTTP client breadcrumbs (#4004) by @sentrivana +- build(deps): bump actions/create-github-app-token from 1.11.2 to 1.11.3 (#4023) by @dependabot +- Don't set transaction status to error on sys.exit(0) (#4025) by @sentrivana +- feat(litestar): Add `failed_request_status_codes` (#4021) by @vrslev +- build(deps): bump actions/create-github-app-token from 1.11.1 to 1.11.2 (#4015) by @dependabot +- Fix mypy (#4019) by @sentrivana +- feat(profiling): Continuous profiling sample rate (#4002) by @Zylphrex +- feat(spans): track and report spans that were dropped (#4005) by @constantinius +- chore(profiling): Change continuous profile buffer size (#3987) by @Zylphrex +- Handle MultiPartParserError to avoid internal sentry crash (#4001) by @orhanhenrik +- fix(ci): Various errors on master (#4009) by @Zylphrex +- build(deps): bump codecov/codecov-action from 5.1.2 to 5.3.1 (#3995) by @dependabot +- Deprecate `enable_tracing` option (#3935) by @antonpirker +- Split gevent tests off (#3964) by @sentrivana +- Add support for Python 3.12 and 3.13 to AWS Lambda integration. (#3965) by @antonpirker +- Use httpx_mock in test_httpx (#3967) by @sentrivana +- fix(utils): Check that `__module__` is `str` (#3942) by @szokeasaurusrex + +_Plus 4 more_ + ## 2.20.0 - **New integration:** Add [Typer](https://typer.tiangolo.com/) integration (#3869) by @patrick91 diff --git a/docs/conf.py b/docs/conf.py index 1d58274beb..b7ae919e9a 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.20.0" +release = "2.21.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index ce435de36b..876556776c 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -582,4 +582,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.20.0" +VERSION = "2.21.0" diff --git a/setup.py b/setup.py index 1bfbb6f7e4..760ce2d60f 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.20.0", + version="2.21.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From dc1460aedddf96befe56cd09815af31bc09a33a0 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 12 Feb 2025 11:05:47 +0100 Subject: [PATCH 127/591] Update CHANGELOG.md --- CHANGELOG.md | 54 +++++++++++++++++++++++++++++----------------------- 1 file changed, 30 insertions(+), 24 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8402a18f81..0229aac66f 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,32 +4,38 @@ ### Various fixes & improvements -- feat(tracing): Add `__repr__` to `Baggage` (#4043) by @szokeasaurusrex -- ref(utils): Explicitly use None default when checking metadata (#4039) by @mpurnell1 -- fix(flags): Fix bug where concurrent accesses to the flags property could raise a RunTime error (#4034) by @cmanallen -- fix(integrations): Do not patch `execute` (#4026) by @sentrivana -- test: Fix typo in test name (#4036) by @szokeasaurusrex -- ref(integrations): Add more min versions of frameworks (#3973) by @sentrivana -- [1] Add tox generation script, but don't use it yet (#3971) by @sentrivana -- Set level based on status code for HTTP client breadcrumbs (#4004) by @sentrivana -- build(deps): bump actions/create-github-app-token from 1.11.2 to 1.11.3 (#4023) by @dependabot -- Don't set transaction status to error on sys.exit(0) (#4025) by @sentrivana -- feat(litestar): Add `failed_request_status_codes` (#4021) by @vrslev -- build(deps): bump actions/create-github-app-token from 1.11.1 to 1.11.2 (#4015) by @dependabot -- Fix mypy (#4019) by @sentrivana -- feat(profiling): Continuous profiling sample rate (#4002) by @Zylphrex -- feat(spans): track and report spans that were dropped (#4005) by @constantinius -- chore(profiling): Change continuous profile buffer size (#3987) by @Zylphrex -- Handle MultiPartParserError to avoid internal sentry crash (#4001) by @orhanhenrik -- fix(ci): Various errors on master (#4009) by @Zylphrex -- build(deps): bump codecov/codecov-action from 5.1.2 to 5.3.1 (#3995) by @dependabot +- Fix incompatibility with new Strawberry version (#4026) by @sentrivana +- Add `failed_request_status_codes` to Litestar (#4021) by @vrslev + + See https://docs.sentry.io/platforms/python/integrations/litestar/ for details. - Deprecate `enable_tracing` option (#3935) by @antonpirker -- Split gevent tests off (#3964) by @sentrivana -- Add support for Python 3.12 and 3.13 to AWS Lambda integration. (#3965) by @antonpirker -- Use httpx_mock in test_httpx (#3967) by @sentrivana -- fix(utils): Check that `__module__` is `str` (#3942) by @szokeasaurusrex -_Plus 4 more_ + The `enable_tracing` option is now deprecated. Please use `traces_sample_rate` instead. See https://docs.sentry.io/platforms/python/configuration/options/#traces_sample_rate for more information. +- Explicitly use `None` default when checking metadata (#4039) by @mpurnell1 +- Fix bug where concurrent accesses to the flags property could raise a `RuntimeError` (#4034) by @cmanallen +- Add more min versions of frameworks (#3973) by @sentrivana +- Set level based on status code for HTTP client breadcrumbs (#4004) by @sentrivana +- Don't set transaction status to error on `sys.exit(0)` (#4025) by @sentrivana +- Continuous profiling sample rate (#4002) by @Zylphrex +- Track and report spans that were dropped (#4005) by @constantinius +- Change continuous profile buffer size (#3987) by @Zylphrex +- Handle `MultiPartParserError` to avoid internal sentry crash (#4001) by @orhanhenrik +- Handle `None` lineno in `get_source_context` (#3925) by @sentrivana +- Add support for Python 3.12 and 3.13 to AWS Lambda integration (#3965) by @antonpirker +- Add `propagate_traces` deprecation warning (#3899) by @mgaligniana +- Check that `__module__` is `str` (#3942) by @szokeasaurusrex +- Add `__repr__` to `Baggage` (#4043) by @szokeasaurusrex +- Fix a typo (#3923) by @antonpirker +- Fix various CI errors on master (#4009) by @Zylphrex +- Split gevent tests off (#3964) by @sentrivana +- Add tox generation script, but don't use it yet (#3971) by @sentrivana +- Use `httpx_mock` in `test_httpx` (#3967) by @sl0thentr0py +- Fix typo in test name (#4036) by @szokeasaurusrex +- Fix mypy (#4019) by @sentrivana +- Test Celery's latest RC (#3938) by @sentrivana +- Bump `actions/create-github-app-token` from `1.11.2` to `1.11.3` (#4023) by @dependabot +- Bump `actions/create-github-app-token` from `1.11.1` to `1.11.2` (#4015) by @dependabot +- Bump `codecov/codecov-action` from `5.1.2` to `5.3.1` (#3995) by @dependabot ## 2.20.0 From 221f105bf8ef65ddfe4f20d57947e9b13fc10f42 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 12 Feb 2025 14:22:17 +0100 Subject: [PATCH 128/591] Update sample rate in DSC (#4018) - update `sample_rate` in DSC after the initial sampling decision is made - fix some typos Part of https://github.com/getsentry/sentry-python/issues/3999 --- sentry_sdk/scope.py | 12 ++++ sentry_sdk/tracing.py | 1 - sentry_sdk/tracing_utils.py | 2 +- tests/integrations/stdlib/test_httplib.py | 15 ++-- tests/test_dsc.py | 83 ++++++++++++++++++++++- tests/tracing/test_integration_tests.py | 21 ++++-- tests/tracing/test_sampling.py | 21 +++--- 7 files changed, 127 insertions(+), 28 deletions(-) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index c22cdfb030..53191c45da 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -1043,6 +1043,18 @@ def start_transaction( sampling_context.update(custom_sampling_context) transaction._set_initial_sampling_decision(sampling_context=sampling_context) + # update the sample rate in the dsc + if transaction.sample_rate is not None: + propagation_context = self.get_active_propagation_context() + if propagation_context: + dsc = propagation_context.dynamic_sampling_context + if dsc is not None: + dsc["sample_rate"] = str(transaction.sample_rate) + if transaction._baggage: + transaction._baggage.sentry_items["sample_rate"] = str( + transaction.sample_rate + ) + if transaction.sampled: profile = Profile( transaction.sampled, transaction._start_timestamp_monotonic_ns diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 59473d752c..2692944cf9 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1070,7 +1070,6 @@ def get_baggage(self): The first time a new baggage with Sentry items is made, it will be frozen.""" - if not self._baggage or self._baggage.mutable: self._baggage = Baggage.populate_from_transaction(self) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index a1cfd729c2..ae72b8cce9 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -392,7 +392,7 @@ def __init__( self.parent_sampled = parent_sampled """Boolean indicator if the parent span was sampled. Important when the parent span originated in an upstream service, - because we watn to sample the whole trace, or nothing from the trace.""" + because we want to sample the whole trace, or nothing from the trace.""" self.dynamic_sampling_context = dynamic_sampling_context """Data that is used for dynamic sampling decisions.""" diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 7f2c5d68b2..f2de190de0 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -185,12 +185,13 @@ def test_outgoing_trace_headers(sentry_init, monkeypatch): sentry_init(traces_sample_rate=1.0) - headers = {} - headers["baggage"] = ( - "other-vendor-value-1=foo;bar;baz, sentry-trace_id=771a43a4192642f0b136d5159a501700, " - "sentry-public_key=49d0f7386ad645858ae85020e393bef3, sentry-sample_rate=0.01337, " - "sentry-user_id=Am%C3%A9lie, other-vendor-value-2=foo;bar;" - ) + headers = { + "baggage": ( + "other-vendor-value-1=foo;bar;baz, sentry-trace_id=771a43a4192642f0b136d5159a501700, " + "sentry-public_key=49d0f7386ad645858ae85020e393bef3, sentry-sample_rate=0.01337, " + "sentry-user_id=Am%C3%A9lie, other-vendor-value-2=foo;bar;" + ), + } transaction = Transaction.continue_from_headers(headers) @@ -220,7 +221,7 @@ def test_outgoing_trace_headers(sentry_init, monkeypatch): expected_outgoing_baggage = ( "sentry-trace_id=771a43a4192642f0b136d5159a501700," "sentry-public_key=49d0f7386ad645858ae85020e393bef3," - "sentry-sample_rate=0.01337," + "sentry-sample_rate=1.0," "sentry-user_id=Am%C3%A9lie" ) diff --git a/tests/test_dsc.py b/tests/test_dsc.py index 3b8cff5baf..4837384a8e 100644 --- a/tests/test_dsc.py +++ b/tests/test_dsc.py @@ -8,6 +8,9 @@ This is not tested in this file. """ +import random +from unittest import mock + import pytest import sentry_sdk @@ -115,7 +118,85 @@ def test_dsc_continuation_of_trace(sentry_init, capture_envelopes): assert "sample_rate" in envelope_trace_header assert type(envelope_trace_header["sample_rate"]) == str - assert envelope_trace_header["sample_rate"] == "0.01337" + assert envelope_trace_header["sample_rate"] == "1.0" + + assert "sampled" in envelope_trace_header + assert type(envelope_trace_header["sampled"]) == str + assert envelope_trace_header["sampled"] == "true" + + assert "release" in envelope_trace_header + assert type(envelope_trace_header["release"]) == str + assert envelope_trace_header["release"] == "myfrontend@1.2.3" + + assert "environment" in envelope_trace_header + assert type(envelope_trace_header["environment"]) == str + assert envelope_trace_header["environment"] == "bird" + + assert "transaction" in envelope_trace_header + assert type(envelope_trace_header["transaction"]) == str + assert envelope_trace_header["transaction"] == "bar" + + +def test_dsc_continuation_of_trace_sample_rate_changed_in_traces_sampler( + sentry_init, capture_envelopes +): + """ + Another service calls our service and passes tracing information to us. + Our service is continuing the trace, but modifies the sample rate. + The DSC propagated further should contain the updated sample rate. + """ + + def my_traces_sampler(sampling_context): + return 0.25 + + sentry_init( + dsn="https://mysecret@bla.ingest.sentry.io/12312012", + release="myapp@0.0.1", + environment="canary", + traces_sampler=my_traces_sampler, + ) + envelopes = capture_envelopes() + + # This is what the upstream service sends us + sentry_trace = "771a43a4192642f0b136d5159a501700-1234567890abcdef-1" + baggage = ( + "other-vendor-value-1=foo;bar;baz, " + "sentry-trace_id=771a43a4192642f0b136d5159a501700, " + "sentry-public_key=frontendpublickey, " + "sentry-sample_rate=1.0, " + "sentry-sampled=true, " + "sentry-release=myfrontend@1.2.3, " + "sentry-environment=bird, " + "sentry-transaction=bar, " + "other-vendor-value-2=foo;bar;" + ) + incoming_http_headers = { + "HTTP_SENTRY_TRACE": sentry_trace, + "HTTP_BAGGAGE": baggage, + } + + # We continue the incoming trace and start a new transaction + with mock.patch.object(random, "random", return_value=0.2): + transaction = sentry_sdk.continue_trace(incoming_http_headers) + with sentry_sdk.start_transaction(transaction, name="foo"): + pass + + assert len(envelopes) == 1 + + transaction_envelope = envelopes[0] + envelope_trace_header = transaction_envelope.headers["trace"] + + assert "trace_id" in envelope_trace_header + assert type(envelope_trace_header["trace_id"]) == str + assert envelope_trace_header["trace_id"] == "771a43a4192642f0b136d5159a501700" + + assert "public_key" in envelope_trace_header + assert type(envelope_trace_header["public_key"]) == str + assert envelope_trace_header["public_key"] == "frontendpublickey" + + assert "sample_rate" in envelope_trace_header + assert type(envelope_trace_header["sample_rate"]) == str + assert envelope_trace_header["sample_rate"] == "0.25" assert "sampled" in envelope_trace_header assert type(envelope_trace_header["sampled"]) == str diff --git a/tests/tracing/test_integration_tests.py b/tests/tracing/test_integration_tests.py index f269023f87..13d1a7a77b 100644 --- a/tests/tracing/test_integration_tests.py +++ b/tests/tracing/test_integration_tests.py @@ -53,9 +53,11 @@ def test_basic(sentry_init, capture_events, sample_rate): assert not events -@pytest.mark.parametrize("sampled", [True, False, None]) +@pytest.mark.parametrize("parent_sampled", [True, False, None]) @pytest.mark.parametrize("sample_rate", [0.0, 1.0]) -def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_rate): +def test_continue_from_headers( + sentry_init, capture_envelopes, parent_sampled, sample_rate +): """ Ensure data is actually passed along via headers, and that they are read correctly. @@ -66,7 +68,7 @@ def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_r # make a parent transaction (normally this would be in a different service) with start_transaction(name="hi", sampled=True if sample_rate == 0 else None): with start_span() as old_span: - old_span.sampled = sampled + old_span.sampled = parent_sampled headers = dict( sentry_sdk.get_current_scope().iter_trace_propagation_headers(old_span) ) @@ -81,7 +83,7 @@ def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_r # child transaction, to prove that we can read 'sentry-trace' header data correctly child_transaction = Transaction.continue_from_headers(headers, name="WRONG") assert child_transaction is not None - assert child_transaction.parent_sampled == sampled + assert child_transaction.parent_sampled == parent_sampled assert child_transaction.trace_id == old_span.trace_id assert child_transaction.same_process_as_parent is False assert child_transaction.parent_span_id == old_span.span_id @@ -106,8 +108,8 @@ def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_r sentry_sdk.get_current_scope().transaction = "ho" capture_message("hello") - # in this case the child transaction won't be captured - if sampled is False or (sample_rate == 0 and sampled is None): + if parent_sampled is False or (sample_rate == 0 and parent_sampled is None): + # in this case the child transaction won't be captured trace1, message = envelopes message_payload = message.get_event() trace1_payload = trace1.get_transaction_event() @@ -129,12 +131,17 @@ def test_continue_from_headers(sentry_init, capture_envelopes, sampled, sample_r == message_payload["contexts"]["trace"]["trace_id"] ) + if parent_sampled is not None: + expected_sample_rate = str(float(parent_sampled)) + else: + expected_sample_rate = str(sample_rate) + assert trace2.headers["trace"] == baggage.dynamic_sampling_context() assert trace2.headers["trace"] == { "public_key": "49d0f7386ad645858ae85020e393bef3", "trace_id": "771a43a4192642f0b136d5159a501700", "user_id": "Amelie", - "sample_rate": "0.01337", + "sample_rate": expected_sample_rate, } assert message_payload["message"] == "hello" diff --git a/tests/tracing/test_sampling.py b/tests/tracing/test_sampling.py index 2e6ed0dab3..1ad08ecec2 100644 --- a/tests/tracing/test_sampling.py +++ b/tests/tracing/test_sampling.py @@ -198,20 +198,19 @@ def test_passes_parent_sampling_decision_in_sampling_context( transaction = Transaction.continue_from_headers( headers={"sentry-trace": sentry_trace_header}, name="dogpark" ) - spy = mock.Mock(wraps=transaction) - start_transaction(transaction=spy) - # there's only one call (so index at 0) and kwargs are always last in a call - # tuple (so index at -1) - sampling_context = spy._set_initial_sampling_decision.mock_calls[0][-1][ - "sampling_context" - ] - assert "parent_sampled" in sampling_context - # because we passed in a spy, attribute access requires unwrapping - assert sampling_context["parent_sampled"]._mock_wraps is parent_sampling_decision + def mock_set_initial_sampling_decision(_, sampling_context): + assert "parent_sampled" in sampling_context + assert sampling_context["parent_sampled"] is parent_sampling_decision + with mock.patch( + "sentry_sdk.tracing.Transaction._set_initial_sampling_decision", + mock_set_initial_sampling_decision, + ): + start_transaction(transaction=transaction) -def test_passes_custom_samling_context_from_start_transaction_to_traces_sampler( + +def test_passes_custom_sampling_context_from_start_transaction_to_traces_sampler( sentry_init, DictionaryContaining # noqa: N803 ): traces_sampler = mock.Mock() From a78af17e1935a8992a7d5d7ae835320c5b1e2eb8 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 12 Feb 2025 15:35:28 +0100 Subject: [PATCH 129/591] Move the GraphQL group over to the tox gen script (#3975) - remove hardcoded entries for `ariadne`, `gql`, `graphene`, `strawberry` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini Note that this effectively eliminates the `-latest` tests for the GraphQL group. The script doesn't generate any `-latest` tests since it always makes sure to add a pinned entry for the latest version. So in case all of the integrations in a single group are using the script, the whole `-latest` test category is removed. --- .../workflows/test-integrations-graphql.yml | 70 +-------------- scripts/populate_tox/config.py | 26 +++++- scripts/populate_tox/populate_tox.py | 14 ++- scripts/populate_tox/tox.jinja | 44 --------- tox.ini | 89 ++++++++++--------- 5 files changed, 76 insertions(+), 167 deletions(-) diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index d7cf8d80c1..f3015ae5bf 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -22,74 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-graphql-latest: - name: GraphQL (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.7","3.8","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] - steps: - - uses: actions/checkout@v4.2.2 - - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test ariadne latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-ariadne-latest" - - name: Test gql latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-gql-latest" - - name: Test graphene latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-graphene-latest" - - name: Test strawberry latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-strawberry-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-graphql-pinned: name: GraphQL (pinned) timeout-minutes: 30 @@ -97,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.8","3.11","3.12"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 9e1366c25b..8cdd36c05d 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -5,4 +5,28 @@ # # See scripts/populate_tox/README.md for more info on the format and examples. -TEST_SUITE_CONFIG = {} +TEST_SUITE_CONFIG = { + "ariadne": { + "package": "ariadne", + "deps": { + "*": ["fastapi", "flask", "httpx"], + }, + "python": ">=3.8", + }, + "gql": { + "package": "gql[all]", + }, + "graphene": { + "package": "graphene", + "deps": { + "*": ["blinker", "fastapi", "flask", "httpx"], + "py3.6": ["aiocontextvars"], + }, + }, + "strawberry": { + "package": "strawberry-graphql[fastapi,flask]", + "deps": { + "*": ["httpx"], + }, + }, +} diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 83db87bd35..60770d5832 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -56,7 +56,6 @@ "potel", "aiohttp", "anthropic", - "ariadne", "arq", "asgi", "asyncpg", @@ -76,8 +75,6 @@ "fastapi", "flask", "gcp", - "gql", - "graphene", "grpc", "httpx", "huey", @@ -104,7 +101,6 @@ "starlette", "starlite", "sqlalchemy", - "strawberry", "tornado", "trytond", "typer", @@ -464,7 +460,9 @@ def _compare_min_version_with_defined( ) -def _add_python_versions_to_release(integration: str, package: str, release: Version): +def _add_python_versions_to_release( + integration: str, package: str, release: Version +) -> None: release_pypi_data = fetch_release(package, release) time.sleep(0.1) # give PYPI some breathing room @@ -522,10 +520,8 @@ def main() -> None: test_releases = pick_releases_to_test(releases) for release in test_releases: - py_versions = _add_python_versions_to_release( - integration, package, release - ) - if not py_versions: + _add_python_versions_to_release(integration, package, release) + if not release.python_versions: print(f" Release {release} has no Python versions, skipping.") test_releases = [ diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index b60c6f137a..ad569b17a6 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -43,10 +43,6 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} {py3.7,py3.11,py3.12}-anthropic-latest - # Ariadne - {py3.8,py3.11}-ariadne-v{0.20} - {py3.8,py3.12,py3.13}-ariadne-latest - # Arq {py3.7,py3.11}-arq-v{0.23} {py3.7,py3.12,py3.13}-arq-latest @@ -140,14 +136,6 @@ envlist = # GCP {py3.7}-gcp - # GQL - {py3.7,py3.11}-gql-v{3.4} - {py3.7,py3.12,py3.13}-gql-latest - - # Graphene - {py3.7,py3.11}-graphene-v{3.3} - {py3.7,py3.12,py3.13}-graphene-latest - # gRPC {py3.7,py3.9}-grpc-v{1.39} {py3.7,py3.10}-grpc-v{1.49} @@ -276,11 +264,6 @@ envlist = {py3.7,py3.11}-sqlalchemy-v{2.0} {py3.7,py3.12,py3.13}-sqlalchemy-latest - # Strawberry - {py3.8,py3.11}-strawberry-v{0.209} - {py3.8,py3.11,py3.12}-strawberry-v{0.222} - {py3.8,py3.12,py3.13}-strawberry-latest - # Tornado {py3.8,py3.11,py3.12}-tornado-v{6.0} {py3.8,py3.11,py3.12}-tornado-v{6.2} @@ -362,13 +345,6 @@ deps = anthropic-v0.40: anthropic~=0.40.0 anthropic-latest: anthropic - # Ariadne - ariadne-v0.20: ariadne~=0.20.0 - ariadne-latest: ariadne - ariadne: fastapi - ariadne: flask - ariadne: httpx - # Arq arq-v0.23: arq~=0.23.0 arq-v0.23: pydantic<2 @@ -495,18 +471,6 @@ deps = flask-v3: Flask~=3.0 flask-latest: Flask - # GQL - gql-v{3.4}: gql[all]~=3.4.0 - gql-latest: gql[all] - - # Graphene - graphene: blinker - graphene: fastapi - graphene: flask - graphene: httpx - graphene-v{3.3}: graphene~=3.3.0 - graphene-latest: graphene - # gRPC grpc: protobuf grpc: mypy-protobuf @@ -731,14 +695,6 @@ deps = sqlalchemy-v2.0: sqlalchemy~=2.0.0 sqlalchemy-latest: sqlalchemy - # Strawberry - strawberry: fastapi - strawberry: flask - strawberry: httpx - strawberry-v0.209: strawberry-graphql[fastapi,flask]~=0.209.0 - strawberry-v0.222: strawberry-graphql[fastapi,flask]~=0.222.0 - strawberry-latest: strawberry-graphql[fastapi,flask] - # Tornado # Tornado <6.4.1 is incompatible with Pytest ≥8.2 # See https://github.com/tornadoweb/tornado/pull/3382. diff --git a/tox.ini b/tox.ini index c82d7d9159..4504c48c15 100644 --- a/tox.ini +++ b/tox.ini @@ -43,10 +43,6 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} {py3.7,py3.11,py3.12}-anthropic-latest - # Ariadne - {py3.8,py3.11}-ariadne-v{0.20} - {py3.8,py3.12,py3.13}-ariadne-latest - # Arq {py3.7,py3.11}-arq-v{0.23} {py3.7,py3.12,py3.13}-arq-latest @@ -140,14 +136,6 @@ envlist = # GCP {py3.7}-gcp - # GQL - {py3.7,py3.11}-gql-v{3.4} - {py3.7,py3.12,py3.13}-gql-latest - - # Graphene - {py3.7,py3.11}-graphene-v{3.3} - {py3.7,py3.12,py3.13}-graphene-latest - # gRPC {py3.7,py3.9}-grpc-v{1.39} {py3.7,py3.10}-grpc-v{1.49} @@ -276,11 +264,6 @@ envlist = {py3.7,py3.11}-sqlalchemy-v{2.0} {py3.7,py3.12,py3.13}-sqlalchemy-latest - # Strawberry - {py3.8,py3.11}-strawberry-v{0.209} - {py3.8,py3.11,py3.12}-strawberry-v{0.222} - {py3.8,py3.12,py3.13}-strawberry-latest - # Tornado {py3.8,py3.11,py3.12}-tornado-v{6.0} {py3.8,py3.11,py3.12}-tornado-v{6.2} @@ -305,6 +288,24 @@ envlist = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ GraphQL ~~~ + {py3.8,py3.10,py3.11}-ariadne-v0.20.1 + {py3.8,py3.11,py3.12}-ariadne-v0.22 + {py3.8,py3.11,py3.12}-ariadne-v0.24.0 + {py3.8,py3.11,py3.12}-ariadne-v0.25.2 + + {py3.6,py3.9,py3.10}-gql-v3.4.1 + {py3.7,py3.11,py3.12}-gql-v3.5.0 + + {py3.6,py3.9,py3.10}-graphene-v3.3 + {py3.8,py3.12,py3.13}-graphene-v3.4.3 + + {py3.8,py3.10,py3.11}-strawberry-v0.209.8 + {py3.8,py3.11,py3.12}-strawberry-v0.226.2 + {py3.8,py3.11,py3.12}-strawberry-v0.243.1 + {py3.9,py3.12,py3.13}-strawberry-v0.259.0 + + [testenv] deps = @@ -352,13 +353,6 @@ deps = anthropic-v0.40: anthropic~=0.40.0 anthropic-latest: anthropic - # Ariadne - ariadne-v0.20: ariadne~=0.20.0 - ariadne-latest: ariadne - ariadne: fastapi - ariadne: flask - ariadne: httpx - # Arq arq-v0.23: arq~=0.23.0 arq-v0.23: pydantic<2 @@ -485,18 +479,6 @@ deps = flask-v3: Flask~=3.0 flask-latest: Flask - # GQL - gql-v{3.4}: gql[all]~=3.4.0 - gql-latest: gql[all] - - # Graphene - graphene: blinker - graphene: fastapi - graphene: flask - graphene: httpx - graphene-v{3.3}: graphene~=3.3.0 - graphene-latest: graphene - # gRPC grpc: protobuf grpc: mypy-protobuf @@ -721,14 +703,6 @@ deps = sqlalchemy-v2.0: sqlalchemy~=2.0.0 sqlalchemy-latest: sqlalchemy - # Strawberry - strawberry: fastapi - strawberry: flask - strawberry: httpx - strawberry-v0.209: strawberry-graphql[fastapi,flask]~=0.209.0 - strawberry-v0.222: strawberry-graphql[fastapi,flask]~=0.222.0 - strawberry-latest: strawberry-graphql[fastapi,flask] - # Tornado # Tornado <6.4.1 is incompatible with Pytest ≥8.2 # See https://github.com/tornadoweb/tornado/pull/3382. @@ -754,6 +728,33 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ GraphQL ~~~ + ariadne-v0.20.1: ariadne==0.20.1 + ariadne-v0.22: ariadne==0.22 + ariadne-v0.24.0: ariadne==0.24.0 + ariadne-v0.25.2: ariadne==0.25.2 + ariadne: fastapi + ariadne: flask + ariadne: httpx + + gql-v3.4.1: gql[all]==3.4.1 + gql-v3.5.0: gql[all]==3.5.0 + + graphene-v3.3: graphene==3.3 + graphene-v3.4.3: graphene==3.4.3 + graphene: blinker + graphene: fastapi + graphene: flask + graphene: httpx + py3.6-graphene: aiocontextvars + + strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 + strawberry-v0.226.2: strawberry-graphql[fastapi,flask]==0.226.2 + strawberry-v0.243.1: strawberry-graphql[fastapi,flask]==0.243.1 + strawberry-v0.259.0: strawberry-graphql[fastapi,flask]==0.259.0 + strawberry: httpx + + setenv = PYTHONDONTWRITEBYTECODE=1 From 73a61c686472c4e590a1972a14b63f2ed3fda2e2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 12 Feb 2025 17:04:52 +0100 Subject: [PATCH 130/591] Update changelog with `profile_session_sample_rate` (#4046) --- CHANGELOG.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 0229aac66f..5da35ac676 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -17,6 +17,8 @@ - Set level based on status code for HTTP client breadcrumbs (#4004) by @sentrivana - Don't set transaction status to error on `sys.exit(0)` (#4025) by @sentrivana - Continuous profiling sample rate (#4002) by @Zylphrex + + Set `profile_session_sample_rate=1.0` in your `init()` to collect continuous profiles for 100% of profile sessions. See https://docs.sentry.io/platforms/python/profiling/#enable-continuous-profiling for more information. - Track and report spans that were dropped (#4005) by @constantinius - Change continuous profile buffer size (#3987) by @Zylphrex - Handle `MultiPartParserError` to avoid internal sentry crash (#4001) by @orhanhenrik @@ -40,7 +42,7 @@ ## 2.20.0 - **New integration:** Add [Typer](https://typer.tiangolo.com/) integration (#3869) by @patrick91 - + For more information, see the documentation for the [TyperIntegration](https://docs.sentry.io/platforms/python/integrations/typer/). - **New integration:** Add [Unleash](https://www.getunleash.io/) feature flagging integration (#3888) by @aliu39 @@ -122,7 +124,7 @@ ### Various fixes & improvements - **New integration:** Add [LaunchDarkly](https://launchdarkly.com/) integration (#3648) by @cmanallen - + For more information, see the documentation for the [LaunchDarklyIntegration](https://docs.sentry.io/platforms/python/integrations/launchdarkly/). - **New integration:** Add [OpenFeature](https://openfeature.dev/) feature flagging integration (#3648) by @cmanallen From 7a1c0103f3d023a1a3acd480324af86f8c783b1d Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Wed, 12 Feb 2025 09:55:52 -0800 Subject: [PATCH 131/591] feat(flags): add Statsig integration (#4022) New integration for tracking [Statsig](https://docs.statsig.com/server/pythonSDK) ([pypi](https://pypi.org/project/statsig/)) flag evaluations, specifically the checkGate method which is used for boolean release flags. Unlike JS, there's no support for event callbacks for Statsig's server SDKs. Instead we wrap the module-level `check_gate` function. Ref https://develop.sentry.dev/sdk/expected-features/#feature-flags Ref - https://github.com/getsentry/team-replay/issues/538 --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-flags.yml | 8 + requirements-linting.txt | 3 +- scripts/populate_tox/populate_tox.py | 1 + .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/__init__.py | 1 + sentry_sdk/integrations/statsig.py | 37 ++++ sentry_sdk/integrations/unleash.py | 2 +- setup.py | 1 + tests/integrations/statsig/__init__.py | 3 + tests/integrations/statsig/test_statsig.py | 183 ++++++++++++++++++ tox.ini | 10 + 11 files changed, 248 insertions(+), 2 deletions(-) create mode 100644 sentry_sdk/integrations/statsig.py create mode 100644 tests/integrations/statsig/__init__.py create mode 100644 tests/integrations/statsig/test_statsig.py diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 096da8d672..f56e1a082a 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -55,6 +55,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature-latest" + - name: Test statsig latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-statsig-latest" - name: Test unleash latest run: | set -x # print commands that are executed @@ -119,6 +123,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openfeature" + - name: Test statsig pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-statsig" - name: Test unleash pinned run: | set -x # print commands that are executed diff --git a/requirements-linting.txt b/requirements-linting.txt index 014e177793..4255685b5e 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -15,8 +15,9 @@ flake8-bugbear pep8-naming pre-commit # local linting httpcore -openfeature-sdk launchdarkly-server-sdk +openfeature-sdk +statsig UnleashClient typer strawberry-graphql diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 60770d5832..801aaeccb2 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -100,6 +100,7 @@ "spark", "starlette", "starlite", + "statsig", "sqlalchemy", "tornado", "trytond", diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 43307c3093..5218b0675f 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -87,6 +87,7 @@ "Flags": [ "launchdarkly", "openfeature", + "statsig", "unleash", ], "Gevent": [ diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 45235a41c4..f2b02e8b19 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -151,6 +151,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "sanic": (0, 8), "sqlalchemy": (1, 2), "starlite": (1, 48), + "statsig": (0, 55, 3), "strawberry": (0, 209, 5), "tornado": (6, 0), "typer": (0, 15), diff --git a/sentry_sdk/integrations/statsig.py b/sentry_sdk/integrations/statsig.py new file mode 100644 index 0000000000..1d84eb8aa2 --- /dev/null +++ b/sentry_sdk/integrations/statsig.py @@ -0,0 +1,37 @@ +from functools import wraps +from typing import Any, TYPE_CHECKING + +from sentry_sdk.feature_flags import add_feature_flag +from sentry_sdk.integrations import Integration, DidNotEnable, _check_minimum_version +from sentry_sdk.utils import parse_version + +try: + from statsig import statsig as statsig_module + from statsig.version import __version__ as STATSIG_VERSION +except ImportError: + raise DidNotEnable("statsig is not installed") + +if TYPE_CHECKING: + from statsig.statsig_user import StatsigUser + + +class StatsigIntegration(Integration): + identifier = "statsig" + + @staticmethod + def setup_once(): + # type: () -> None + version = parse_version(STATSIG_VERSION) + _check_minimum_version(StatsigIntegration, version, "statsig") + + # Wrap and patch evaluation method(s) in the statsig module + old_check_gate = statsig_module.check_gate + + @wraps(old_check_gate) + def sentry_check_gate(user, gate, *args, **kwargs): + # type: (StatsigUser, str, *Any, **Any) -> Any + enabled = old_check_gate(user, gate, *args, **kwargs) + add_feature_flag(gate, enabled) + return enabled + + statsig_module.check_gate = sentry_check_gate diff --git a/sentry_sdk/integrations/unleash.py b/sentry_sdk/integrations/unleash.py index c7108394d0..873f36c68b 100644 --- a/sentry_sdk/integrations/unleash.py +++ b/sentry_sdk/integrations/unleash.py @@ -16,7 +16,7 @@ class UnleashIntegration(Integration): @staticmethod def setup_once(): # type: () -> None - # Wrap and patch evaluation methods (instance methods) + # Wrap and patch evaluation methods (class methods) old_is_enabled = UnleashClient.is_enabled @wraps(old_is_enabled) diff --git a/setup.py b/setup.py index 760ce2d60f..21793220d4 100644 --- a/setup.py +++ b/setup.py @@ -79,6 +79,7 @@ def get_file_text(file_name): "sqlalchemy": ["sqlalchemy>=1.2"], "starlette": ["starlette>=0.19.1"], "starlite": ["starlite>=1.48"], + "statsig": ["statsig>=0.55.3"], "tornado": ["tornado>=6"], "unleash": ["UnleashClient>=6.0.1"], }, diff --git a/tests/integrations/statsig/__init__.py b/tests/integrations/statsig/__init__.py new file mode 100644 index 0000000000..6abc08235b --- /dev/null +++ b/tests/integrations/statsig/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("statsig") diff --git a/tests/integrations/statsig/test_statsig.py b/tests/integrations/statsig/test_statsig.py new file mode 100644 index 0000000000..c1666bde4d --- /dev/null +++ b/tests/integrations/statsig/test_statsig.py @@ -0,0 +1,183 @@ +import concurrent.futures as cf +import sys +from contextlib import contextmanager +from statsig import statsig +from statsig.statsig_user import StatsigUser +from random import random +from unittest.mock import Mock + +import pytest + +import sentry_sdk +from sentry_sdk.integrations.statsig import StatsigIntegration + + +@contextmanager +def mock_statsig(gate_dict): + old_check_gate = statsig.check_gate + + def mock_check_gate(user, gate, *args, **kwargs): + return gate_dict.get(gate, False) + + statsig.check_gate = Mock(side_effect=mock_check_gate) + + yield + + statsig.check_gate = old_check_gate + + +def test_check_gate(sentry_init, capture_events, uninstall_integration): + uninstall_integration(StatsigIntegration.identifier) + + with mock_statsig({"hello": True, "world": False}): + sentry_init(integrations=[StatsigIntegration()]) + events = capture_events() + user = StatsigUser(user_id="user-id") + + statsig.check_gate(user, "hello") + statsig.check_gate(user, "world") + statsig.check_gate(user, "other") # unknown gates default to False. + + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 1 + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + {"flag": "other", "result": False}, + ] + } + + +def test_check_gate_threaded(sentry_init, capture_events, uninstall_integration): + uninstall_integration(StatsigIntegration.identifier) + + with mock_statsig({"hello": True, "world": False}): + sentry_init(integrations=[StatsigIntegration()]) + events = capture_events() + user = StatsigUser(user_id="user-id") + + # Capture an eval before we split isolation scopes. + statsig.check_gate(user, "hello") + + def task(flag_key): + # Creates a new isolation scope for the thread. + # This means the evaluations in each task are captured separately. + with sentry_sdk.isolation_scope(): + statsig.check_gate(user, flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + with cf.ThreadPoolExecutor(max_workers=2) as pool: + pool.map(task, ["world", "other"]) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } + + +@pytest.mark.skipif(sys.version_info < (3, 7), reason="requires python3.7 or higher") +def test_check_gate_asyncio(sentry_init, capture_events, uninstall_integration): + asyncio = pytest.importorskip("asyncio") + uninstall_integration(StatsigIntegration.identifier) + + with mock_statsig({"hello": True, "world": False}): + sentry_init(integrations=[StatsigIntegration()]) + events = capture_events() + user = StatsigUser(user_id="user-id") + + # Capture an eval before we split isolation scopes. + statsig.check_gate(user, "hello") + + async def task(flag_key): + with sentry_sdk.isolation_scope(): + statsig.check_gate(user, flag_key) + # use a tag to identify to identify events later on + sentry_sdk.set_tag("task_id", flag_key) + sentry_sdk.capture_exception(Exception("something wrong!")) + + async def runner(): + return asyncio.gather(task("world"), task("other")) + + asyncio.run(runner()) + + # Capture error in original scope + sentry_sdk.set_tag("task_id", "0") + sentry_sdk.capture_exception(Exception("something wrong!")) + + assert len(events) == 3 + events.sort(key=lambda e: e["tags"]["task_id"]) + + assert events[0]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + ] + } + assert events[1]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "other", "result": False}, + ] + } + assert events[2]["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": True}, + {"flag": "world", "result": False}, + ] + } + + +def test_wraps_original(sentry_init, uninstall_integration): + uninstall_integration(StatsigIntegration.identifier) + flag_value = random() < 0.5 + + with mock_statsig( + {"test-flag": flag_value} + ): # patches check_gate with a Mock object. + mock_check_gate = statsig.check_gate + sentry_init(integrations=[StatsigIntegration()]) # wraps check_gate. + user = StatsigUser(user_id="user-id") + + res = statsig.check_gate(user, "test-flag", "extra-arg", kwarg=1) # type: ignore[arg-type] + + assert res == flag_value + assert mock_check_gate.call_args == ( # type: ignore[attr-defined] + (user, "test-flag", "extra-arg"), + {"kwarg": 1}, + ) + + +def test_wrapper_attributes(sentry_init, uninstall_integration): + uninstall_integration(StatsigIntegration.identifier) + original_check_gate = statsig.check_gate + sentry_init(integrations=[StatsigIntegration()]) + + # Methods have not lost their qualified names after decoration. + assert statsig.check_gate.__name__ == "check_gate" + assert statsig.check_gate.__qualname__ == original_check_gate.__qualname__ + + # Clean up + statsig.check_gate = original_check_gate diff --git a/tox.ini b/tox.ini index 4504c48c15..d5778a9fe1 100644 --- a/tox.ini +++ b/tox.ini @@ -259,6 +259,10 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar + # Statsig + {py3.8,py3.12,py3.13}-statsig-v0.55.3 + {py3.8,py3.12,py3.13}-statsig-latest + # SQL Alchemy {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} {py3.7,py3.11}-sqlalchemy-v{2.0} @@ -697,6 +701,11 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 + # Statsig + statsig: typing_extensions + statsig-v0.55.3: statsig~=0.55.3 + statsig-latest: statsig + # SQLAlchemy sqlalchemy-v1.2: sqlalchemy~=1.2.0 sqlalchemy-v1.4: sqlalchemy~=1.4.0 @@ -815,6 +824,7 @@ setenv = starlette: TESTPATH=tests/integrations/starlette starlite: TESTPATH=tests/integrations/starlite sqlalchemy: TESTPATH=tests/integrations/sqlalchemy + statsig: TESTPATH=tests/integrations/statsig strawberry: TESTPATH=tests/integrations/strawberry tornado: TESTPATH=tests/integrations/tornado trytond: TESTPATH=tests/integrations/trytond From 2b067e953470f93dadc726d331c7e91d9ec08f1b Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Wed, 12 Feb 2025 14:06:29 -0500 Subject: [PATCH 132/591] feat(profiling): Continuous profiling lifecycle (#4017) This introduces auto lifecycle setting for continuous profiling to only profile while there is an active transaction. This replaces the experimental auto start setting. --- sentry_sdk/consts.py | 2 + sentry_sdk/profiler/continuous_profiler.py | 172 +++++++++++++++--- sentry_sdk/profiler/transaction_profiler.py | 2 +- sentry_sdk/scope.py | 14 +- sentry_sdk/tracing.py | 12 +- tests/profiler/test_continuous_profiler.py | 188 ++++++++++++++++++-- 6 files changed, 347 insertions(+), 43 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 876556776c..df2c2b52a0 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -38,6 +38,7 @@ class CompressionAlgo(Enum): from typing import Any from typing import Sequence from typing import Tuple + from typing_extensions import Literal from typing_extensions import TypedDict from sentry_sdk._types import ( @@ -528,6 +529,7 @@ def __init__( profiles_sample_rate=None, # type: Optional[float] profiles_sampler=None, # type: Optional[TracesSampler] profiler_mode=None, # type: Optional[ProfilerMode] + profile_lifecycle="manual", # type: Literal["manual", "trace"] profile_session_sample_rate=None, # type: Optional[float] auto_enabling_integrations=True, # type: bool disabled_integrations=None, # type: Optional[Sequence[sentry_sdk.integrations.Integration]] diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index b07fbec998..1619925bd2 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -5,6 +5,7 @@ import threading import time import uuid +from collections import deque from datetime import datetime, timezone from sentry_sdk.consts import VERSION @@ -27,9 +28,11 @@ if TYPE_CHECKING: from typing import Any from typing import Callable + from typing import Deque from typing import Dict from typing import List from typing import Optional + from typing import Set from typing import Type from typing import Union from typing_extensions import TypedDict @@ -120,6 +123,9 @@ def setup_continuous_profiler(options, sdk_info, capture_func): def try_autostart_continuous_profiler(): # type: () -> None + + # TODO: deprecate this as it'll be replaced by the auto lifecycle option + if _scheduler is None: return @@ -129,6 +135,14 @@ def try_autostart_continuous_profiler(): _scheduler.manual_start() +def try_profile_lifecycle_trace_start(): + # type: () -> Union[ContinuousProfile, None] + if _scheduler is None: + return None + + return _scheduler.auto_start() + + def start_profiler(): # type: () -> None if _scheduler is None: @@ -170,6 +184,14 @@ def determine_profile_session_sampling_decision(sample_rate): return random.random() < float(sample_rate) +class ContinuousProfile: + active: bool = True + + def stop(self): + # type: () -> None + self.active = False + + class ContinuousScheduler: mode = "unknown" # type: ContinuousProfilerMode @@ -179,16 +201,21 @@ def __init__(self, frequency, options, sdk_info, capture_func): self.options = options self.sdk_info = sdk_info self.capture_func = capture_func + + self.lifecycle = self.options.get("profile_lifecycle") + profile_session_sample_rate = self.options.get("profile_session_sample_rate") + self.sampled = determine_profile_session_sampling_decision( + profile_session_sample_rate + ) + self.sampler = self.make_sampler() self.buffer = None # type: Optional[ProfileBuffer] self.pid = None # type: Optional[int] self.running = False - profile_session_sample_rate = self.options.get("profile_session_sample_rate") - self.sampled = determine_profile_session_sampling_decision( - profile_session_sample_rate - ) + self.new_profiles = deque(maxlen=128) # type: Deque[ContinuousProfile] + self.active_profiles = set() # type: Set[ContinuousProfile] def is_auto_start_enabled(self): # type: () -> bool @@ -207,15 +234,38 @@ def is_auto_start_enabled(self): return experiments.get("continuous_profiling_auto_start") + def auto_start(self): + # type: () -> Union[ContinuousProfile, None] + if not self.sampled: + return None + + if self.lifecycle != "trace": + return None + + logger.debug("[Profiling] Auto starting profiler") + + profile = ContinuousProfile() + + self.new_profiles.append(profile) + self.ensure_running() + + return profile + def manual_start(self): # type: () -> None if not self.sampled: return + if self.lifecycle != "manual": + return + self.ensure_running() def manual_stop(self): # type: () -> None + if self.lifecycle != "manual": + return + self.teardown() def ensure_running(self): @@ -249,28 +299,97 @@ def make_sampler(self): cache = LRUCache(max_size=256) - def _sample_stack(*args, **kwargs): - # type: (*Any, **Any) -> None - """ - Take a sample of the stack on all the threads in the process. - This should be called at a regular interval to collect samples. - """ - - ts = now() - - try: - sample = [ - (str(tid), extract_stack(frame, cache, cwd)) - for tid, frame in sys._current_frames().items() - ] - except AttributeError: - # For some reason, the frame we get doesn't have certain attributes. - # When this happens, we abandon the current sample as it's bad. - capture_internal_exception(sys.exc_info()) - return - - if self.buffer is not None: - self.buffer.write(ts, sample) + if self.lifecycle == "trace": + + def _sample_stack(*args, **kwargs): + # type: (*Any, **Any) -> None + """ + Take a sample of the stack on all the threads in the process. + This should be called at a regular interval to collect samples. + """ + + # no profiles taking place, so we can stop early + if not self.new_profiles and not self.active_profiles: + self.running = False + return + + # This is the number of profiles we want to pop off. + # It's possible another thread adds a new profile to + # the list and we spend longer than we want inside + # the loop below. + # + # Also make sure to set this value before extracting + # frames so we do not write to any new profiles that + # were started after this point. + new_profiles = len(self.new_profiles) + + ts = now() + + try: + sample = [ + (str(tid), extract_stack(frame, cache, cwd)) + for tid, frame in sys._current_frames().items() + ] + except AttributeError: + # For some reason, the frame we get doesn't have certain attributes. + # When this happens, we abandon the current sample as it's bad. + capture_internal_exception(sys.exc_info()) + return + + # Move the new profiles into the active_profiles set. + # + # We cannot directly add the to active_profiles set + # in `start_profiling` because it is called from other + # threads which can cause a RuntimeError when it the + # set sizes changes during iteration without a lock. + # + # We also want to avoid using a lock here so threads + # that are starting profiles are not blocked until it + # can acquire the lock. + for _ in range(new_profiles): + self.active_profiles.add(self.new_profiles.popleft()) + inactive_profiles = [] + + for profile in self.active_profiles: + if profile.active: + pass + else: + # If a profile is marked inactive, we buffer it + # to `inactive_profiles` so it can be removed. + # We cannot remove it here as it would result + # in a RuntimeError. + inactive_profiles.append(profile) + + for profile in inactive_profiles: + self.active_profiles.remove(profile) + + if self.buffer is not None: + self.buffer.write(ts, sample) + + else: + + def _sample_stack(*args, **kwargs): + # type: (*Any, **Any) -> None + """ + Take a sample of the stack on all the threads in the process. + This should be called at a regular interval to collect samples. + """ + + ts = now() + + try: + sample = [ + (str(tid), extract_stack(frame, cache, cwd)) + for tid, frame in sys._current_frames().items() + ] + except AttributeError: + # For some reason, the frame we get doesn't have certain attributes. + # When this happens, we abandon the current sample as it's bad. + capture_internal_exception(sys.exc_info()) + return + + if self.buffer is not None: + self.buffer.write(ts, sample) return _sample_stack @@ -294,6 +413,7 @@ def run(self): if self.buffer is not None: self.buffer.flush() + self.buffer = None class ThreadContinuousScheduler(ContinuousScheduler): diff --git a/sentry_sdk/profiler/transaction_profiler.py b/sentry_sdk/profiler/transaction_profiler.py index f579c441fa..3743b7c905 100644 --- a/sentry_sdk/profiler/transaction_profiler.py +++ b/sentry_sdk/profiler/transaction_profiler.py @@ -644,7 +644,7 @@ def _sample_stack(*args, **kwargs): if profile.active: profile.write(now, sample) else: - # If a thread is marked inactive, we buffer it + # If a profile is marked inactive, we buffer it # to `inactive_profiles` so it can be removed. # We cannot remove it here as it would result # in a RuntimeError. diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 53191c45da..4e3bb87489 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -12,7 +12,11 @@ from sentry_sdk.attachments import Attachment from sentry_sdk.consts import DEFAULT_MAX_BREADCRUMBS, FALSE_VALUES, INSTRUMENTER from sentry_sdk.feature_flags import FlagBuffer, DEFAULT_FLAG_CAPACITY -from sentry_sdk.profiler.continuous_profiler import try_autostart_continuous_profiler +from sentry_sdk.profiler.continuous_profiler import ( + get_profiler_id, + try_autostart_continuous_profiler, + try_profile_lifecycle_trace_start, +) from sentry_sdk.profiler.transaction_profiler import Profile from sentry_sdk.session import Session from sentry_sdk.tracing_utils import ( @@ -1063,6 +1067,14 @@ def start_transaction( transaction._profile = profile + transaction._continuous_profile = try_profile_lifecycle_trace_start() + + # Typically, the profiler is set when the transaction is created. But when + # using the auto lifecycle, the profiler isn't running when the first + # transaction is started. So make sure we update the profiler id on it. + if transaction._continuous_profile is not None: + transaction.set_profiler_id(get_profiler_id()) + # we don't bother to keep spans if we already know we're not going to # send the transaction max_spans = (client.options["_experiments"].get("max_spans")) or 1000 diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 2692944cf9..9d50d38963 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -34,7 +34,8 @@ P = ParamSpec("P") R = TypeVar("R") - import sentry_sdk.profiler + from sentry_sdk.profiler.continuous_profiler import ContinuousProfile + from sentry_sdk.profiler.transaction_profiler import Profile from sentry_sdk._types import ( Event, MeasurementUnit, @@ -767,6 +768,7 @@ class Transaction(Span): "_measurements", "_contexts", "_profile", + "_continuous_profile", "_baggage", ) @@ -788,9 +790,8 @@ def __init__( # type: ignore[misc] self.parent_sampled = parent_sampled self._measurements = {} # type: Dict[str, MeasurementValue] self._contexts = {} # type: Dict[str, Any] - self._profile = ( - None - ) # type: Optional[sentry_sdk.profiler.transaction_profiler.Profile] + self._profile = None # type: Optional[Profile] + self._continuous_profile = None # type: Optional[ContinuousProfile] self._baggage = baggage def __repr__(self): @@ -843,6 +844,9 @@ def __exit__(self, ty, value, tb): if self._profile is not None: self._profile.__exit__(ty, value, tb) + if self._continuous_profile is not None: + self._continuous_profile.stop() + super().__exit__(ty, value, tb) @property diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 6f4893e59d..331080df83 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -8,6 +8,7 @@ import sentry_sdk from sentry_sdk.consts import VERSION from sentry_sdk.profiler.continuous_profiler import ( + get_profiler_id, setup_continuous_profiler, start_profiler, stop_profiler, @@ -24,9 +25,12 @@ def get_client_options(use_top_level_profiler_mode): - def client_options(mode=None, auto_start=None, profile_session_sample_rate=1.0): + def client_options( + mode=None, auto_start=None, profile_session_sample_rate=1.0, lifecycle="manual" + ): if use_top_level_profiler_mode: return { + "profile_lifecycle": lifecycle, "profiler_mode": mode, "profile_session_sample_rate": profile_session_sample_rate, "_experiments": { @@ -34,6 +38,7 @@ def client_options(mode=None, auto_start=None, profile_session_sample_rate=1.0): }, } return { + "profile_lifecycle": lifecycle, "profile_session_sample_rate": profile_session_sample_rate, "_experiments": { "continuous_profiling_auto_start": auto_start, @@ -121,14 +126,17 @@ def test_continuous_profiler_setup_twice(mode, make_options, teardown_profiling) ) -def assert_single_transaction_with_profile_chunks(envelopes, thread): +def assert_single_transaction_with_profile_chunks( + envelopes, thread, max_chunks, transactions=1 +): items = defaultdict(list) for envelope in envelopes: for item in envelope.items: items[item.type].append(item) - assert len(items["transaction"]) == 1 + assert len(items["transaction"]) == transactions assert len(items["profile_chunk"]) > 0 + assert len(items["profile_chunk"]) <= max_chunks transaction = items["transaction"][0].payload.json @@ -163,6 +171,7 @@ def assert_single_transaction_with_profile_chunks(envelopes, thread): for profile_chunk_item in items["profile_chunk"]: profile_chunk = profile_chunk_item.payload.json + del profile_chunk["profile"] # make the diff easier to read assert profile_chunk == ApproxDict( { "client_sdk": { @@ -224,9 +233,9 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.1) + time.sleep(0.05) - assert_single_transaction_with_profile_chunks(envelopes, thread) + assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) for _ in range(3): stop_profiler() @@ -235,7 +244,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.1) + time.sleep(0.05) assert_single_transaction_without_profile_chunks(envelopes) @@ -245,9 +254,9 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): - time.sleep(0.1) + time.sleep(0.05) - assert_single_transaction_with_profile_chunks(envelopes, thread) + assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) @pytest.mark.parametrize( @@ -272,7 +281,9 @@ def test_continuous_profiler_manual_start_and_stop_sampled( make_options, teardown_profiling, ): - options = make_options(mode=mode) + options = make_options( + mode=mode, profile_session_sample_rate=1.0, lifecycle="manual" + ) sentry_init( traces_sample_rate=1.0, **options, @@ -291,7 +302,7 @@ def test_continuous_profiler_manual_start_and_stop_sampled( with sentry_sdk.start_span(op="op"): time.sleep(0.05) - assert_single_transaction_with_profile_chunks(envelopes, thread) + assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) stop_profiler() @@ -325,7 +336,9 @@ def test_continuous_profiler_manual_start_and_stop_unsampled( make_options, teardown_profiling, ): - options = make_options(mode=mode, profile_session_sample_rate=0.0) + options = make_options( + mode=mode, profile_session_sample_rate=0.0, lifecycle="manual" + ) sentry_init( traces_sample_rate=1.0, **options, @@ -342,3 +355,156 @@ def test_continuous_profiler_manual_start_and_stop_unsampled( assert_single_transaction_without_profile_chunks(envelopes) stop_profiler() + + +@pytest.mark.parametrize( + "mode", + [ + pytest.param("thread"), + pytest.param("gevent", marks=requires_gevent), + ], +) +@pytest.mark.parametrize( + "make_options", + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], +) +@mock.patch("sentry_sdk.profiler.continuous_profiler.DEFAULT_SAMPLING_FREQUENCY", 21) +def test_continuous_profiler_auto_start_and_stop_sampled( + sentry_init, + capture_envelopes, + mode, + make_options, + teardown_profiling, +): + options = make_options( + mode=mode, profile_session_sample_rate=1.0, lifecycle="trace" + ) + sentry_init( + traces_sample_rate=1.0, + **options, + ) + + envelopes = capture_envelopes() + + thread = threading.current_thread() + + for _ in range(3): + envelopes.clear() + + with sentry_sdk.start_transaction(name="profiling 1"): + assert get_profiler_id() is not None, "profiler should be running" + with sentry_sdk.start_span(op="op"): + time.sleep(0.03) + assert get_profiler_id() is not None, "profiler should be running" + + # the profiler takes a while to stop so if we start a transaction + # immediately, it'll be part of the same chunk + assert get_profiler_id() is not None, "profiler should be running" + + with sentry_sdk.start_transaction(name="profiling 2"): + assert get_profiler_id() is not None, "profiler should be running" + with sentry_sdk.start_span(op="op"): + time.sleep(0.03) + assert get_profiler_id() is not None, "profiler should be running" + + # wait at least 1 cycle for the profiler to stop + time.sleep(0.2) + assert get_profiler_id() is None, "profiler should not be running" + + assert_single_transaction_with_profile_chunks( + envelopes, thread, max_chunks=1, transactions=2 + ) + + +@pytest.mark.parametrize( + "mode", + [ + pytest.param("thread"), + pytest.param("gevent", marks=requires_gevent), + ], +) +@pytest.mark.parametrize( + "make_options", + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], +) +@mock.patch("sentry_sdk.profiler.continuous_profiler.PROFILE_BUFFER_SECONDS", 0.01) +def test_continuous_profiler_auto_start_and_stop_unsampled( + sentry_init, + capture_envelopes, + mode, + make_options, + teardown_profiling, +): + options = make_options( + mode=mode, profile_session_sample_rate=0.0, lifecycle="trace" + ) + sentry_init( + traces_sample_rate=1.0, + **options, + ) + + envelopes = capture_envelopes() + + for _ in range(3): + envelopes.clear() + + with sentry_sdk.start_transaction(name="profiling"): + assert get_profiler_id() is None, "profiler should not be running" + with sentry_sdk.start_span(op="op"): + time.sleep(0.05) + assert get_profiler_id() is None, "profiler should not be running" + + assert get_profiler_id() is None, "profiler should not be running" + assert_single_transaction_without_profile_chunks(envelopes) + + +@pytest.mark.parametrize( + ["mode", "class_name"], + [ + pytest.param("thread", "ThreadContinuousScheduler"), + pytest.param( + "gevent", + "GeventContinuousScheduler", + marks=requires_gevent, + ), + ], +) +@pytest.mark.parametrize( + "make_options", + [ + pytest.param(get_client_options(True), id="non-experiment"), + pytest.param(get_client_options(False), id="experiment"), + ], +) +def test_continuous_profiler_manual_start_and_stop_noop_when_using_trace_lifecyle( + sentry_init, + mode, + class_name, + make_options, + teardown_profiling, +): + options = make_options( + mode=mode, profile_session_sample_rate=0.0, lifecycle="trace" + ) + sentry_init( + traces_sample_rate=1.0, + **options, + ) + + with mock.patch( + f"sentry_sdk.profiler.continuous_profiler.{class_name}.ensure_running" + ) as mock_ensure_running: + start_profiler() + mock_ensure_running.assert_not_called() + + with mock.patch( + f"sentry_sdk.profiler.continuous_profiler.{class_name}.teardown" + ) as mock_teardown: + stop_profiler() + mock_teardown.assert_not_called() From 7c9f402f1ca2824405b5c72609d7865c25a5d05a Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Thu, 13 Feb 2025 08:52:49 -0500 Subject: [PATCH 133/591] tests(profiling): Reduce continuous profiling test flakiness (#4052) Not too sure what the problem is exactly but my suspicion is that the profiler runs in a separate thread and needs time to flush the chunk, the test wasn't waiting long enough. --- tests/profiler/test_continuous_profiler.py | 32 ++++++++++++++-------- 1 file changed, 21 insertions(+), 11 deletions(-) diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 331080df83..525616c9a8 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -127,7 +127,7 @@ def test_continuous_profiler_setup_twice(mode, make_options, teardown_profiling) def assert_single_transaction_with_profile_chunks( - envelopes, thread, max_chunks, transactions=1 + envelopes, thread, max_chunks=None, transactions=1 ): items = defaultdict(list) for envelope in envelopes: @@ -136,7 +136,8 @@ def assert_single_transaction_with_profile_chunks( assert len(items["transaction"]) == transactions assert len(items["profile_chunk"]) > 0 - assert len(items["profile_chunk"]) <= max_chunks + if max_chunks is not None: + assert len(items["profile_chunk"]) <= max_chunks transaction = items["transaction"][0].payload.json @@ -235,7 +236,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_span(op="op"): time.sleep(0.05) - assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) + assert_single_transaction_with_profile_chunks(envelopes, thread) for _ in range(3): stop_profiler() @@ -256,7 +257,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( with sentry_sdk.start_span(op="op"): time.sleep(0.05) - assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) + assert_single_transaction_with_profile_chunks(envelopes, thread) @pytest.mark.parametrize( @@ -299,18 +300,27 @@ def test_continuous_profiler_manual_start_and_stop_sampled( envelopes.clear() with sentry_sdk.start_transaction(name="profiling"): + assert get_profiler_id() is not None, "profiler should be running" with sentry_sdk.start_span(op="op"): - time.sleep(0.05) + time.sleep(0.1) + assert get_profiler_id() is not None, "profiler should be running" - assert_single_transaction_with_profile_chunks(envelopes, thread, max_chunks=10) + assert_single_transaction_with_profile_chunks(envelopes, thread) + + assert get_profiler_id() is not None, "profiler should be running" stop_profiler() + # the profiler stops immediately in manual mode + assert get_profiler_id() is None, "profiler should not be running" + envelopes.clear() with sentry_sdk.start_transaction(name="profiling"): + assert get_profiler_id() is None, "profiler should not be running" with sentry_sdk.start_span(op="op"): - time.sleep(0.05) + time.sleep(0.1) + assert get_profiler_id() is None, "profiler should not be running" assert_single_transaction_without_profile_chunks(envelopes) @@ -397,17 +407,17 @@ def test_continuous_profiler_auto_start_and_stop_sampled( with sentry_sdk.start_transaction(name="profiling 1"): assert get_profiler_id() is not None, "profiler should be running" with sentry_sdk.start_span(op="op"): - time.sleep(0.03) + time.sleep(0.1) assert get_profiler_id() is not None, "profiler should be running" - # the profiler takes a while to stop so if we start a transaction - # immediately, it'll be part of the same chunk + # the profiler takes a while to stop in auto mode so if we start + # a transaction immediately, it'll be part of the same chunk assert get_profiler_id() is not None, "profiler should be running" with sentry_sdk.start_transaction(name="profiling 2"): assert get_profiler_id() is not None, "profiler should be running" with sentry_sdk.start_span(op="op"): - time.sleep(0.03) + time.sleep(0.1) assert get_profiler_id() is not None, "profiler should be running" # wait at least 1 cycle for the profiler to stop From c2a3c08e7bc913aae7dbde74b6cb16c3d0165c25 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 15:19:14 +0100 Subject: [PATCH 134/591] Fix clickhouse test (#4053) We're not interested in random breadcrumbs from random logs like ``` + { + 'category': 'tzlocal', + 'data': {}, + 'level': 'warning', + 'message': '/etc/timezone is deprecated on Debian, and no longer reliable. ' + 'Ignoring.', + 'type': 'log', }, ``` --- .../clickhouse_driver/test_clickhouse_driver.py | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/tests/integrations/clickhouse_driver/test_clickhouse_driver.py b/tests/integrations/clickhouse_driver/test_clickhouse_driver.py index 3b07a82f03..0675ad9ff5 100644 --- a/tests/integrations/clickhouse_driver/test_clickhouse_driver.py +++ b/tests/integrations/clickhouse_driver/test_clickhouse_driver.py @@ -109,7 +109,13 @@ def test_clickhouse_client_breadcrumbs(sentry_init, capture_events) -> None: for crumb in event["breadcrumbs"]["values"]: crumb.pop("timestamp", None) - assert event["breadcrumbs"]["values"] == expected_breadcrumbs + actual_query_breadcrumbs = [ + breadcrumb + for breadcrumb in event["breadcrumbs"]["values"] + if breadcrumb["category"] == "query" + ] + + assert actual_query_breadcrumbs == expected_breadcrumbs def test_clickhouse_client_breadcrumbs_with_pii(sentry_init, capture_events) -> None: From 5a66a04e36922f1ee2a722eec073366bf5d8d3d2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 15:30:46 +0100 Subject: [PATCH 135/591] tests: Remove toxgen cutoff, add statsig (#4048) - a new integration was added and added to tox.ini, but not the template - remove cutoff in favor of https://github.com/getsentry/sentry-python/issues/4047 --- scripts/populate_tox/populate_tox.py | 8 +------- scripts/populate_tox/tox.jinja | 12 +++++++++++- tox.ini | 6 +++--- 3 files changed, 15 insertions(+), 11 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 801aaeccb2..fe6d9d216a 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -1,5 +1,5 @@ """ -This script populates tox.ini automatically using release data from PYPI. +This script populates tox.ini automatically using release data from PyPI. """ import functools @@ -8,7 +8,6 @@ import time from bisect import bisect_left from collections import defaultdict -from datetime import datetime, timedelta from importlib.metadata import metadata from packaging.specifiers import SpecifierSet from packaging.version import Version @@ -27,9 +26,6 @@ from split_tox_gh_actions.split_tox_gh_actions import GROUPS -# Only consider package versions going back this far -CUTOFF = datetime.now() - timedelta(days=365 * 5) - TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" ENV = Environment( loader=FileSystemLoader(Path(__file__).resolve().parent), @@ -157,8 +153,6 @@ def _prefilter_releases(integration: str, releases: dict[str, dict]) -> list[Ver continue meta = data[0] - if datetime.fromisoformat(meta["upload_time"]) < CUTOFF: - continue if meta["yanked"]: continue diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index ad569b17a6..5d8a931aec 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -259,6 +259,10 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar + # Statsig + {py3.8,py3.12,py3.13}-statsig-v0.55.3 + {py3.8,py3.12,py3.13}-statsig-latest + # SQL Alchemy {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} {py3.7,py3.11}-sqlalchemy-v{2.0} @@ -689,6 +693,11 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 + # Statsig + statsig: typing_extensions + statsig-v0.55.3: statsig~=0.55.3 + statsig-latest: statsig + # SQLAlchemy sqlalchemy-v1.2: sqlalchemy~=1.2.0 sqlalchemy-v1.4: sqlalchemy~=1.4.0 @@ -794,9 +803,10 @@ setenv = rq: TESTPATH=tests/integrations/rq sanic: TESTPATH=tests/integrations/sanic spark: TESTPATH=tests/integrations/spark + sqlalchemy: TESTPATH=tests/integrations/sqlalchemy starlette: TESTPATH=tests/integrations/starlette starlite: TESTPATH=tests/integrations/starlite - sqlalchemy: TESTPATH=tests/integrations/sqlalchemy + statsig: TESTPATH=tests/integrations/statsig strawberry: TESTPATH=tests/integrations/strawberry tornado: TESTPATH=tests/integrations/tornado trytond: TESTPATH=tests/integrations/trytond diff --git a/tox.ini b/tox.ini index d5778a9fe1..4fb410568d 100644 --- a/tox.ini +++ b/tox.ini @@ -307,7 +307,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.226.2 {py3.8,py3.11,py3.12}-strawberry-v0.243.1 - {py3.9,py3.12,py3.13}-strawberry-v0.259.0 + {py3.9,py3.12,py3.13}-strawberry-v0.260.0 @@ -760,7 +760,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.226.2: strawberry-graphql[fastapi,flask]==0.226.2 strawberry-v0.243.1: strawberry-graphql[fastapi,flask]==0.243.1 - strawberry-v0.259.0: strawberry-graphql[fastapi,flask]==0.259.0 + strawberry-v0.260.0: strawberry-graphql[fastapi,flask]==0.260.0 strawberry: httpx @@ -821,9 +821,9 @@ setenv = rq: TESTPATH=tests/integrations/rq sanic: TESTPATH=tests/integrations/sanic spark: TESTPATH=tests/integrations/spark + sqlalchemy: TESTPATH=tests/integrations/sqlalchemy starlette: TESTPATH=tests/integrations/starlette starlite: TESTPATH=tests/integrations/starlite - sqlalchemy: TESTPATH=tests/integrations/sqlalchemy statsig: TESTPATH=tests/integrations/statsig strawberry: TESTPATH=tests/integrations/strawberry tornado: TESTPATH=tests/integrations/tornado From 5a5a1cf8549ddb2448d6c89d7ce474edfc0677b2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 15:41:34 +0100 Subject: [PATCH 136/591] tests: Generate Flags tox entries with toxgen script (#3974) - remove hardcoded entries for `openfeature`, `launchdarkly`, `statsig`, and `unleash` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new `tox.ini` Note that this effectively eliminates the `-latest` tests for the Flags group. The script doesn't generate any `-latest` tests since it always makes sure to add a pinned entry for the latest version. So in case all of the integrations in a single group are using the script, the whole `-latest` test category is removed. --- .github/workflows/test-integrations-flags.yml | 70 +------------------ scripts/populate_tox/config.py | 15 ++++ scripts/populate_tox/populate_tox.py | 4 -- scripts/populate_tox/tox.jinja | 33 --------- tox.ini | 66 ++++++++--------- 5 files changed, 47 insertions(+), 141 deletions(-) diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index f56e1a082a..ad344762ae 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -22,74 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-flags-latest: - name: Flags (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.8","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] - steps: - - uses: actions/checkout@v4.2.2 - - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test launchdarkly latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-launchdarkly-latest" - - name: Test openfeature latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature-latest" - - name: Test statsig latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-statsig-latest" - - name: Test unleash latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-unleash-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-flags-pinned: name: Flags (pinned) timeout-minutes: 30 @@ -97,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.12","3.13"] + python-version: ["3.7","3.8","3.9","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 8cdd36c05d..402ecf7a82 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -23,10 +23,25 @@ "py3.6": ["aiocontextvars"], }, }, + "launchdarkly": { + "package": "launchdarkly-server-sdk", + }, + "openfeature": { + "package": "openfeature-sdk", + }, + "statsig": { + "package": "statsig", + "deps": { + "*": ["typing_extensions"], + }, + }, "strawberry": { "package": "strawberry-graphql[fastapi,flask]", "deps": { "*": ["httpx"], }, }, + "unleash": { + "package": "UnleashClient", + }, } diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index fe6d9d216a..b8969b8987 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -77,12 +77,10 @@ "huggingface_hub", "langchain", "langchain_notiktoken", - "launchdarkly", "litestar", "loguru", "openai", "openai_notiktoken", - "openfeature", "pure_eval", "pymongo", "pyramid", @@ -96,12 +94,10 @@ "spark", "starlette", "starlite", - "statsig", "sqlalchemy", "tornado", "trytond", "typer", - "unleash", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 5d8a931aec..8086411f7b 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -163,10 +163,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # LaunchDarkly - {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 - {py3.8,py3.12,py3.13}-launchdarkly-latest - # Litestar {py3.8,py3.11}-litestar-v{2.0} {py3.8,py3.11,py3.12}-litestar-v{2.6} @@ -184,10 +180,6 @@ envlist = {py3.9,py3.11,py3.12}-openai-latest {py3.9,py3.11,py3.12}-openai-notiktoken - # OpenFeature - {py3.8,py3.12,py3.13}-openfeature-v0.7 - {py3.8,py3.12,py3.13}-openfeature-latest - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -259,10 +251,6 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar - # Statsig - {py3.8,py3.12,py3.13}-statsig-v0.55.3 - {py3.8,py3.12,py3.13}-statsig-latest - # SQL Alchemy {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} {py3.7,py3.11}-sqlalchemy-v{2.0} @@ -284,10 +272,6 @@ envlist = {py3.7,py3.12,py3.13}-typer-v{0.15} {py3.7,py3.12,py3.13}-typer-latest - # Unleash - {py3.8,py3.12,py3.13}-unleash-v6.0.1 - {py3.8,py3.12,py3.13}-unleash-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -557,18 +541,6 @@ deps = openai-latest: tiktoken~=0.6.0 openai-notiktoken: openai - # OpenFeature - openfeature-v0.7: openfeature-sdk~=0.7.1 - openfeature-latest: openfeature-sdk - - # LaunchDarkly - launchdarkly-v9.8.0: launchdarkly-server-sdk~=9.8.0 - launchdarkly-latest: launchdarkly-server-sdk - - # Unleash - unleash-v6.0.1: UnleashClient~=6.0.1 - unleash-latest: UnleashClient - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -693,11 +665,6 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 - # Statsig - statsig: typing_extensions - statsig-v0.55.3: statsig~=0.55.3 - statsig-latest: statsig - # SQLAlchemy sqlalchemy-v1.2: sqlalchemy~=1.2.0 sqlalchemy-v1.4: sqlalchemy~=1.4.0 diff --git a/tox.ini b/tox.ini index 4fb410568d..b8d1e6a74e 100644 --- a/tox.ini +++ b/tox.ini @@ -163,10 +163,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # LaunchDarkly - {py3.8,py3.12,py3.13}-launchdarkly-v9.8.0 - {py3.8,py3.12,py3.13}-launchdarkly-latest - # Litestar {py3.8,py3.11}-litestar-v{2.0} {py3.8,py3.11,py3.12}-litestar-v{2.6} @@ -184,10 +180,6 @@ envlist = {py3.9,py3.11,py3.12}-openai-latest {py3.9,py3.11,py3.12}-openai-notiktoken - # OpenFeature - {py3.8,py3.12,py3.13}-openfeature-v0.7 - {py3.8,py3.12,py3.13}-openfeature-latest - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -259,10 +251,6 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar - # Statsig - {py3.8,py3.12,py3.13}-statsig-v0.55.3 - {py3.8,py3.12,py3.13}-statsig-latest - # SQL Alchemy {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} {py3.7,py3.11}-sqlalchemy-v{2.0} @@ -284,14 +272,24 @@ envlist = {py3.7,py3.12,py3.13}-typer-v{0.15} {py3.7,py3.12,py3.13}-typer-latest - # Unleash - {py3.8,py3.12,py3.13}-unleash-v6.0.1 - {py3.8,py3.12,py3.13}-unleash-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ Flags ~~~ + {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 + {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 + + {py3.8,py3.12,py3.13}-openfeature-v0.7.5 + {py3.9,py3.12,py3.13}-openfeature-v0.8.0 + + {py3.7,py3.12,py3.13}-statsig-v0.55.3 + {py3.7,py3.12,py3.13}-statsig-v0.56.0 + + {py3.8,py3.12,py3.13}-unleash-v6.0.1 + {py3.8,py3.12,py3.13}-unleash-v6.1.0 + + # ~~~ GraphQL ~~~ {py3.8,py3.10,py3.11}-ariadne-v0.20.1 {py3.8,py3.11,py3.12}-ariadne-v0.22 @@ -307,7 +305,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.226.2 {py3.8,py3.11,py3.12}-strawberry-v0.243.1 - {py3.9,py3.12,py3.13}-strawberry-v0.260.0 + {py3.9,py3.12,py3.13}-strawberry-v0.260.1 @@ -565,18 +563,6 @@ deps = openai-latest: tiktoken~=0.6.0 openai-notiktoken: openai - # OpenFeature - openfeature-v0.7: openfeature-sdk~=0.7.1 - openfeature-latest: openfeature-sdk - - # LaunchDarkly - launchdarkly-v9.8.0: launchdarkly-server-sdk~=9.8.0 - launchdarkly-latest: launchdarkly-server-sdk - - # Unleash - unleash-v6.0.1: UnleashClient~=6.0.1 - unleash-latest: UnleashClient - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -701,11 +687,6 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 - # Statsig - statsig: typing_extensions - statsig-v0.55.3: statsig~=0.55.3 - statsig-latest: statsig - # SQLAlchemy sqlalchemy-v1.2: sqlalchemy~=1.2.0 sqlalchemy-v1.4: sqlalchemy~=1.4.0 @@ -737,6 +718,21 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ Flags ~~~ + launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 + launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 + + openfeature-v0.7.5: openfeature-sdk==0.7.5 + openfeature-v0.8.0: openfeature-sdk==0.8.0 + + statsig-v0.55.3: statsig==0.55.3 + statsig-v0.56.0: statsig==0.56.0 + statsig: typing_extensions + + unleash-v6.0.1: UnleashClient==6.0.1 + unleash-v6.1.0: UnleashClient==6.1.0 + + # ~~~ GraphQL ~~~ ariadne-v0.20.1: ariadne==0.20.1 ariadne-v0.22: ariadne==0.22 @@ -760,7 +756,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.226.2: strawberry-graphql[fastapi,flask]==0.226.2 strawberry-v0.243.1: strawberry-graphql[fastapi,flask]==0.243.1 - strawberry-v0.260.0: strawberry-graphql[fastapi,flask]==0.260.0 + strawberry-v0.260.1: strawberry-graphql[fastapi,flask]==0.260.1 strawberry: httpx From c6b599402732cb89f020eff3316a983ca308f0ab Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 16:33:10 +0100 Subject: [PATCH 137/591] Generate Misc tox entries via toxgen script (#3982) - remove hardcoded entries for `loguru`, `trytond`, and `typer` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new `tox.ini` Note that this effectively eliminates the `-latest` tests for the Misc group. The script doesn't generate any `-latest` tests since it always makes sure to add a pinned entry for the latest version. So in case all of the integrations in a single group are using the script, the whole `-latest` test category is removed. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-misc.yml | 76 -------------------- scripts/populate_tox/config.py | 13 ++++ scripts/populate_tox/populate_tox.py | 3 - scripts/populate_tox/tox.jinja | 32 --------- sentry_sdk/integrations/__init__.py | 1 + tox.ini | 61 ++++++++-------- 6 files changed, 43 insertions(+), 143 deletions(-) diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 82577c7be6..4e582c6c71 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -22,82 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-misc-latest: - name: Misc (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.6","3.7","3.8","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] - steps: - - uses: actions/checkout@v4.2.2 - - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test loguru latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-loguru-latest" - - name: Test opentelemetry latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-opentelemetry-latest" - - name: Test potel latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-potel-latest" - - name: Test pure_eval latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-pure_eval-latest" - - name: Test trytond latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-trytond-latest" - - name: Test typer latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-typer-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-misc-pinned: name: Misc (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 402ecf7a82..ac75753825 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -26,6 +26,9 @@ "launchdarkly": { "package": "launchdarkly-server-sdk", }, + "loguru": { + "package": "loguru", + }, "openfeature": { "package": "openfeature-sdk", }, @@ -41,6 +44,16 @@ "*": ["httpx"], }, }, + "trytond": { + "package": "trytond", + "deps": { + "*": ["werkzeug"], + "<=5.0": ["werkzeug<1.0"], + }, + }, + "typer": { + "package": "typer", + }, "unleash": { "package": "UnleashClient", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index b8969b8987..73c7277fd2 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -78,7 +78,6 @@ "langchain", "langchain_notiktoken", "litestar", - "loguru", "openai", "openai_notiktoken", "pure_eval", @@ -96,8 +95,6 @@ "starlite", "sqlalchemy", "tornado", - "trytond", - "typer", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 8086411f7b..06cd50c9a1 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -169,10 +169,6 @@ envlist = {py3.8,py3.11,py3.12}-litestar-v{2.12} {py3.8,py3.11,py3.12}-litestar-latest - # Loguru - {py3.6,py3.11,py3.12}-loguru-v{0.5} - {py3.6,py3.12,py3.13}-loguru-latest - # OpenAI {py3.9,py3.11,py3.12}-openai-v1.0 {py3.9,py3.11,py3.12}-openai-v1.22 @@ -261,17 +257,6 @@ envlist = {py3.8,py3.11,py3.12}-tornado-v{6.2} {py3.8,py3.11,py3.12}-tornado-latest - # Trytond - {py3.6}-trytond-v{4} - {py3.6,py3.8}-trytond-v{5} - {py3.6,py3.11}-trytond-v{6} - {py3.8,py3.11,py3.12}-trytond-v{7} - {py3.8,py3.12,py3.13}-trytond-latest - - # Typer - {py3.7,py3.12,py3.13}-typer-v{0.15} - {py3.7,py3.12,py3.13}-typer-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -523,10 +508,6 @@ deps = litestar-v2.12: litestar~=2.12.0 litestar-latest: litestar - # Loguru - loguru-v0.5: loguru~=0.5.0 - loguru-latest: loguru - # OpenAI openai: pytest-asyncio openai-v1.0: openai~=1.0.0 @@ -679,19 +660,6 @@ deps = tornado-v6.2: tornado~=6.2.0 tornado-latest: tornado - # Trytond - trytond: werkzeug - trytond-v4: werkzeug<1.0 - trytond-v4: trytond~=4.0 - trytond-v5: trytond~=5.0 - trytond-v6: trytond~=6.0 - trytond-v7: trytond~=7.0 - trytond-latest: trytond - - # Typer - typer-v0.15: typer~=0.15.0 - typer-latest: typer - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index f2b02e8b19..d803a0b169 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -142,6 +142,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "huggingface_hub": (0, 22), "langchain": (0, 0, 210), "launchdarkly": (9, 8, 0), + "loguru": (0, 7, 0), "openai": (1, 0, 0), "openfeature": (0, 7, 1), "quart": (0, 16, 0), diff --git a/tox.ini b/tox.ini index b8d1e6a74e..fa6240b094 100644 --- a/tox.ini +++ b/tox.ini @@ -169,10 +169,6 @@ envlist = {py3.8,py3.11,py3.12}-litestar-v{2.12} {py3.8,py3.11,py3.12}-litestar-latest - # Loguru - {py3.6,py3.11,py3.12}-loguru-v{0.5} - {py3.6,py3.12,py3.13}-loguru-latest - # OpenAI {py3.9,py3.11,py3.12}-openai-v1.0 {py3.9,py3.11,py3.12}-openai-v1.22 @@ -261,17 +257,6 @@ envlist = {py3.8,py3.11,py3.12}-tornado-v{6.2} {py3.8,py3.11,py3.12}-tornado-latest - # Trytond - {py3.6}-trytond-v{4} - {py3.6,py3.8}-trytond-v{5} - {py3.6,py3.11}-trytond-v{6} - {py3.8,py3.11,py3.12}-trytond-v{7} - {py3.8,py3.12,py3.13}-trytond-latest - - # Typer - {py3.7,py3.12,py3.13}-typer-v{0.15} - {py3.7,py3.12,py3.13}-typer-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -308,6 +293,19 @@ envlist = {py3.9,py3.12,py3.13}-strawberry-v0.260.1 + # ~~~ Misc ~~~ + {py3.6,py3.12,py3.13}-loguru-v0.7.3 + + {py3.6}-trytond-v4.6.9 + {py3.6}-trytond-v4.8.18 + {py3.6,py3.7,py3.8}-trytond-v5.8.16 + {py3.8,py3.10,py3.11}-trytond-v6.8.17 + {py3.8,py3.11,py3.12}-trytond-v7.0.9 + {py3.8,py3.11,py3.12}-trytond-v7.4.5 + + {py3.7,py3.11,py3.12}-typer-v0.15.1 + + [testenv] deps = @@ -545,10 +543,6 @@ deps = litestar-v2.12: litestar~=2.12.0 litestar-latest: litestar - # Loguru - loguru-v0.5: loguru~=0.5.0 - loguru-latest: loguru - # OpenAI openai: pytest-asyncio openai-v1.0: openai~=1.0.0 @@ -701,19 +695,6 @@ deps = tornado-v6.2: tornado~=6.2.0 tornado-latest: tornado - # Trytond - trytond: werkzeug - trytond-v4: werkzeug<1.0 - trytond-v4: trytond~=4.0 - trytond-v5: trytond~=5.0 - trytond-v6: trytond~=6.0 - trytond-v7: trytond~=7.0 - trytond-latest: trytond - - # Typer - typer-v0.15: typer~=0.15.0 - typer-latest: typer - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -760,6 +741,22 @@ deps = strawberry: httpx + # ~~~ Misc ~~~ + loguru-v0.7.3: loguru==0.7.3 + + trytond-v4.6.9: trytond==4.6.9 + trytond-v4.8.18: trytond==4.8.18 + trytond-v5.8.16: trytond==5.8.16 + trytond-v6.8.17: trytond==6.8.17 + trytond-v7.0.9: trytond==7.0.9 + trytond-v7.4.5: trytond==7.4.5 + trytond: werkzeug + trytond-v4.6.9: werkzeug<1.0 + trytond-v4.8.18: werkzeug<1.0 + + typer-v0.15.1: typer==0.15.1 + + setenv = PYTHONDONTWRITEBYTECODE=1 From feb642b5dc20ef848d7c2c6dde6c19c78188eac6 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 16:47:34 +0100 Subject: [PATCH 138/591] tests: Generate DB group by toxgen script (#3978) - remove hardcoded entries for `sqlalchemy`, `pymongo`, `redis_py_cluster_legacy`, `clickhouse_driver` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to port to the script, I'll do this step by step in follow-up PRs. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-dbs.yml | 2 +- scripts/populate_tox/config.py | 15 ++++ scripts/populate_tox/populate_tox.py | 4 -- scripts/populate_tox/tox.jinja | 43 ----------- tox.ini | 80 ++++++++++----------- 5 files changed, 53 insertions(+), 91 deletions(-) diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 0f5c37306a..d525e353ed 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -124,7 +124,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index ac75753825..df99681e77 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -13,6 +13,9 @@ }, "python": ">=3.8", }, + "clickhouse_driver": { + "package": "clickhouse-driver", + }, "gql": { "package": "gql[all]", }, @@ -32,6 +35,18 @@ "openfeature": { "package": "openfeature-sdk", }, + "pymongo": { + "package": "pymongo", + "deps": { + "*": ["mockupdb"], + }, + }, + "redis_py_cluster_legacy": { + "package": "redis-py-cluster", + }, + "sqlalchemy": { + "package": "sqlalchemy", + }, "statsig": { "package": "statsig", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 73c7277fd2..09c31923e6 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -61,7 +61,6 @@ "bottle", "celery", "chalice", - "clickhouse_driver", "cohere", "cloud_resource_context", "cohere", @@ -81,19 +80,16 @@ "openai", "openai_notiktoken", "pure_eval", - "pymongo", "pyramid", "quart", "ray", "redis", - "redis_py_cluster_legacy", "requests", "rq", "sanic", "spark", "starlette", "starlite", - "sqlalchemy", "tornado", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 06cd50c9a1..a7a7ff2615 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -85,10 +85,6 @@ envlist = {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest - # Clickhouse Driver - {py3.8,py3.11}-clickhouse_driver-v{0.2.0} - {py3.8,py3.12,py3.13}-clickhouse_driver-latest - # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context @@ -185,13 +181,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # PyMongo (Mongo DB) - {py3.6}-pymongo-v{3.1} - {py3.6,py3.9}-pymongo-v{3.12} - {py3.6,py3.11}-pymongo-v{4.0} - {py3.7,py3.11,py3.12}-pymongo-v{4.3,4.7} - {py3.7,py3.12,py3.13}-pymongo-latest - # Pyramid {py3.6,py3.11}-pyramid-v{1.6} {py3.6,py3.11,py3.12}-pyramid-v{1.10} @@ -213,10 +202,6 @@ envlist = {py3.7,py3.11,py3.12}-redis-v{5} {py3.7,py3.12,py3.13}-redis-latest - # Redis Cluster - {py3.6,py3.8}-redis_py_cluster_legacy-v{1,2} - # no -latest, not developed anymore - # Requests {py3.6,py3.8,py3.12,py3.13}-requests @@ -247,11 +232,6 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar - # SQL Alchemy - {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} - {py3.7,py3.11}-sqlalchemy-v{2.0} - {py3.7,py3.12,py3.13}-sqlalchemy-latest - # Tornado {py3.8,py3.11,py3.12}-tornado-v{6.0} {py3.8,py3.11,py3.12}-tornado-v{6.2} @@ -373,10 +353,6 @@ deps = chalice-v1.16: chalice~=1.16.0 chalice-latest: chalice - # Clickhouse Driver - clickhouse_driver-v0.2.0: clickhouse_driver~=0.2.0 - clickhouse_driver-latest: clickhouse_driver - # Cohere cohere-v5: cohere~=5.3.3 cohere-latest: cohere @@ -531,15 +507,6 @@ deps = # pure_eval pure_eval: pure_eval - # PyMongo (MongoDB) - pymongo: mockupdb - pymongo-v3.1: pymongo~=3.1.0 - pymongo-v3.13: pymongo~=3.13.0 - pymongo-v4.0: pymongo~=4.0.0 - pymongo-v4.3: pymongo~=4.3.0 - pymongo-v4.7: pymongo~=4.7.0 - pymongo-latest: pymongo - # Pyramid pyramid: Werkzeug<2.1.0 pyramid-v1.6: pyramid~=1.6.0 @@ -574,10 +541,6 @@ deps = redis-v5: redis~=5.0 redis-latest: redis - # Redis Cluster - redis_py_cluster_legacy-v1: redis-py-cluster~=1.0 - redis_py_cluster_legacy-v2: redis-py-cluster~=2.0 - # Requests requests: requests>=2.0 @@ -646,12 +609,6 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 - # SQLAlchemy - sqlalchemy-v1.2: sqlalchemy~=1.2.0 - sqlalchemy-v1.4: sqlalchemy~=1.4.0 - sqlalchemy-v2.0: sqlalchemy~=2.0.0 - sqlalchemy-latest: sqlalchemy - # Tornado # Tornado <6.4.1 is incompatible with Pytest ≥8.2 # See https://github.com/tornadoweb/tornado/pull/3382. diff --git a/tox.ini b/tox.ini index fa6240b094..0487b3c595 100644 --- a/tox.ini +++ b/tox.ini @@ -85,10 +85,6 @@ envlist = {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest - # Clickhouse Driver - {py3.8,py3.11}-clickhouse_driver-v{0.2.0} - {py3.8,py3.12,py3.13}-clickhouse_driver-latest - # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context @@ -185,13 +181,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # PyMongo (Mongo DB) - {py3.6}-pymongo-v{3.1} - {py3.6,py3.9}-pymongo-v{3.12} - {py3.6,py3.11}-pymongo-v{4.0} - {py3.7,py3.11,py3.12}-pymongo-v{4.3,4.7} - {py3.7,py3.12,py3.13}-pymongo-latest - # Pyramid {py3.6,py3.11}-pyramid-v{1.6} {py3.6,py3.11,py3.12}-pyramid-v{1.10} @@ -213,10 +202,6 @@ envlist = {py3.7,py3.11,py3.12}-redis-v{5} {py3.7,py3.12,py3.13}-redis-latest - # Redis Cluster - {py3.6,py3.8}-redis_py_cluster_legacy-v{1,2} - # no -latest, not developed anymore - # Requests {py3.6,py3.8,py3.12,py3.13}-requests @@ -247,11 +232,6 @@ envlist = {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar - # SQL Alchemy - {py3.6,py3.9}-sqlalchemy-v{1.2,1.4} - {py3.7,py3.11}-sqlalchemy-v{2.0} - {py3.7,py3.12,py3.13}-sqlalchemy-latest - # Tornado {py3.8,py3.11,py3.12}-tornado-v{6.0} {py3.8,py3.11,py3.12}-tornado-v{6.2} @@ -261,6 +241,24 @@ envlist = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ DBs ~~~ + {py3.7,py3.11,py3.12}-clickhouse_driver-v0.2.9 + + {py3.6}-pymongo-v3.5.1 + {py3.6,py3.10,py3.11}-pymongo-v3.13.0 + {py3.6,py3.9,py3.10}-pymongo-v4.0.2 + {py3.9,py3.12,py3.13}-pymongo-v4.11.1 + + {py3.6}-redis_py_cluster_legacy-v1.3.6 + {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 + {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 + + {py3.6,py3.7}-sqlalchemy-v1.3.9 + {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 + {py3.7,py3.10,py3.11}-sqlalchemy-v2.0.9 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.38 + + # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 @@ -408,10 +406,6 @@ deps = chalice-v1.16: chalice~=1.16.0 chalice-latest: chalice - # Clickhouse Driver - clickhouse_driver-v0.2.0: clickhouse_driver~=0.2.0 - clickhouse_driver-latest: clickhouse_driver - # Cohere cohere-v5: cohere~=5.3.3 cohere-latest: cohere @@ -566,15 +560,6 @@ deps = # pure_eval pure_eval: pure_eval - # PyMongo (MongoDB) - pymongo: mockupdb - pymongo-v3.1: pymongo~=3.1.0 - pymongo-v3.13: pymongo~=3.13.0 - pymongo-v4.0: pymongo~=4.0.0 - pymongo-v4.3: pymongo~=4.3.0 - pymongo-v4.7: pymongo~=4.7.0 - pymongo-latest: pymongo - # Pyramid pyramid: Werkzeug<2.1.0 pyramid-v1.6: pyramid~=1.6.0 @@ -609,10 +594,6 @@ deps = redis-v5: redis~=5.0 redis-latest: redis - # Redis Cluster - redis_py_cluster_legacy-v1: redis-py-cluster~=1.0 - redis_py_cluster_legacy-v2: redis-py-cluster~=2.0 - # Requests requests: requests>=2.0 @@ -681,12 +662,6 @@ deps = starlite-v{1.48}: starlite~=1.48.0 starlite-v{1.51}: starlite~=1.51.0 - # SQLAlchemy - sqlalchemy-v1.2: sqlalchemy~=1.2.0 - sqlalchemy-v1.4: sqlalchemy~=1.4.0 - sqlalchemy-v2.0: sqlalchemy~=2.0.0 - sqlalchemy-latest: sqlalchemy - # Tornado # Tornado <6.4.1 is incompatible with Pytest ≥8.2 # See https://github.com/tornadoweb/tornado/pull/3382. @@ -699,6 +674,25 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ DBs ~~~ + clickhouse_driver-v0.2.9: clickhouse-driver==0.2.9 + + pymongo-v3.5.1: pymongo==3.5.1 + pymongo-v3.13.0: pymongo==3.13.0 + pymongo-v4.0.2: pymongo==4.0.2 + pymongo-v4.11.1: pymongo==4.11.1 + pymongo: mockupdb + + redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 + redis_py_cluster_legacy-v2.0.0: redis-py-cluster==2.0.0 + redis_py_cluster_legacy-v2.1.3: redis-py-cluster==2.1.3 + + sqlalchemy-v1.3.9: sqlalchemy==1.3.9 + sqlalchemy-v1.4.54: sqlalchemy==1.4.54 + sqlalchemy-v2.0.9: sqlalchemy==2.0.9 + sqlalchemy-v2.0.38: sqlalchemy==2.0.38 + + # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 From 85879b49bc715ea459864768ce9649ca6c6a9db9 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Feb 2025 17:23:12 +0100 Subject: [PATCH 139/591] tests: Generate some of the Web 1 tox entries with toxgen (#3980) - remove hardcoded entries for `flask`, `starlette` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to port to the script, I'll do this step by step in follow-up PRs. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-web-1.yml | 2 +- scripts/populate_tox/config.py | 23 ++++++ scripts/populate_tox/populate_tox.py | 2 - scripts/populate_tox/tox.jinja | 40 --------- sentry_sdk/integrations/__init__.py | 3 +- tox.ini | 82 +++++++++---------- 6 files changed, 66 insertions(+), 86 deletions(-) diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 9b3a2f06ec..e243ceb69a 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -115,7 +115,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index df99681e77..8982a8c53a 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -16,6 +16,13 @@ "clickhouse_driver": { "package": "clickhouse-driver", }, + "flask": { + "package": "flask", + "deps": { + "*": ["flask-login", "werkzeug"], + "<2.0": ["werkzeug<2.1.0", "markupsafe<2.1.0"], + }, + }, "gql": { "package": "gql[all]", }, @@ -47,6 +54,22 @@ "sqlalchemy": { "package": "sqlalchemy", }, + "starlette": { + "package": "starlette", + "deps": { + "*": [ + "pytest-asyncio", + "python-multipart", + "requests", + "anyio<4.0.0", + "jinja2", + "httpx", + ], + "<0.37": ["httpx<0.28.0"], + "<0.15": ["jinja2<3.1"], + "py3.6": ["aiocontextvars"], + }, + }, "statsig": { "package": "statsig", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 09c31923e6..01e5a7c463 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -68,7 +68,6 @@ "dramatiq", "falcon", "fastapi", - "flask", "gcp", "grpc", "httpx", @@ -88,7 +87,6 @@ "rq", "sanic", "spark", - "starlette", "starlite", "tornado", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index a7a7ff2615..70c570ba25 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -123,12 +123,6 @@ envlist = {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest - # Flask - {py3.6,py3.8}-flask-v{1} - {py3.8,py3.11,py3.12}-flask-v{2} - {py3.10,py3.11,py3.12}-flask-v{3} - {py3.10,py3.12,py3.13}-flask-latest - # GCP {py3.7}-gcp @@ -222,12 +216,6 @@ envlist = {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} {py3.8,py3.10,py3.11,py3.12}-spark-latest - # Starlette - {py3.7,py3.10}-starlette-v{0.19} - {py3.7,py3.11}-starlette-v{0.24,0.28} - {py3.8,py3.11,py3.12}-starlette-v{0.32,0.36,0.40} - {py3.8,py3.12,py3.13}-starlette-latest - # Starlite {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar @@ -410,16 +398,6 @@ deps = fastapi-v{0.79}: fastapi~=0.79.0 fastapi-latest: fastapi - # Flask - flask: flask-login - flask-v{1,2.0}: Werkzeug<2.1.0 - flask-v{1,2.0}: markupsafe<2.1.0 - flask-v{3}: Werkzeug - flask-v1: Flask~=1.0 - flask-v2: Flask~=2.0 - flask-v3: Flask~=3.0 - flask-latest: Flask - # gRPC grpc: protobuf grpc: mypy-protobuf @@ -581,24 +559,6 @@ deps = spark-v4.0: pyspark==4.0.0.dev2 spark-latest: pyspark - # Starlette - starlette: pytest-asyncio - starlette: python-multipart - starlette: requests - # (this is a dependency of httpx) - starlette: anyio<4.0.0 - starlette: jinja2 - starlette-v{0.19,0.24,0.28,0.32,0.36}: httpx<0.28.0 - starlette-v0.40: httpx - starlette-latest: httpx - starlette-v0.19: starlette~=0.19.0 - starlette-v0.24: starlette~=0.24.0 - starlette-v0.28: starlette~=0.28.0 - starlette-v0.32: starlette~=0.32.0 - starlette-v0.36: starlette~=0.36.0 - starlette-v0.40: starlette~=0.40.0 - starlette-latest: starlette - # Starlite starlite: pytest-asyncio starlite: python-multipart diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index d803a0b169..9bff264752 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -135,7 +135,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "dramatiq": (1, 9), "falcon": (1, 4), "fastapi": (0, 79, 0), - "flask": (0, 10), + "flask": (1, 1, 4), "gql": (3, 4, 1), "graphene": (3, 3), "grpc": (1, 32, 0), # grpcio @@ -151,6 +151,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "rq": (0, 6), "sanic": (0, 8), "sqlalchemy": (1, 2), + "starlette": (0, 16), "starlite": (1, 48), "statsig": (0, 55, 3), "strawberry": (0, 209, 5), diff --git a/tox.ini b/tox.ini index 0487b3c595..73085eb243 100644 --- a/tox.ini +++ b/tox.ini @@ -123,12 +123,6 @@ envlist = {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest - # Flask - {py3.6,py3.8}-flask-v{1} - {py3.8,py3.11,py3.12}-flask-v{2} - {py3.10,py3.11,py3.12}-flask-v{3} - {py3.10,py3.12,py3.13}-flask-latest - # GCP {py3.7}-gcp @@ -222,12 +216,6 @@ envlist = {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} {py3.8,py3.10,py3.11,py3.12}-spark-latest - # Starlette - {py3.7,py3.10}-starlette-v{0.19} - {py3.7,py3.11}-starlette-v{0.24,0.28} - {py3.8,py3.11,py3.12}-starlette-v{0.32,0.36,0.40} - {py3.8,py3.12,py3.13}-starlette-latest - # Starlite {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar @@ -288,7 +276,19 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.226.2 {py3.8,py3.11,py3.12}-strawberry-v0.243.1 - {py3.9,py3.12,py3.13}-strawberry-v0.260.1 + {py3.9,py3.12,py3.13}-strawberry-v0.260.2 + + + # ~~~ Web 1 ~~~ + {py3.6,py3.7,py3.8}-flask-v1.1.4 + {py3.8,py3.12,py3.13}-flask-v2.3.3 + {py3.8,py3.12,py3.13}-flask-v3.0.3 + {py3.9,py3.12,py3.13}-flask-v3.1.0 + + {py3.6,py3.9,py3.10}-starlette-v0.16.0 + {py3.7,py3.10,py3.11}-starlette-v0.26.1 + {py3.8,py3.11,py3.12}-starlette-v0.36.3 + {py3.9,py3.12,py3.13}-starlette-v0.45.3 # ~~~ Misc ~~~ @@ -463,16 +463,6 @@ deps = fastapi-v{0.79}: fastapi~=0.79.0 fastapi-latest: fastapi - # Flask - flask: flask-login - flask-v{1,2.0}: Werkzeug<2.1.0 - flask-v{1,2.0}: markupsafe<2.1.0 - flask-v{3}: Werkzeug - flask-v1: Flask~=1.0 - flask-v2: Flask~=2.0 - flask-v3: Flask~=3.0 - flask-latest: Flask - # gRPC grpc: protobuf grpc: mypy-protobuf @@ -634,24 +624,6 @@ deps = spark-v4.0: pyspark==4.0.0.dev2 spark-latest: pyspark - # Starlette - starlette: pytest-asyncio - starlette: python-multipart - starlette: requests - # (this is a dependency of httpx) - starlette: anyio<4.0.0 - starlette: jinja2 - starlette-v{0.19,0.24,0.28,0.32,0.36}: httpx<0.28.0 - starlette-v0.40: httpx - starlette-latest: httpx - starlette-v0.19: starlette~=0.19.0 - starlette-v0.24: starlette~=0.24.0 - starlette-v0.28: starlette~=0.28.0 - starlette-v0.32: starlette~=0.32.0 - starlette-v0.36: starlette~=0.36.0 - starlette-v0.40: starlette~=0.40.0 - starlette-latest: starlette - # Starlite starlite: pytest-asyncio starlite: python-multipart @@ -731,10 +703,36 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.226.2: strawberry-graphql[fastapi,flask]==0.226.2 strawberry-v0.243.1: strawberry-graphql[fastapi,flask]==0.243.1 - strawberry-v0.260.1: strawberry-graphql[fastapi,flask]==0.260.1 + strawberry-v0.260.2: strawberry-graphql[fastapi,flask]==0.260.2 strawberry: httpx + # ~~~ Web 1 ~~~ + flask-v1.1.4: flask==1.1.4 + flask-v2.3.3: flask==2.3.3 + flask-v3.0.3: flask==3.0.3 + flask-v3.1.0: flask==3.1.0 + flask: flask-login + flask: werkzeug + flask-v1.1.4: werkzeug<2.1.0 + flask-v1.1.4: markupsafe<2.1.0 + + starlette-v0.16.0: starlette==0.16.0 + starlette-v0.26.1: starlette==0.26.1 + starlette-v0.36.3: starlette==0.36.3 + starlette-v0.45.3: starlette==0.45.3 + starlette: pytest-asyncio + starlette: python-multipart + starlette: requests + starlette: anyio<4.0.0 + starlette: jinja2 + starlette: httpx + starlette-v0.16.0: httpx<0.28.0 + starlette-v0.26.1: httpx<0.28.0 + starlette-v0.36.3: httpx<0.28.0 + py3.6-starlette: aiocontextvars + + # ~~~ Misc ~~~ loguru-v0.7.3: loguru==0.7.3 From 25ddbcad9642cf38b7a9668e348f80fb9b1c892e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 14 Feb 2025 10:44:35 +0100 Subject: [PATCH 140/591] tests: Generate some of the AI tox entries by toxgen (#3977) - remove hardcoded entries for `huggingface_hub` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to port to the script, I'll do this step by step in follow-up PRs. This group in particular needs special treatment because of the `notiktoken` versions of some of the integrations. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-ai.yml | 4 ++-- scripts/populate_tox/config.py | 3 +++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 8 -------- tox.ini | 22 ++++++++++++++-------- 5 files changed, 19 insertions(+), 19 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index b9ade22f08..c3a2de036b 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.9","3.11","3.12","3.13"] + python-version: ["3.7","3.9","3.11","3.12"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -101,7 +101,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 8982a8c53a..0bfe1b618c 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -33,6 +33,9 @@ "py3.6": ["aiocontextvars"], }, }, + "huggingface_hub": { + "package": "huggingface_hub", + }, "launchdarkly": { "package": "launchdarkly-server-sdk", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 01e5a7c463..ff19ec3a5f 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -72,7 +72,6 @@ "grpc", "httpx", "huey", - "huggingface_hub", "langchain", "langchain_notiktoken", "litestar", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 70c570ba25..812bdf052a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -143,10 +143,6 @@ envlist = {py3.6,py3.11,py3.12}-huey-v{2.0} {py3.6,py3.12,py3.13}-huey-latest - # Huggingface Hub - {py3.9,py3.12,py3.13}-huggingface_hub-{v0.22} - {py3.9,py3.12,py3.13}-huggingface_hub-latest - # Langchain {py3.9,py3.11,py3.12}-langchain-v0.1 {py3.9,py3.11,py3.12}-langchain-v0.3 @@ -433,10 +429,6 @@ deps = huey-v2.0: huey~=2.0.0 huey-latest: huey - # Huggingface Hub - huggingface_hub-v0.22: huggingface_hub~=0.22.2 - huggingface_hub-latest: huggingface_hub - # Langchain langchain-v0.1: openai~=1.0.0 langchain-v0.1: langchain~=0.1.11 diff --git a/tox.ini b/tox.ini index 73085eb243..deea74b328 100644 --- a/tox.ini +++ b/tox.ini @@ -143,10 +143,6 @@ envlist = {py3.6,py3.11,py3.12}-huey-v{2.0} {py3.6,py3.12,py3.13}-huey-latest - # Huggingface Hub - {py3.9,py3.12,py3.13}-huggingface_hub-{v0.22} - {py3.9,py3.12,py3.13}-huggingface_hub-latest - # Langchain {py3.9,py3.11,py3.12}-langchain-v0.1 {py3.9,py3.11,py3.12}-langchain-v0.3 @@ -229,6 +225,13 @@ envlist = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ AI ~~~ + {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 + {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 + {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 + + # ~~~ DBs ~~~ {py3.7,py3.11,py3.12}-clickhouse_driver-v0.2.9 @@ -498,10 +501,6 @@ deps = huey-v2.0: huey~=2.0.0 huey-latest: huey - # Huggingface Hub - huggingface_hub-v0.22: huggingface_hub~=0.22.2 - huggingface_hub-latest: huggingface_hub - # Langchain langchain-v0.1: openai~=1.0.0 langchain-v0.1: langchain~=0.1.11 @@ -646,6 +645,13 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ AI ~~~ + huggingface_hub-v0.22.2: huggingface_hub==0.22.2 + huggingface_hub-v0.24.7: huggingface_hub==0.24.7 + huggingface_hub-v0.26.5: huggingface_hub==0.26.5 + huggingface_hub-v0.28.1: huggingface_hub==0.28.1 + + # ~~~ DBs ~~~ clickhouse_driver-v0.2.9: clickhouse-driver==0.2.9 From 8f22defb70d43ce79c12e6efc6437bd02d18d42d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 14 Feb 2025 11:18:45 +0100 Subject: [PATCH 141/591] tests: Generate part of the Tasks tox entries by a script (#3976) - remove hardcoded entries for `celery`, `spark`, `huey`, `dramatiq` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to switch over to the script, I'll do this step by step in follow-up PRs. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-tasks.yml | 4 +- scripts/populate_tox/config.py | 17 +++ scripts/populate_tox/populate_tox.py | 4 +- scripts/populate_tox/tox.jinja | 54 --------- tox.ini | 103 ++++++------------ 6 files changed, 56 insertions(+), 128 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index c3a2de036b..1a5df1d00f 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -101,7 +101,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.11","3.12"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 31e6f3c97a..6abefa29f4 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.10","3.11","3.12","3.13"] + python-version: ["3.7","3.8","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -115,7 +115,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 0bfe1b618c..3b6cb9b3d4 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -13,9 +13,19 @@ }, "python": ">=3.8", }, + "celery": { + "package": "celery", + "deps": { + "*": ["newrelic", "redis"], + "py3.7": ["importlib-metadata<5.0"], + }, + }, "clickhouse_driver": { "package": "clickhouse-driver", }, + "dramatiq": { + "package": "dramatiq", + }, "flask": { "package": "flask", "deps": { @@ -33,6 +43,9 @@ "py3.6": ["aiocontextvars"], }, }, + "huey": { + "package": "huey", + }, "huggingface_hub": { "package": "huggingface_hub", }, @@ -54,6 +67,10 @@ "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, + "spark": { + "package": "pyspark", + "python": ">=3.8", + }, "sqlalchemy": { "package": "sqlalchemy", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index ff19ec3a5f..855caa135d 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -59,19 +59,18 @@ "beam", "boto3", "bottle", - "celery", "chalice", "cohere", "cloud_resource_context", "cohere", "django", - "dramatiq", "falcon", "fastapi", "gcp", "grpc", "httpx", "huey", + "huggingface_hub", "langchain", "langchain_notiktoken", "litestar", @@ -85,7 +84,6 @@ "requests", "rq", "sanic", - "spark", "starlite", "tornado", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 812bdf052a..2e8d654d55 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -74,13 +74,6 @@ envlist = {py3.6,py3.9}-bottle-v{0.12} {py3.6,py3.12,py3.13}-bottle-latest - # Celery - {py3.6,py3.8}-celery-v{4} - {py3.6,py3.8}-celery-v{5.0} - {py3.7,py3.10}-celery-v{5.1,5.2} - {py3.8,py3.11,py3.12}-celery-v{5.3,5.4,5.5} - {py3.8,py3.12,py3.13}-celery-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -107,12 +100,6 @@ envlist = {py3.10,py3.11,py3.12}-django-v{5.0,5.1} {py3.10,py3.12,py3.13}-django-latest - # dramatiq - {py3.6,py3.9}-dramatiq-v{1.13} - {py3.7,py3.10,py3.11}-dramatiq-v{1.15} - {py3.8,py3.11,py3.12}-dramatiq-v{1.17} - {py3.8,py3.11,py3.12}-dramatiq-latest - # Falcon {py3.6,py3.7}-falcon-v{1,1.4,2} {py3.6,py3.11,py3.12}-falcon-v{3} @@ -139,10 +126,6 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} {py3.9,py3.12,py3.13}-httpx-latest - # Huey - {py3.6,py3.11,py3.12}-huey-v{2.0} - {py3.6,py3.12,py3.13}-huey-latest - # Langchain {py3.9,py3.11,py3.12}-langchain-v0.1 {py3.9,py3.11,py3.12}-langchain-v0.3 @@ -208,10 +191,6 @@ envlist = {py3.8,py3.11,py3.12}-sanic-v{24.6} {py3.9,py3.12,py3.13}-sanic-latest - # Spark - {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} - {py3.8,py3.10,py3.11,py3.12}-spark-latest - # Starlite {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar @@ -317,21 +296,6 @@ deps = bottle-v0.12: bottle~=0.12.0 bottle-latest: bottle - # Celery - celery: redis - celery-v4: Celery~=4.0 - celery-v5.0: Celery~=5.0.0 - celery-v5.1: Celery~=5.1.0 - celery-v5.2: Celery~=5.2.0 - celery-v5.3: Celery~=5.3.0 - celery-v5.4: Celery~=5.4.0 - # TODO: update when stable is out - celery-v5.5: Celery==5.5.0rc4 - celery-latest: Celery - - celery: newrelic - {py3.7}-celery: importlib-metadata<5.0 - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 @@ -370,12 +334,6 @@ deps = django-v5.1: Django==5.1rc1 django-latest: Django - # dramatiq - dramatiq-v1.13: dramatiq>=1.13,<1.14 - dramatiq-v1.15: dramatiq>=1.15,<1.16 - dramatiq-v1.17: dramatiq>=1.17,<1.18 - dramatiq-latest: dramatiq - # Falcon falcon-v1.4: falcon~=1.4.0 falcon-v1: falcon~=1.0 @@ -425,10 +383,6 @@ deps = httpx-v0.27: httpx~=0.27.0 httpx-latest: httpx - # Huey - huey-v2.0: huey~=2.0.0 - huey-latest: huey - # Langchain langchain-v0.1: openai~=1.0.0 langchain-v0.1: langchain~=0.1.11 @@ -543,14 +497,6 @@ deps = sanic-v24.6: sanic~=24.6.0 sanic-latest: sanic - # Spark - spark-v3.1: pyspark~=3.1.0 - spark-v3.3: pyspark~=3.3.0 - spark-v3.5: pyspark~=3.5.0 - # TODO: update to ~=4.0.0 once stable is out - spark-v4.0: pyspark==4.0.0.dev2 - spark-latest: pyspark - # Starlite starlite: pytest-asyncio starlite: python-multipart diff --git a/tox.ini b/tox.ini index deea74b328..71a9588f3e 100644 --- a/tox.ini +++ b/tox.ini @@ -74,13 +74,6 @@ envlist = {py3.6,py3.9}-bottle-v{0.12} {py3.6,py3.12,py3.13}-bottle-latest - # Celery - {py3.6,py3.8}-celery-v{4} - {py3.6,py3.8}-celery-v{5.0} - {py3.7,py3.10}-celery-v{5.1,5.2} - {py3.8,py3.11,py3.12}-celery-v{5.3,5.4,5.5} - {py3.8,py3.12,py3.13}-celery-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -107,12 +100,6 @@ envlist = {py3.10,py3.11,py3.12}-django-v{5.0,5.1} {py3.10,py3.12,py3.13}-django-latest - # dramatiq - {py3.6,py3.9}-dramatiq-v{1.13} - {py3.7,py3.10,py3.11}-dramatiq-v{1.15} - {py3.8,py3.11,py3.12}-dramatiq-v{1.17} - {py3.8,py3.11,py3.12}-dramatiq-latest - # Falcon {py3.6,py3.7}-falcon-v{1,1.4,2} {py3.6,py3.11,py3.12}-falcon-v{3} @@ -139,10 +126,6 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} {py3.9,py3.12,py3.13}-httpx-latest - # Huey - {py3.6,py3.11,py3.12}-huey-v{2.0} - {py3.6,py3.12,py3.13}-huey-latest - # Langchain {py3.9,py3.11,py3.12}-langchain-v0.1 {py3.9,py3.11,py3.12}-langchain-v0.3 @@ -208,10 +191,6 @@ envlist = {py3.8,py3.11,py3.12}-sanic-v{24.6} {py3.9,py3.12,py3.13}-sanic-latest - # Spark - {py3.8,py3.10,py3.11}-spark-v{3.1,3.3,3.5} - {py3.8,py3.10,py3.11,py3.12}-spark-latest - # Starlite {py3.8,py3.11}-starlite-v{1.48,1.51} # 1.51.14 is the last starlite version; the project continues as litestar @@ -225,13 +204,6 @@ envlist = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. - # ~~~ AI ~~~ - {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 - {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 - {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - - # ~~~ DBs ~~~ {py3.7,py3.11,py3.12}-clickhouse_driver-v0.2.9 @@ -282,6 +254,22 @@ envlist = {py3.9,py3.12,py3.13}-strawberry-v0.260.2 + # ~~~ Tasks ~~~ + {py3.6,py3.7,py3.8}-celery-v4.4.7 + {py3.6,py3.7,py3.8}-celery-v5.0.5 + {py3.8,py3.11,py3.12}-celery-v5.4.0 + + {py3.6,py3.7}-dramatiq-v1.9.0 + {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 + {py3.7,py3.10,py3.11}-dramatiq-v1.15.0 + {py3.8,py3.12,py3.13}-dramatiq-v1.17.1 + + {py3.8,py3.9}-spark-v3.0.3 + {py3.8,py3.9}-spark-v3.2.4 + {py3.8,py3.10,py3.11}-spark-v3.4.4 + {py3.8,py3.10,py3.11}-spark-v3.5.4 + + # ~~~ Web 1 ~~~ {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -389,21 +377,6 @@ deps = bottle-v0.12: bottle~=0.12.0 bottle-latest: bottle - # Celery - celery: redis - celery-v4: Celery~=4.0 - celery-v5.0: Celery~=5.0.0 - celery-v5.1: Celery~=5.1.0 - celery-v5.2: Celery~=5.2.0 - celery-v5.3: Celery~=5.3.0 - celery-v5.4: Celery~=5.4.0 - # TODO: update when stable is out - celery-v5.5: Celery==5.5.0rc4 - celery-latest: Celery - - celery: newrelic - {py3.7}-celery: importlib-metadata<5.0 - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 @@ -442,12 +415,6 @@ deps = django-v5.1: Django==5.1rc1 django-latest: Django - # dramatiq - dramatiq-v1.13: dramatiq>=1.13,<1.14 - dramatiq-v1.15: dramatiq>=1.15,<1.16 - dramatiq-v1.17: dramatiq>=1.17,<1.18 - dramatiq-latest: dramatiq - # Falcon falcon-v1.4: falcon~=1.4.0 falcon-v1: falcon~=1.0 @@ -497,10 +464,6 @@ deps = httpx-v0.27: httpx~=0.27.0 httpx-latest: httpx - # Huey - huey-v2.0: huey~=2.0.0 - huey-latest: huey - # Langchain langchain-v0.1: openai~=1.0.0 langchain-v0.1: langchain~=0.1.11 @@ -615,14 +578,6 @@ deps = sanic-v24.6: sanic~=24.6.0 sanic-latest: sanic - # Spark - spark-v3.1: pyspark~=3.1.0 - spark-v3.3: pyspark~=3.3.0 - spark-v3.5: pyspark~=3.5.0 - # TODO: update to ~=4.0.0 once stable is out - spark-v4.0: pyspark==4.0.0.dev2 - spark-latest: pyspark - # Starlite starlite: pytest-asyncio starlite: python-multipart @@ -645,13 +600,6 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. - # ~~~ AI ~~~ - huggingface_hub-v0.22.2: huggingface_hub==0.22.2 - huggingface_hub-v0.24.7: huggingface_hub==0.24.7 - huggingface_hub-v0.26.5: huggingface_hub==0.26.5 - huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - - # ~~~ DBs ~~~ clickhouse_driver-v0.2.9: clickhouse-driver==0.2.9 @@ -713,6 +661,25 @@ deps = strawberry: httpx + # ~~~ Tasks ~~~ + celery-v4.4.7: celery==4.4.7 + celery-v5.0.5: celery==5.0.5 + celery-v5.4.0: celery==5.4.0 + celery: newrelic + celery: redis + py3.7-celery: importlib-metadata<5.0 + + dramatiq-v1.9.0: dramatiq==1.9.0 + dramatiq-v1.12.3: dramatiq==1.12.3 + dramatiq-v1.15.0: dramatiq==1.15.0 + dramatiq-v1.17.1: dramatiq==1.17.1 + + spark-v3.0.3: pyspark==3.0.3 + spark-v3.2.4: pyspark==3.2.4 + spark-v3.4.4: pyspark==3.4.4 + spark-v3.5.4: pyspark==3.5.4 + + # ~~~ Web 1 ~~~ flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 From 24afdb36f27f2ca7f4484edc523c58942030696c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 14 Feb 2025 11:33:09 +0100 Subject: [PATCH 142/591] tests: Generate some of the Web 2 tox entries by toxgen (#3981) - remove hardcoded entries for `falcon`, `starlite`, `pyramid`, `bottle`, `tornado` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to port to the script, I'll do this step by step in follow-up PRs. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/test-integrations-web-2.yml | 4 +- scripts/populate_tox/config.py | 40 ++++++ scripts/populate_tox/populate_tox.py | 5 - scripts/populate_tox/tox.jinja | 63 --------- tox.ini | 124 +++++++++--------- 5 files changed, 103 insertions(+), 133 deletions(-) diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 3c010fc0bd..b3973aa960 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -121,7 +121,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 3b6cb9b3d4..0f0e150a4f 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -13,6 +13,12 @@ }, "python": ">=3.8", }, + "bottle": { + "package": "bottle", + "deps": { + "*": ["werkzeug<2.1.0"], + }, + }, "celery": { "package": "celery", "deps": { @@ -26,6 +32,10 @@ "dramatiq": { "package": "dramatiq", }, + "falcon": { + "package": "falcon", + "python": "<3.13", + }, "flask": { "package": "flask", "deps": { @@ -64,6 +74,12 @@ "*": ["mockupdb"], }, }, + "pyramid": { + "package": "pyramid", + "deps": { + "*": ["werkzeug<2.1.0"], + }, + }, "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, @@ -90,6 +106,20 @@ "py3.6": ["aiocontextvars"], }, }, + "starlite": { + "package": "starlite", + "deps": { + "*": [ + "pytest-asyncio", + "python-multipart", + "requests", + "cryptography", + "pydantic<2.0.0", + "httpx<0.28", + ], + }, + "python": "<=3.11", + }, "statsig": { "package": "statsig", "deps": { @@ -102,6 +132,16 @@ "*": ["httpx"], }, }, + "tornado": { + "package": "tornado", + "deps": { + "*": ["pytest"], + "<=6.4.1": [ + "pytest<8.2" + ], # https://github.com/tornadoweb/tornado/pull/3382 + "py3.6": ["aiocontextvars"], + }, + }, "trytond": { "package": "trytond", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 855caa135d..e6cb0e4de1 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -58,13 +58,11 @@ "aws_lambda", "beam", "boto3", - "bottle", "chalice", "cohere", "cloud_resource_context", "cohere", "django", - "falcon", "fastapi", "gcp", "grpc", @@ -77,15 +75,12 @@ "openai", "openai_notiktoken", "pure_eval", - "pyramid", "quart", "ray", "redis", "requests", "rq", "sanic", - "starlite", - "tornado", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 2e8d654d55..a6fc55c7e4 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -70,10 +70,6 @@ envlist = {py3.11,py3.12}-boto3-v{1.34} {py3.11,py3.12,py3.13}-boto3-latest - # Bottle - {py3.6,py3.9}-bottle-v{0.12} - {py3.6,py3.12,py3.13}-bottle-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -100,12 +96,6 @@ envlist = {py3.10,py3.11,py3.12}-django-v{5.0,5.1} {py3.10,py3.12,py3.13}-django-latest - # Falcon - {py3.6,py3.7}-falcon-v{1,1.4,2} - {py3.6,py3.11,py3.12}-falcon-v{3} - {py3.8,py3.11,py3.12}-falcon-v{4} - {py3.7,py3.11,py3.12}-falcon-latest - # FastAPI {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest @@ -154,12 +144,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # Pyramid - {py3.6,py3.11}-pyramid-v{1.6} - {py3.6,py3.11,py3.12}-pyramid-v{1.10} - {py3.6,py3.11,py3.12}-pyramid-v{2.0} - {py3.6,py3.11,py3.12}-pyramid-latest - # Quart {py3.7,py3.11}-quart-v{0.16} {py3.8,py3.11,py3.12}-quart-v{0.19} @@ -191,15 +175,6 @@ envlist = {py3.8,py3.11,py3.12}-sanic-v{24.6} {py3.9,py3.12,py3.13}-sanic-latest - # Starlite - {py3.8,py3.11}-starlite-v{1.48,1.51} - # 1.51.14 is the last starlite version; the project continues as litestar - - # Tornado - {py3.8,py3.11,py3.12}-tornado-v{6.0} - {py3.8,py3.11,py3.12}-tornado-v{6.2} - {py3.8,py3.11,py3.12}-tornado-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -291,11 +266,6 @@ deps = boto3-v1.34: boto3~=1.34.0 boto3-latest: boto3 - # Bottle - bottle: Werkzeug<2.1.0 - bottle-v0.12: bottle~=0.12.0 - bottle-latest: bottle - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 @@ -334,14 +304,6 @@ deps = django-v5.1: Django==5.1rc1 django-latest: Django - # Falcon - falcon-v1.4: falcon~=1.4.0 - falcon-v1: falcon~=1.0 - falcon-v2: falcon~=2.0 - falcon-v3: falcon~=3.0 - falcon-v4: falcon~=4.0 - falcon-latest: falcon - # FastAPI fastapi: httpx # (this is a dependency of httpx) @@ -431,13 +393,6 @@ deps = # pure_eval pure_eval: pure_eval - # Pyramid - pyramid: Werkzeug<2.1.0 - pyramid-v1.6: pyramid~=1.6.0 - pyramid-v1.10: pyramid~=1.10.0 - pyramid-v2.0: pyramid~=2.0.0 - pyramid-latest: pyramid - # Quart quart: quart-auth quart: pytest-asyncio @@ -497,24 +452,6 @@ deps = sanic-v24.6: sanic~=24.6.0 sanic-latest: sanic - # Starlite - starlite: pytest-asyncio - starlite: python-multipart - starlite: requests - starlite: cryptography - starlite: pydantic<2.0.0 - starlite: httpx<0.28 - starlite-v{1.48}: starlite~=1.48.0 - starlite-v{1.51}: starlite~=1.51.0 - - # Tornado - # Tornado <6.4.1 is incompatible with Pytest ≥8.2 - # See https://github.com/tornadoweb/tornado/pull/3382. - tornado-{v6.0,v6.2}: pytest<8.2 - tornado-v6.0: tornado~=6.0.0 - tornado-v6.2: tornado~=6.2.0 - tornado-latest: tornado - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. diff --git a/tox.ini b/tox.ini index 71a9588f3e..cb3538e1aa 100644 --- a/tox.ini +++ b/tox.ini @@ -70,10 +70,6 @@ envlist = {py3.11,py3.12}-boto3-v{1.34} {py3.11,py3.12,py3.13}-boto3-latest - # Bottle - {py3.6,py3.9}-bottle-v{0.12} - {py3.6,py3.12,py3.13}-bottle-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -100,12 +96,6 @@ envlist = {py3.10,py3.11,py3.12}-django-v{5.0,5.1} {py3.10,py3.12,py3.13}-django-latest - # Falcon - {py3.6,py3.7}-falcon-v{1,1.4,2} - {py3.6,py3.11,py3.12}-falcon-v{3} - {py3.8,py3.11,py3.12}-falcon-v{4} - {py3.7,py3.11,py3.12}-falcon-latest - # FastAPI {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest @@ -154,12 +144,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # Pyramid - {py3.6,py3.11}-pyramid-v{1.6} - {py3.6,py3.11,py3.12}-pyramid-v{1.10} - {py3.6,py3.11,py3.12}-pyramid-v{2.0} - {py3.6,py3.11,py3.12}-pyramid-latest - # Quart {py3.7,py3.11}-quart-v{0.16} {py3.8,py3.11,py3.12}-quart-v{0.19} @@ -191,15 +175,6 @@ envlist = {py3.8,py3.11,py3.12}-sanic-v{24.6} {py3.9,py3.12,py3.13}-sanic-latest - # Starlite - {py3.8,py3.11}-starlite-v{1.48,1.51} - # 1.51.14 is the last starlite version; the project continues as litestar - - # Tornado - {py3.8,py3.11,py3.12}-tornado-v{6.0} - {py3.8,py3.11,py3.12}-tornado-v{6.2} - {py3.8,py3.11,py3.12}-tornado-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -282,6 +257,30 @@ envlist = {py3.9,py3.12,py3.13}-starlette-v0.45.3 + # ~~~ Web 2 ~~~ + {py3.6,py3.7}-bottle-v0.12.25 + {py3.6,py3.8,py3.9}-bottle-v0.13.2 + + {py3.6}-falcon-v1.4.1 + {py3.6,py3.7}-falcon-v2.0.0 + {py3.6,py3.11,py3.12}-falcon-v3.1.3 + {py3.8,py3.11,py3.12}-falcon-v4.0.2 + + {py3.6}-pyramid-v1.8.6 + {py3.6,py3.8,py3.9}-pyramid-v1.10.8 + {py3.6,py3.10,py3.11}-pyramid-v2.0.2 + + {py3.8,py3.10,py3.11}-starlite-v1.48.1 + {py3.8,py3.10,py3.11}-starlite-v1.49.0 + {py3.8,py3.10,py3.11}-starlite-v1.50.2 + {py3.8,py3.10,py3.11}-starlite-v1.51.16 + + {py3.6,py3.7,py3.8}-tornado-v6.0.4 + {py3.6,py3.8,py3.9}-tornado-v6.1 + {py3.7,py3.9,py3.10}-tornado-v6.2 + {py3.8,py3.10,py3.11}-tornado-v6.4.2 + + # ~~~ Misc ~~~ {py3.6,py3.12,py3.13}-loguru-v0.7.3 @@ -372,11 +371,6 @@ deps = boto3-v1.34: boto3~=1.34.0 boto3-latest: boto3 - # Bottle - bottle: Werkzeug<2.1.0 - bottle-v0.12: bottle~=0.12.0 - bottle-latest: bottle - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 @@ -415,14 +409,6 @@ deps = django-v5.1: Django==5.1rc1 django-latest: Django - # Falcon - falcon-v1.4: falcon~=1.4.0 - falcon-v1: falcon~=1.0 - falcon-v2: falcon~=2.0 - falcon-v3: falcon~=3.0 - falcon-v4: falcon~=4.0 - falcon-latest: falcon - # FastAPI fastapi: httpx # (this is a dependency of httpx) @@ -512,13 +498,6 @@ deps = # pure_eval pure_eval: pure_eval - # Pyramid - pyramid: Werkzeug<2.1.0 - pyramid-v1.6: pyramid~=1.6.0 - pyramid-v1.10: pyramid~=1.10.0 - pyramid-v2.0: pyramid~=2.0.0 - pyramid-latest: pyramid - # Quart quart: quart-auth quart: pytest-asyncio @@ -578,24 +557,6 @@ deps = sanic-v24.6: sanic~=24.6.0 sanic-latest: sanic - # Starlite - starlite: pytest-asyncio - starlite: python-multipart - starlite: requests - starlite: cryptography - starlite: pydantic<2.0.0 - starlite: httpx<0.28 - starlite-v{1.48}: starlite~=1.48.0 - starlite-v{1.51}: starlite~=1.51.0 - - # Tornado - # Tornado <6.4.1 is incompatible with Pytest ≥8.2 - # See https://github.com/tornadoweb/tornado/pull/3382. - tornado-{v6.0,v6.2}: pytest<8.2 - tornado-v6.0: tornado~=6.0.0 - tornado-v6.2: tornado~=6.2.0 - tornado-latest: tornado - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -706,6 +667,43 @@ deps = py3.6-starlette: aiocontextvars + # ~~~ Web 2 ~~~ + bottle-v0.12.25: bottle==0.12.25 + bottle-v0.13.2: bottle==0.13.2 + bottle: werkzeug<2.1.0 + + falcon-v1.4.1: falcon==1.4.1 + falcon-v2.0.0: falcon==2.0.0 + falcon-v3.1.3: falcon==3.1.3 + falcon-v4.0.2: falcon==4.0.2 + + pyramid-v1.8.6: pyramid==1.8.6 + pyramid-v1.10.8: pyramid==1.10.8 + pyramid-v2.0.2: pyramid==2.0.2 + pyramid: werkzeug<2.1.0 + + starlite-v1.48.1: starlite==1.48.1 + starlite-v1.49.0: starlite==1.49.0 + starlite-v1.50.2: starlite==1.50.2 + starlite-v1.51.16: starlite==1.51.16 + starlite: pytest-asyncio + starlite: python-multipart + starlite: requests + starlite: cryptography + starlite: pydantic<2.0.0 + starlite: httpx<0.28 + + tornado-v6.0.4: tornado==6.0.4 + tornado-v6.1: tornado==6.1 + tornado-v6.2: tornado==6.2 + tornado-v6.4.2: tornado==6.4.2 + tornado: pytest + tornado-v6.0.4: pytest<8.2 + tornado-v6.1: pytest<8.2 + tornado-v6.2: pytest<8.2 + py3.6-tornado: aiocontextvars + + # ~~~ Misc ~~~ loguru-v0.7.3: loguru==0.7.3 From 7b5904a17bd51521c5c5ee58ba60b3460ec1806d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 14 Feb 2025 14:09:40 +0100 Subject: [PATCH 143/591] tests: Generate tox entries for grpc via script (#3979) - remove hardcoded entries for `grpc` from the tox template - remove them from the ignore list in `populate_tox.py` - run `populate_tox.py` to fill in entries for them - run `split_gh_tox_actions.py` to generate the CI yaml files so that they correspond to the new tox.ini The remaining integrations in this group are not trivial to port to the script, I'll do this step by step in follow-up PRs. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .../workflows/test-integrations-network.yml | 2 +- scripts/populate_tox/config.py | 10 ++++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 16 --------- tox.ini | 34 ++++++++++--------- 5 files changed, 29 insertions(+), 34 deletions(-) diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 56f4bcfd57..aae29ab7f9 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.9","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 0f0e150a4f..2c2920e7ac 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -53,6 +53,13 @@ "py3.6": ["aiocontextvars"], }, }, + "grpc": { + "package": "grpcio", + "deps": { + "*": ["protobuf", "mypy-protobuf", "types-protobuf", "pytest-asyncio"], + }, + "python": ">=3.7", + }, "huey": { "package": "huey", }, @@ -83,6 +90,9 @@ "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, + "requests": { + "package": "requests", + }, "spark": { "package": "pyspark", "python": ">=3.8", diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index e6cb0e4de1..4bfce80ce7 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -65,7 +65,6 @@ "django", "fastapi", "gcp", - "grpc", "httpx", "huey", "huggingface_hub", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index a6fc55c7e4..15119b4768 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -103,12 +103,6 @@ envlist = # GCP {py3.7}-gcp - # gRPC - {py3.7,py3.9}-grpc-v{1.39} - {py3.7,py3.10}-grpc-v{1.49} - {py3.7,py3.11}-grpc-v{1.59} - {py3.8,py3.11,py3.12}-grpc-latest - # HTTPX {py3.6,py3.9}-httpx-v{0.16,0.18} {py3.6,py3.10}-httpx-v{0.20,0.22} @@ -314,16 +308,6 @@ deps = fastapi-v{0.79}: fastapi~=0.79.0 fastapi-latest: fastapi - # gRPC - grpc: protobuf - grpc: mypy-protobuf - grpc: types-protobuf - grpc: pytest-asyncio - grpc-v1.39: grpcio~=1.39.0 - grpc-v1.49: grpcio~=1.49.1 - grpc-v1.59: grpcio~=1.59.0 - grpc-latest: grpcio - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 diff --git a/tox.ini b/tox.ini index cb3538e1aa..9ce3d40a21 100644 --- a/tox.ini +++ b/tox.ini @@ -103,12 +103,6 @@ envlist = # GCP {py3.7}-gcp - # gRPC - {py3.7,py3.9}-grpc-v{1.39} - {py3.7,py3.10}-grpc-v{1.49} - {py3.7,py3.11}-grpc-v{1.59} - {py3.8,py3.11,py3.12}-grpc-latest - # HTTPX {py3.6,py3.9}-httpx-v{0.16,0.18} {py3.6,py3.10}-httpx-v{0.20,0.22} @@ -229,6 +223,13 @@ envlist = {py3.9,py3.12,py3.13}-strawberry-v0.260.2 + # ~~~ Network ~~~ + {py3.7,py3.8}-grpc-v1.32.0 + {py3.7,py3.9,py3.10}-grpc-v1.44.0 + {py3.7,py3.10,py3.11}-grpc-v1.58.3 + {py3.8,py3.12,py3.13}-grpc-v1.70.0 + + # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 @@ -419,16 +420,6 @@ deps = fastapi-v{0.79}: fastapi~=0.79.0 fastapi-latest: fastapi - # gRPC - grpc: protobuf - grpc: mypy-protobuf - grpc: types-protobuf - grpc: pytest-asyncio - grpc-v1.39: grpcio~=1.39.0 - grpc-v1.49: grpcio~=1.49.1 - grpc-v1.59: grpcio~=1.59.0 - grpc-latest: grpcio - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 @@ -622,6 +613,17 @@ deps = strawberry: httpx + # ~~~ Network ~~~ + grpc-v1.32.0: grpcio==1.32.0 + grpc-v1.44.0: grpcio==1.44.0 + grpc-v1.58.3: grpcio==1.58.3 + grpc-v1.70.0: grpcio==1.70.0 + grpc: protobuf + grpc: mypy-protobuf + grpc: types-protobuf + grpc: pytest-asyncio + + # ~~~ Tasks ~~~ celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 From ae68d8536e5712ed00cbe088372bcd7873d742b1 Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Fri, 14 Feb 2025 08:51:37 -0600 Subject: [PATCH 144/591] Revert "feat(tracing): Add `propagate_traces` deprecation warning (#3899)" (#4055) Closes: https://github.com/getsentry/sentry-python/issues/4054 We should log deprecation notices but since this notice is not actionable it should be removed. --- sentry_sdk/integrations/celery/__init__.py | 7 ------- tests/integrations/celery/test_celery.py | 11 ++++------- 2 files changed, 4 insertions(+), 14 deletions(-) diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index 80decb6064..dc48aac0e6 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -1,6 +1,4 @@ import sys -import warnings - from collections.abc import Mapping from functools import wraps @@ -70,11 +68,6 @@ def __init__( exclude_beat_tasks=None, ): # type: (bool, bool, Optional[List[str]]) -> None - warnings.warn( - "The `propagate_traces` parameter is deprecated. Please use `trace_propagation_targets` instead.", - DeprecationWarning, - stacklevel=2, - ) self.propagate_traces = propagate_traces self.monitor_beat_tasks = monitor_beat_tasks self.exclude_beat_tasks = exclude_beat_tasks diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index f8d118e7e9..e51341599f 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -268,9 +268,7 @@ def dummy_task(): def test_simple_no_propagation(capture_events, init_celery): - with pytest.warns(DeprecationWarning): - celery = init_celery(propagate_traces=False) - + celery = init_celery(propagate_traces=False) events = capture_events() @celery.task(name="dummy_task") @@ -534,10 +532,9 @@ def test_sentry_propagate_traces_override(init_celery): Test if the `sentry-propagate-traces` header given to `apply_async` overrides the `propagate_traces` parameter in the integration constructor. """ - with pytest.warns(DeprecationWarning): - celery = init_celery( - propagate_traces=True, traces_sample_rate=1.0, release="abcdef" - ) + celery = init_celery( + propagate_traces=True, traces_sample_rate=1.0, release="abcdef" + ) @celery.task(name="dummy_task", bind=True) def dummy_task(self, message): From 6a1b7d4798a4aa48557e39a3e922cc49213dc007 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Mon, 17 Feb 2025 04:33:15 -0500 Subject: [PATCH 145/591] tests(httplib): Fix flakey https test (#4057) Ideally this test shouldn't even make a request anywhere but this should make it a little more stable. This test failed 3 times on the same PR - https://github.com/getsentry/sentry-python/actions/runs/13337072005/job/37254546574?pr=4056 - https://github.com/getsentry/sentry-python/actions/runs/13337072005/job/37254551103?pr=4056 - https://github.com/getsentry/sentry-python/actions/runs/13337072011/job/37254546356?pr=4056 --- tests/integrations/stdlib/test_httplib.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index f2de190de0..227a24336c 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -380,7 +380,7 @@ def test_span_origin(sentry_init, capture_events): events = capture_events() with start_transaction(name="foo"): - conn = HTTPSConnection("example.com") + conn = HTTPConnection("example.com") conn.request("GET", "/foo") conn.getresponse() From 1abad47110887960c50865b7f93963bbccf6458d Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 17 Feb 2025 13:43:50 +0000 Subject: [PATCH 146/591] release: 2.22.0 --- CHANGELOG.md | 23 +++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 26 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5da35ac676..54f565c4e6 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,28 @@ # Changelog +## 2.22.0 + +### Various fixes & improvements + +- tests(httplib): Fix flakey https test (#4057) by @Zylphrex +- Revert "feat(tracing): Add `propagate_traces` deprecation warning (#3899)" (#4055) by @cmanallen +- tests: Generate tox entries for grpc via script (#3979) by @sentrivana +- tests: Generate some of the Web 2 tox entries by toxgen (#3981) by @sentrivana +- tests: Generate part of the Tasks tox entries by a script (#3976) by @sentrivana +- tests: Generate some of the AI tox entries by toxgen (#3977) by @sentrivana +- tests: Generate some of the Web 1 tox entries with toxgen (#3980) by @sentrivana +- tests: Generate DB group by toxgen script (#3978) by @sentrivana +- Generate Misc tox entries via toxgen script (#3982) by @sentrivana +- tests: Generate Flags tox entries with toxgen script (#3974) by @sentrivana +- tests: Remove toxgen cutoff, add statsig (#4048) by @sentrivana +- Fix clickhouse test (#4053) by @sentrivana +- tests(profiling): Reduce continuous profiling test flakiness (#4052) by @Zylphrex +- feat(profiling): Continuous profiling lifecycle (#4017) by @Zylphrex +- feat(flags): add Statsig integration (#4022) by @aliu39 +- Update changelog with `profile_session_sample_rate` (#4046) by @sentrivana +- Move the GraphQL group over to the tox gen script (#3975) by @sentrivana +- Update sample rate in DSC (#4018) by @sentrivana + ## 2.21.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index b7ae919e9a..0928eea74f 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.21.0" +release = "2.22.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index df2c2b52a0..20179e2231 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -584,4 +584,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.21.0" +VERSION = "2.22.0" diff --git a/setup.py b/setup.py index 21793220d4..675f5bb1bc 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.21.0", + version="2.22.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 1fcd36414d45de2fcf661806a8803fea80cf3498 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 17 Feb 2025 15:01:24 +0100 Subject: [PATCH 147/591] Updated Changelog.md --- CHANGELOG.md | 37 ++++++++++++++++++++----------------- 1 file changed, 20 insertions(+), 17 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 54f565c4e6..acc018f65c 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,24 +4,27 @@ ### Various fixes & improvements -- tests(httplib): Fix flakey https test (#4057) by @Zylphrex -- Revert "feat(tracing): Add `propagate_traces` deprecation warning (#3899)" (#4055) by @cmanallen -- tests: Generate tox entries for grpc via script (#3979) by @sentrivana -- tests: Generate some of the Web 2 tox entries by toxgen (#3981) by @sentrivana -- tests: Generate part of the Tasks tox entries by a script (#3976) by @sentrivana -- tests: Generate some of the AI tox entries by toxgen (#3977) by @sentrivana -- tests: Generate some of the Web 1 tox entries with toxgen (#3980) by @sentrivana -- tests: Generate DB group by toxgen script (#3978) by @sentrivana -- Generate Misc tox entries via toxgen script (#3982) by @sentrivana -- tests: Generate Flags tox entries with toxgen script (#3974) by @sentrivana -- tests: Remove toxgen cutoff, add statsig (#4048) by @sentrivana -- Fix clickhouse test (#4053) by @sentrivana -- tests(profiling): Reduce continuous profiling test flakiness (#4052) by @Zylphrex -- feat(profiling): Continuous profiling lifecycle (#4017) by @Zylphrex -- feat(flags): add Statsig integration (#4022) by @aliu39 -- Update changelog with `profile_session_sample_rate` (#4046) by @sentrivana -- Move the GraphQL group over to the tox gen script (#3975) by @sentrivana +- **New integration:** Add [Statsig](https://statsig.com/) integration (#4022) by @aliu39 + + For more information, see the documentation for the [TyperIntegration](https://docs.sentry.io/platforms/python/integrations/statsig/). + +- Profiling: Continuous profiling lifecycle (#4017) by @Zylphrex +- Fix: Revert "feat(tracing): Add `propagate_traces` deprecation warning (#3899)" (#4055) by @cmanallen +- Tests: Generate Web 1 group tox entries by toxgen script (#3980) by @sentrivana +- Tests: Generate Web 2 group tox entries by toxgen script (#3981) by @sentrivana +- Tests: Generate Tasks group tox entries by toxgen script (#3976) by @sentrivana +- Tests: Generate AI group tox entries by toxgen script (#3977) by @sentrivana +- Tests: Generate DB group tox entries by toxgen script (#3978) by @sentrivana +- Tests: Generate Misc group tox entries by toxgen script (#3982) by @sentrivana +- Tests: Generate Flags group tox entries by toxgen script (#3974) by @sentrivana +- Tests: Generate gRPC tox entries by toxgen script (#3979) by @sentrivana +- Tests: Remove toxgen cutoff, add statsig (#4048) by @sentrivana +- Tests: Reduce continuous profiling test flakiness (#4052) by @Zylphrex +- Tests: Fix Clickhouse test (#4053) by @sentrivana +- Tests: Fix flaky HTTPS test (#4057) by @Zylphrex - Update sample rate in DSC (#4018) by @sentrivana +- Move the GraphQL group over to the tox gen script (#3975) by @sentrivana +- Update changelog with `profile_session_sample_rate` (#4046) by @sentrivana ## 2.21.0 From 651e28fefa6d1375027a0f623e6ff7bd0812b111 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 17 Feb 2025 15:25:01 +0100 Subject: [PATCH 148/591] Fixed typo in changelog (#4068) oops... (also changed in on Github) --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index acc018f65c..e6857c34ae 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -6,7 +6,7 @@ - **New integration:** Add [Statsig](https://statsig.com/) integration (#4022) by @aliu39 - For more information, see the documentation for the [TyperIntegration](https://docs.sentry.io/platforms/python/integrations/statsig/). + For more information, see the documentation for the [StatsigIntegration](https://docs.sentry.io/platforms/python/integrations/statsig/). - Profiling: Continuous profiling lifecycle (#4017) by @Zylphrex - Fix: Revert "feat(tracing): Add `propagate_traces` deprecation warning (#3899)" (#4055) by @cmanallen From 74b3bbf9d949e2f2225d4100976baf20098b5e7b Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 18 Feb 2025 10:00:41 +0000 Subject: [PATCH 149/591] build(deps): bump actions/create-github-app-token from 1.11.3 to 1.11.5 (#4059) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.3 to 1.11.5.
Release notes

Sourced from actions/create-github-app-token's releases.

v1.11.5

1.11.5 (2025-02-15)

Bug Fixes

v1.11.4

1.11.4 (2025-02-15)

Bug Fixes

Commits
  • 0d56448 build(release): 1.11.5 [skip ci]
  • 8cedd97 fix(deps): bump @​octokit/request from 9.2.0 to 9.2.2 (#209)
  • 415f6a5 fix(deps): bump @​octokit/request-error from 6.1.6 to 6.1.7 (#208)
  • c14f92a build(release): 1.11.4 [skip ci]
  • d30def8 fix(deps): bump @​octokit/endpoint from 10.1.1 to 10.1.3 (#207)
  • a5be472 build(deps-dev): bump esbuild from 0.24.2 to 0.25.0 (#206)
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=1.11.3&new-version=1.11.5)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index ae9ae279c7..4d8c060f6a 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@67e27a7eb7db372a1c61a7f9bdab8699e9ee57f7 # v1.11.3 + uses: actions/create-github-app-token@0d564482f06ca65fa9e77e2510873638c82206f2 # v1.11.5 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From a5ce968d6542bdd486ab99ce00d756723d804cdc Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Tue, 18 Feb 2025 11:05:39 -0500 Subject: [PATCH 150/591] feat(profiling): Add new functions to start/stop continuous profiler (#4056) The `start_profiler` and `stop_profiler` functions were renamed to `start_profile_session` and `stop_profile_session` respectively. --- sentry_sdk/profiler/continuous_profiler.py | 14 ++++ tests/profiler/test_continuous_profiler.py | 86 ++++++++++++++++++++-- 2 files changed, 92 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 1619925bd2..9e2aa35fc1 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -145,6 +145,13 @@ def try_profile_lifecycle_trace_start(): def start_profiler(): # type: () -> None + + # TODO: deprecate this as it'll be replaced by `start_profile_session` + start_profile_session() + + +def start_profile_session(): + # type: () -> None if _scheduler is None: return @@ -153,6 +160,13 @@ def start_profiler(): def stop_profiler(): # type: () -> None + + # TODO: deprecate this as it'll be replaced by `stop_profile_session` + stop_profile_session() + + +def stop_profile_session(): + # type: () -> None if _scheduler is None: return diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 525616c9a8..78335d7b87 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -11,7 +11,9 @@ get_profiler_id, setup_continuous_profiler, start_profiler, + start_profile_session, stop_profiler, + stop_profile_session, ) from tests.conftest import ApproxDict @@ -207,6 +209,21 @@ def assert_single_transaction_without_profile_chunks(envelopes): pytest.param("gevent", marks=requires_gevent), ], ) +@pytest.mark.parametrize( + ["start_profiler_func", "stop_profiler_func"], + [ + pytest.param( + start_profile_session, + stop_profile_session, + id="start_profile_session/stop_profile_session", + ), + pytest.param( + start_profiler, + stop_profiler, + id="start_profiler/stop_profiler (deprecated)", + ), + ], +) @pytest.mark.parametrize( "make_options", [ @@ -219,6 +236,8 @@ def test_continuous_profiler_auto_start_and_manual_stop( sentry_init, capture_envelopes, mode, + start_profiler_func, + stop_profiler_func, make_options, teardown_profiling, ): @@ -239,7 +258,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( assert_single_transaction_with_profile_chunks(envelopes, thread) for _ in range(3): - stop_profiler() + stop_profiler_func() envelopes.clear() @@ -249,7 +268,7 @@ def test_continuous_profiler_auto_start_and_manual_stop( assert_single_transaction_without_profile_chunks(envelopes) - start_profiler() + start_profiler_func() envelopes.clear() @@ -267,6 +286,21 @@ def test_continuous_profiler_auto_start_and_manual_stop( pytest.param("gevent", marks=requires_gevent), ], ) +@pytest.mark.parametrize( + ["start_profiler_func", "stop_profiler_func"], + [ + pytest.param( + start_profile_session, + stop_profile_session, + id="start_profile_session/stop_profile_session", + ), + pytest.param( + start_profiler, + stop_profiler, + id="start_profiler/stop_profiler (deprecated)", + ), + ], +) @pytest.mark.parametrize( "make_options", [ @@ -279,6 +313,8 @@ def test_continuous_profiler_manual_start_and_stop_sampled( sentry_init, capture_envelopes, mode, + start_profiler_func, + stop_profiler_func, make_options, teardown_profiling, ): @@ -295,7 +331,7 @@ def test_continuous_profiler_manual_start_and_stop_sampled( thread = threading.current_thread() for _ in range(3): - start_profiler() + start_profiler_func() envelopes.clear() @@ -309,7 +345,7 @@ def test_continuous_profiler_manual_start_and_stop_sampled( assert get_profiler_id() is not None, "profiler should be running" - stop_profiler() + stop_profiler_func() # the profiler stops immediately in manual mode assert get_profiler_id() is None, "profiler should not be running" @@ -332,6 +368,21 @@ def test_continuous_profiler_manual_start_and_stop_sampled( pytest.param("gevent", marks=requires_gevent), ], ) +@pytest.mark.parametrize( + ["start_profiler_func", "stop_profiler_func"], + [ + pytest.param( + start_profile_session, + stop_profile_session, + id="start_profile_session/stop_profile_session", + ), + pytest.param( + start_profiler, + stop_profiler, + id="start_profiler/stop_profiler (deprecated)", + ), + ], +) @pytest.mark.parametrize( "make_options", [ @@ -343,6 +394,8 @@ def test_continuous_profiler_manual_start_and_stop_unsampled( sentry_init, capture_envelopes, mode, + start_profiler_func, + stop_profiler_func, make_options, teardown_profiling, ): @@ -356,7 +409,7 @@ def test_continuous_profiler_manual_start_and_stop_unsampled( envelopes = capture_envelopes() - start_profiler() + start_profiler_func() with sentry_sdk.start_transaction(name="profiling"): with sentry_sdk.start_span(op="op"): @@ -364,7 +417,7 @@ def test_continuous_profiler_manual_start_and_stop_unsampled( assert_single_transaction_without_profile_chunks(envelopes) - stop_profiler() + stop_profiler_func() @pytest.mark.parametrize( @@ -485,6 +538,21 @@ def test_continuous_profiler_auto_start_and_stop_unsampled( ), ], ) +@pytest.mark.parametrize( + ["start_profiler_func", "stop_profiler_func"], + [ + pytest.param( + start_profile_session, + stop_profile_session, + id="start_profile_session/stop_profile_session", + ), + pytest.param( + start_profiler, + stop_profiler, + id="start_profiler/stop_profiler (deprecated)", + ), + ], +) @pytest.mark.parametrize( "make_options", [ @@ -495,6 +563,8 @@ def test_continuous_profiler_auto_start_and_stop_unsampled( def test_continuous_profiler_manual_start_and_stop_noop_when_using_trace_lifecyle( sentry_init, mode, + start_profiler_func, + stop_profiler_func, class_name, make_options, teardown_profiling, @@ -510,11 +580,11 @@ def test_continuous_profiler_manual_start_and_stop_noop_when_using_trace_lifecyl with mock.patch( f"sentry_sdk.profiler.continuous_profiler.{class_name}.ensure_running" ) as mock_ensure_running: - start_profiler() + start_profiler_func() mock_ensure_running.assert_not_called() with mock.patch( f"sentry_sdk.profiler.continuous_profiler.{class_name}.teardown" ) as mock_teardown: - stop_profiler() + stop_profiler_func() mock_teardown.assert_not_called() From 3745d9ad43d9cc925a72d98edaf712166cb6a1a1 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 19 Feb 2025 11:53:40 +0100 Subject: [PATCH 151/591] ci: Fix API doc failure in CI (#4075) Sphinx 8.2 (see [changelog](https://www.sphinx-doc.org/en/master/changes/index.html#release-8-2-0-released-feb-18-2025)) seems to have broken our CI. Looks like an incompatibility between it and the autodoc-typehints extension, so hopefully the two catch up with one another -- I'll pin sphinx to <8.2 for now. --- requirements-docs.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements-docs.txt b/requirements-docs.txt index 15f226aac7..81e04ba3ef 100644 --- a/requirements-docs.txt +++ b/requirements-docs.txt @@ -1,5 +1,5 @@ gevent shibuya -sphinx +sphinx<8.2 sphinx-autodoc-typehints[type_comments]>=1.8.0 typing-extensions From 67f04910a4b2d6928d4ea7d39d3ba5aea4f91d28 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 19 Feb 2025 12:09:32 +0100 Subject: [PATCH 152/591] tests: Add `fail_on_changes` to toxgen (#4072) Add `fail_on_changes` to toxgen. The idea is that the script will now have two modes: - **Normal mode** (when `fail_on_changes` is `False`) that is used to actually generate the `tox.ini` file. This [will be](https://github.com/getsentry/sentry-python/issues/4050) run in a cron job in CI and create a PR with the updated test setup. - The newly added **fail-on-changes mode** (when `fail_on_changes` is `True`) that is used to detect manual changes to one of the affected files without updating the rest (e.g. making a manual change to `tox.ini` without updating the `tox.jinja` template). This will be run in CI similar to the `fail_on_changes` check of `split-tox-gh-actions`. The problem with detecting manual changes is that if we just reran the script on each PR, chances are it would pull in new releases that are not part of the `tox.ini` on master, making the file look different from what was committed as if it had unrelated manual changes. To counteract this, we now store the timestamp when the file was last generated in `tox.ini`. We use this in fail-on-changes mode to filter out releases that popped up after the file was last generated. This way, the package versions should be the same and if there is anything different in `tox.ini`, it's likely to be the manual changes that we want to detect. Closes https://github.com/getsentry/sentry-python/issues/4051 --- .github/workflows/ci.yml | 6 +- scripts/populate_tox/populate_tox.py | 127 +++++++++++++++++++++++++-- scripts/populate_tox/tox.jinja | 2 + tox.ini | 6 +- 4 files changed, 130 insertions(+), 11 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index e8931e229e..03ed8de742 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -44,7 +44,11 @@ jobs: with: python-version: 3.12 - - run: | + - name: Detect unexpected changes to tox.ini or CI + run: | + pip install -e . + pip install -r scripts/populate_tox/requirements.txt + python scripts/populate_tox/populate_tox.py --fail-on-changes pip install -r scripts/split_tox_gh_actions/requirements.txt python scripts/split_tox_gh_actions/split_tox_gh_actions.py --fail-on-changes diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 4bfce80ce7..5906eee5b4 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -3,15 +3,18 @@ """ import functools +import hashlib import os import sys import time from bisect import bisect_left from collections import defaultdict +from datetime import datetime, timezone from importlib.metadata import metadata from packaging.specifiers import SpecifierSet from packaging.version import Version from pathlib import Path +from textwrap import dedent from typing import Optional, Union # Adding the scripts directory to PATH. This is necessary in order to be able @@ -106,7 +109,9 @@ def fetch_release(package: str, version: Version) -> dict: return pypi_data.json() -def _prefilter_releases(integration: str, releases: dict[str, dict]) -> list[Version]: +def _prefilter_releases( + integration: str, releases: dict[str, dict], older_than: Optional[datetime] = None +) -> list[Version]: """ Filter `releases`, removing releases that are for sure unsupported. @@ -135,6 +140,10 @@ def _prefilter_releases(integration: str, releases: dict[str, dict]) -> list[Ver if meta["yanked"]: continue + if older_than is not None: + if datetime.fromisoformat(meta["upload_time_iso_8601"]) > older_than: + continue + version = Version(release) if min_supported and version < min_supported: @@ -160,19 +169,24 @@ def _prefilter_releases(integration: str, releases: dict[str, dict]) -> list[Ver return sorted(filtered_releases) -def get_supported_releases(integration: str, pypi_data: dict) -> list[Version]: +def get_supported_releases( + integration: str, pypi_data: dict, older_than: Optional[datetime] = None +) -> list[Version]: """ Get a list of releases that are currently supported by the SDK. This takes into account a handful of parameters (Python support, the lowest version we've defined for the framework, the date of the release). + + If an `older_than` timestamp is provided, no release newer than that will be + considered. """ package = pypi_data["info"]["name"] # Get a consolidated list without taking into account Python support yet # (because that might require an additional API call for some # of the releases) - releases = _prefilter_releases(integration, pypi_data["releases"]) + releases = _prefilter_releases(integration, pypi_data["releases"], older_than) # Determine Python support expected_python_versions = TEST_SUITE_CONFIG[integration].get("python") @@ -381,7 +395,9 @@ def _render_dependencies(integration: str, releases: list[Version]) -> list[str] return rendered -def write_tox_file(packages: dict) -> None: +def write_tox_file( + packages: dict, update_timestamp: bool, last_updated: datetime +) -> None: template = ENV.get_template("tox.jinja") context = {"groups": {}} @@ -400,6 +416,11 @@ def write_tox_file(packages: dict) -> None: } ) + if update_timestamp: + context["updated"] = datetime.now(tz=timezone.utc).isoformat() + else: + context["updated"] = last_updated.isoformat() + rendered = template.render(context) with open(TOX_FILE, "w") as file: @@ -453,7 +474,59 @@ def _add_python_versions_to_release( release.rendered_python_versions = _render_python_versions(release.python_versions) -def main() -> None: +def get_file_hash() -> str: + """Calculate a hash of the tox.ini file.""" + hasher = hashlib.md5() + + with open(TOX_FILE, "rb") as f: + buf = f.read() + hasher.update(buf) + + return hasher.hexdigest() + + +def get_last_updated() -> Optional[datetime]: + timestamp = None + + with open(TOX_FILE, "r") as f: + for line in f: + if line.startswith("# Last generated:"): + timestamp = datetime.fromisoformat(line.strip().split()[-1]) + break + + if timestamp is None: + print( + "Failed to find out when tox.ini was last generated; the timestamp seems to be missing from the file." + ) + + return timestamp + + +def main(fail_on_changes: bool = False) -> None: + """ + Generate tox.ini from the tox.jinja template. + + The script has two modes of operation: + - fail on changes mode (if `fail_on_changes` is True) + - normal mode (if `fail_on_changes` is False) + + Fail on changes mode is run on every PR to make sure that `tox.ini`, + `tox.jinja` and this script don't go out of sync because of manual changes + in one place but not the other. + + Normal mode is meant to be run as a cron job, regenerating tox.ini and + proposing the changes via a PR. + """ + print(f"Running in {'fail_on_changes' if fail_on_changes else 'normal'} mode.") + last_updated = get_last_updated() + if fail_on_changes: + # We need to make the script ignore any new releases after the `last_updated` + # timestamp so that we don't fail CI on a PR just because a new package + # version was released, leading to unrelated changes in tox.ini. + print( + f"Since we're in fail_on_changes mode, we're only considering releases before the last tox.ini update at {last_updated.isoformat()}." + ) + global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION sdk_python_versions = _parse_python_versions_from_classifiers( metadata("sentry-sdk").get_all("Classifier") @@ -480,7 +553,9 @@ def main() -> None: pypi_data = fetch_package(package) # Get the list of all supported releases - releases = get_supported_releases(integration, pypi_data) + # If in check mode, ignore releases newer than `last_updated` + older_than = last_updated if fail_on_changes else None + releases = get_supported_releases(integration, pypi_data, older_than) if not releases: print(" Found no supported releases.") continue @@ -510,8 +585,44 @@ def main() -> None: } ) - write_tox_file(packages) + if fail_on_changes: + old_file_hash = get_file_hash() + + write_tox_file( + packages, update_timestamp=not fail_on_changes, last_updated=last_updated + ) + + if fail_on_changes: + new_file_hash = get_file_hash() + if old_file_hash != new_file_hash: + raise RuntimeError( + dedent( + """ + Detected that `tox.ini` is out of sync with + `scripts/populate_tox/tox.jinja` and/or + `scripts/populate_tox/populate_tox.py`. This might either mean + that `tox.ini` was changed manually, or the `tox.jinja` + template and/or the `populate_tox.py` script were changed without + regenerating `tox.ini`. + + Please don't make manual changes to `tox.ini`. Instead, make the + changes to the `tox.jinja` template and/or the `populate_tox.py` + script (as applicable) and regenerate the `tox.ini` file with: + + python -m venv toxgen.env + . toxgen.env/bin/activate + pip install -r scripts/populate_tox/requirements.txt + python scripts/populate_tox/populate_tox.py + """ + ) + ) + print("Done checking tox.ini. Looking good!") + else: + print( + "Done generating tox.ini. Make sure to also update the CI YAML files to reflect the new test targets." + ) if __name__ == "__main__": - main() + fail_on_changes = len(sys.argv) == 2 and sys.argv[1] == "--fail-on-changes" + main(fail_on_changes) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 15119b4768..81ab17c919 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -9,6 +9,8 @@ # or in the script (if you want to change the auto-generated part). # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". +# +# Last generated: {{ updated }} [tox] requires = diff --git a/tox.ini b/tox.ini index 9ce3d40a21..0e41500fe1 100644 --- a/tox.ini +++ b/tox.ini @@ -9,6 +9,8 @@ # or in the script (if you want to change the auto-generated part). # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". +# +# Last generated: 2025-02-18T12:57:32.874168+00:00 [tox] requires = @@ -290,7 +292,7 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.9 - {py3.8,py3.11,py3.12}-trytond-v7.4.5 + {py3.8,py3.11,py3.12}-trytond-v7.4.6 {py3.7,py3.11,py3.12}-typer-v0.15.1 @@ -714,7 +716,7 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.9: trytond==7.0.9 - trytond-v7.4.5: trytond==7.4.5 + trytond-v7.4.6: trytond==7.4.6 trytond: werkzeug trytond-v4.6.9: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From a3b6e5d9f3adc515548dabd73462e77bccc4d516 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 19 Feb 2025 15:18:54 +0100 Subject: [PATCH 153/591] tests: Test relevant prereleases and allow to ignore releases (#4073) If a package has a prerelease of a higher version than the highest released stable version, make sure to test it, too. We consider alpha, beta, and RC releases. Also add an option to ignore specific releases (this is related to the above since the script now pulls in two irrelevant alpha releases of starlite). Closes https://github.com/getsentry/sentry-python/issues/4030 --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- scripts/populate_tox/README.md | 35 ++++++++++++++ scripts/populate_tox/config.py | 1 + scripts/populate_tox/populate_tox.py | 72 ++++++++++++++++++++++------ tox.ini | 10 ++-- 4 files changed, 101 insertions(+), 17 deletions(-) diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index aa9884387e..c9a3b67ba0 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -45,9 +45,15 @@ integration_name: { rule2: [package3, package4, ...], }, "python": python_version_specifier, + "include": package_version_specifier, } ``` +When talking about version specifiers, we mean +[version specifiers as defined](https://packaging.python.org/en/latest/specifications/version-specifiers/#id5) +by the Python Packaging Authority. See also the actual implementation +in [packaging.specifiers](https://packaging.pypa.io/en/stable/specifiers.html). + ### `package` The name of the third party package as it's listed on PyPI. The script will @@ -118,6 +124,35 @@ metadata or the SDK is explicitly not supporting some packages on specific Python versions (because of, for example, broken context vars), the `python` key can be used. +### `include` + +Sometimes we only want to consider testing some specific versions of packages. +For example, the Starlite package has two alpha prereleases of version 2.0.0, but +we do not want to test these, since Starlite 2.0 was renamed to Litestar. + +The value of the `include` key expects a version specifier defining which +versions should be considered for testing. For example, since we only want to test +versions below 2.x in Starlite, we can use + +```python +"starlite": { + "include": "<2", + ... +} +``` + +The `include` key can also be used to exclude a set of specific versions by using +`!=` version specifiers. For example, the Starlite restriction above could equivalently +be expressed like so: + + +```python +"starlite": { + "include": "!=2.0.0a1,!=2.0.0a2", + ... +} +``` + ## How-Tos diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 2c2920e7ac..b5da928d80 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -129,6 +129,7 @@ ], }, "python": "<=3.11", + "include": "!=2.0.0a1,!=2.0.0a2", # these are not relevant as there will never be a stable 2.0 release (starlite continues as litestar) }, "statsig": { "package": "statsig", diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 5906eee5b4..544d4bdcb1 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -111,7 +111,7 @@ def fetch_release(package: str, version: Version) -> dict: def _prefilter_releases( integration: str, releases: dict[str, dict], older_than: Optional[datetime] = None -) -> list[Version]: +) -> tuple[list[Version], Optional[Version]]: """ Filter `releases`, removing releases that are for sure unsupported. @@ -120,6 +120,10 @@ def _prefilter_releases( they require additional API calls to be made. The purpose of this function is to slim down the list so that we don't have to make more API calls than necessary for releases that are for sure not supported. + + The function returns a tuple with: + - the list of prefiltered releases + - an optional prerelease if there is one that should be tested """ min_supported = _MIN_VERSIONS.get(integration) if min_supported is not None: @@ -129,7 +133,14 @@ def _prefilter_releases( f" {integration} doesn't have a minimum version defined in sentry_sdk/integrations/__init__.py. Consider defining one" ) + include_versions = None + if TEST_SUITE_CONFIG[integration].get("include") is not None: + include_versions = SpecifierSet( + TEST_SUITE_CONFIG[integration]["include"], prereleases=True + ) + filtered_releases = [] + last_prerelease = None for release, data in releases.items(): if not data: @@ -149,9 +160,15 @@ def _prefilter_releases( if min_supported and version < min_supported: continue - if version.is_prerelease or version.is_postrelease: - # TODO: consider the newest prerelease unless obsolete - # https://github.com/getsentry/sentry-python/issues/4030 + if version.is_postrelease or version.is_devrelease: + continue + + if include_versions is not None and version not in include_versions: + continue + + if version.is_prerelease: + if last_prerelease is None or version > last_prerelease: + last_prerelease = version continue for i, saved_version in enumerate(filtered_releases): @@ -166,18 +183,30 @@ def _prefilter_releases( else: filtered_releases.append(version) - return sorted(filtered_releases) + filtered_releases.sort() + + # Check if the latest prerelease is relevant (i.e., it's for a version higher + # than the last released version); if not, don't consider it + if last_prerelease is not None: + if not filtered_releases or last_prerelease > filtered_releases[-1]: + return filtered_releases, last_prerelease + + return filtered_releases, None def get_supported_releases( integration: str, pypi_data: dict, older_than: Optional[datetime] = None -) -> list[Version]: +) -> tuple[list[Version], Optional[Version]]: """ Get a list of releases that are currently supported by the SDK. This takes into account a handful of parameters (Python support, the lowest version we've defined for the framework, the date of the release). + We return the list of supported releases and optionally also the newest + prerelease, if it should be tested (meaning it's for a version higher than + the current stable version). + If an `older_than` timestamp is provided, no release newer than that will be considered. """ @@ -186,7 +215,9 @@ def get_supported_releases( # Get a consolidated list without taking into account Python support yet # (because that might require an additional API call for some # of the releases) - releases = _prefilter_releases(integration, pypi_data["releases"], older_than) + releases, latest_prerelease = _prefilter_releases( + integration, pypi_data["releases"], older_than + ) # Determine Python support expected_python_versions = TEST_SUITE_CONFIG[integration].get("python") @@ -210,14 +241,18 @@ def _supports_lowest(release: Version) -> bool: # version(s) that we do, cut off the rest releases = releases[i:] - return releases + return releases, latest_prerelease -def pick_releases_to_test(releases: list[Version]) -> list[Version]: +def pick_releases_to_test( + releases: list[Version], last_prerelease: Optional[Version] +) -> list[Version]: """Pick a handful of releases to test from a sorted list of supported releases.""" # If the package has majors (or major-like releases, even if they don't do # semver), we want to make sure we're testing them all. If not, we just pick # the oldest, the newest, and a couple in between. + # + # If there is a relevant prerelease, also test that in addition to the above. has_majors = len(set([v.major for v in releases])) > 1 filtered_releases = set() @@ -252,7 +287,11 @@ def pick_releases_to_test(releases: list[Version]) -> list[Version]: releases[-1], # latest } - return sorted(filtered_releases) + filtered_releases = sorted(filtered_releases) + if last_prerelease is not None: + filtered_releases.append(last_prerelease) + + return filtered_releases def supported_python_versions( @@ -553,9 +592,14 @@ def main(fail_on_changes: bool = False) -> None: pypi_data = fetch_package(package) # Get the list of all supported releases - # If in check mode, ignore releases newer than `last_updated` + + # If in fail-on-changes mode, ignore releases newer than `last_updated` older_than = last_updated if fail_on_changes else None - releases = get_supported_releases(integration, pypi_data, older_than) + + releases, latest_prerelease = get_supported_releases( + integration, pypi_data, older_than + ) + if not releases: print(" Found no supported releases.") continue @@ -563,9 +607,9 @@ def main(fail_on_changes: bool = False) -> None: _compare_min_version_with_defined(integration, releases) # Pick a handful of the supported releases to actually test against - # and fetch the PYPI data for each to determine which Python versions + # and fetch the PyPI data for each to determine which Python versions # to test it on - test_releases = pick_releases_to_test(releases) + test_releases = pick_releases_to_test(releases, latest_prerelease) for release in test_releases: _add_python_versions_to_release(integration, package, release) diff --git a/tox.ini b/tox.ini index 0e41500fe1..360d16342e 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-02-18T12:57:32.874168+00:00 +# Last generated: 2025-02-19T12:41:15.689786+00:00 [tox] requires = @@ -211,10 +211,11 @@ envlist = {py3.8,py3.10,py3.11}-ariadne-v0.20.1 {py3.8,py3.11,py3.12}-ariadne-v0.22 {py3.8,py3.11,py3.12}-ariadne-v0.24.0 - {py3.8,py3.11,py3.12}-ariadne-v0.25.2 + {py3.9,py3.12,py3.13}-ariadne-v0.26.0 {py3.6,py3.9,py3.10}-gql-v3.4.1 {py3.7,py3.11,py3.12}-gql-v3.5.0 + {py3.9,py3.12,py3.13}-gql-v3.6.0b4 {py3.6,py3.9,py3.10}-graphene-v3.3 {py3.8,py3.12,py3.13}-graphene-v3.4.3 @@ -236,6 +237,7 @@ envlist = {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.11,py3.12}-celery-v5.4.0 + {py3.8,py3.12,py3.13}-celery-v5.5.0rc4 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -592,13 +594,14 @@ deps = ariadne-v0.20.1: ariadne==0.20.1 ariadne-v0.22: ariadne==0.22 ariadne-v0.24.0: ariadne==0.24.0 - ariadne-v0.25.2: ariadne==0.25.2 + ariadne-v0.26.0: ariadne==0.26.0 ariadne: fastapi ariadne: flask ariadne: httpx gql-v3.4.1: gql[all]==3.4.1 gql-v3.5.0: gql[all]==3.5.0 + gql-v3.6.0b4: gql[all]==3.6.0b4 graphene-v3.3: graphene==3.3 graphene-v3.4.3: graphene==3.4.3 @@ -630,6 +633,7 @@ deps = celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.4.0: celery==5.4.0 + celery-v5.5.0rc4: celery==5.5.0rc4 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 From ccfd3a80da2fc2eacd95222ab0ac1a3cc720150b Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Thu, 20 Feb 2025 07:39:33 -0500 Subject: [PATCH 154/591] feat(profiling): Export start/stop profile session (#4079) Need to export these explicitly so it can be used. --- sentry_sdk/profiler/__init__.py | 13 ++++++++++--- 1 file changed, 10 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/profiler/__init__.py b/sentry_sdk/profiler/__init__.py index 46382cc29d..d8d4e076d5 100644 --- a/sentry_sdk/profiler/__init__.py +++ b/sentry_sdk/profiler/__init__.py @@ -1,4 +1,9 @@ -from sentry_sdk.profiler.continuous_profiler import start_profiler, stop_profiler +from sentry_sdk.profiler.continuous_profiler import ( + start_profile_session, + start_profiler, + stop_profile_session, + stop_profiler, +) from sentry_sdk.profiler.transaction_profiler import ( MAX_PROFILE_DURATION_NS, PROFILE_MINIMUM_SAMPLES, @@ -20,8 +25,10 @@ ) __all__ = [ - "start_profiler", - "stop_profiler", + "start_profile_session", + "start_profiler", # TODO: Deprecate this in favor of `start_profile_session` + "stop_profile_session", + "stop_profiler", # TODO: Deprecate this in favor of `stop_profile_session` # DEPRECATED: The following was re-exported for backwards compatibility. It # will be removed from sentry_sdk.profiler in a future release. "MAX_PROFILE_DURATION_NS", From 4d64c4e7221ad48b2316c2a45dec57c6c4660402 Mon Sep 17 00:00:00 2001 From: Sviatoslav Abakumov Date: Thu, 20 Feb 2025 16:42:08 +0400 Subject: [PATCH 155/591] fix(typing): Add more typing info to Scope.update_from_kwargs's "contexts" (#4080) The original type hint could be understood as a one-level `dict` of `str` to `Any`, when in fact, it's a two-level dict. --- sentry_sdk/scope.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 4e3bb87489..fbe97ddf44 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -1568,7 +1568,7 @@ def update_from_kwargs( user=None, # type: Optional[Any] level=None, # type: Optional[LogLevelStr] extras=None, # type: Optional[Dict[str, Any]] - contexts=None, # type: Optional[Dict[str, Any]] + contexts=None, # type: Optional[Dict[str, Dict[str, Any]]] tags=None, # type: Optional[Dict[str, str]] fingerprint=None, # type: Optional[List[str]] ): From 24232993da9f1364e0064d155dfe7006ee9b74c2 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Bj=C3=B6rn=20Friedrichs?= <2217052+itsbjoern@users.noreply.github.com> Date: Thu, 20 Feb 2025 13:38:17 +0000 Subject: [PATCH 156/591] AWS Lambda: Fix capturing errors during AWS Lambda INIT phase (#3943) The AWS integration fails to capture errors during the INIT phase (at least in Python 3.8 and above environments). It appears tests for this were disabled after a change in AWS' own runtime environment: https://github.com/getsentry/sentry-python/pull/3592 A change from a few months ago where it seems like string serialisation of the JSON payload was disabled and instead the `post_init_error` is invoked directly with the json payload: https://github.com/aws/aws-lambda-python-runtime-interface-client/commit/a37a43a48bc151c211ad72a6556044aa62b2c671#diff-4513a869520b19ae4e30058106d7c3b5ddbb79216b5e9bd922d83389fb86c603R483 This breaks and causes an error internally when trying to parse the string back into json, and the error is actually swallowed because of `with capture_internal_exceptions()`. Co-authored-by: Anton Pirker --- sentry_sdk/integrations/aws_lambda.py | 5 ++++- tests/integrations/aws_lambda/test_aws.py | 3 --- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/aws_lambda.py b/sentry_sdk/integrations/aws_lambda.py index 831cde8999..c232094256 100644 --- a/sentry_sdk/integrations/aws_lambda.py +++ b/sentry_sdk/integrations/aws_lambda.py @@ -61,7 +61,10 @@ def sentry_init_error(*args, **kwargs): else: # Fall back to AWS lambdas JSON representation of the error - sentry_event = _event_from_error_json(json.loads(args[1])) + error_info = args[1] + if isinstance(error_info, str): + error_info = json.loads(error_info) + sentry_event = _event_from_error_json(error_info) sentry_sdk.capture_event(sentry_event) return init_error(*args, **kwargs) diff --git a/tests/integrations/aws_lambda/test_aws.py b/tests/integrations/aws_lambda/test_aws.py index f60bedc846..8bbd33505b 100644 --- a/tests/integrations/aws_lambda/test_aws.py +++ b/tests/integrations/aws_lambda/test_aws.py @@ -316,9 +316,6 @@ def test_handler(event, context): } -@pytest.mark.xfail( - reason="Amazon changed something (2024-10-01) and on Python 3.9+ our SDK can not capture events in the init phase of the Lambda function anymore. We need to fix this somehow." -) def test_init_error(run_lambda_function, lambda_runtime): envelope_items, _ = run_lambda_function( LAMBDA_PRELUDE From 48ebd7321c6fb2fcc9ddbd2039b1211114532768 Mon Sep 17 00:00:00 2001 From: Nathan Date: Thu, 20 Feb 2025 15:56:22 +0000 Subject: [PATCH 157/591] fix(anthropic): Add partial json support to streams (#3674) Add `partial_json` for tool calling when streaming in Anthropic integrations. (This is an addition to https://github.com/getsentry/sentry-python/pull/3615 --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/anthropic.py | 2 + .../integrations/anthropic/test_anthropic.py | 71 +++++++++++++++++-- 2 files changed, 66 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index f06d8a14db..4cb54309c8 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -101,6 +101,8 @@ def _collect_ai_data(event, input_tokens, output_tokens, content_blocks): elif event.type == "content_block_delta": if hasattr(event.delta, "text"): content_blocks.append(event.delta.text) + elif hasattr(event.delta, "partial_json"): + content_blocks.append(event.delta.partial_json) elif event.type == "content_block_stop": pass elif event.type == "message_delta": diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index 8ce12e70f5..7f6622a1ba 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -1,5 +1,6 @@ from unittest import mock + try: from unittest.mock import AsyncMock except ImportError: @@ -10,7 +11,7 @@ async def __call__(self, *args, **kwargs): import pytest -from anthropic import AsyncAnthropic, Anthropic, AnthropicError, AsyncStream, Stream +from anthropic import Anthropic, AnthropicError, AsyncAnthropic, AsyncStream, Stream from anthropic.types import MessageDeltaUsage, TextDelta, Usage from anthropic.types.content_block_delta_event import ContentBlockDeltaEvent from anthropic.types.content_block_start_event import ContentBlockStartEvent @@ -19,6 +20,7 @@ async def __call__(self, *args, **kwargs): from anthropic.types.message_delta_event import MessageDeltaEvent from anthropic.types.message_start_event import MessageStartEvent +from sentry_sdk.integrations.anthropic import _add_ai_data_to_span, _collect_ai_data from sentry_sdk.utils import package_version try: @@ -42,7 +44,7 @@ async def __call__(self, *args, **kwargs): except ImportError: from anthropic.types.content_block import ContentBlock as TextBlock -from sentry_sdk import start_transaction +from sentry_sdk import start_transaction, start_span from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations.anthropic import AnthropicIntegration @@ -517,9 +519,8 @@ def test_streaming_create_message_with_input_json_delta( if send_default_pii and include_prompts: assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"text": "", "type": "text"} - ] # we do not record InputJSONDelta because it could contain PII - + {"text": "{'location': 'San Francisco, CA'}", "type": "text"} + ] else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] @@ -654,8 +655,8 @@ async def test_streaming_create_message_with_input_json_delta_async( if send_default_pii and include_prompts: assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"text": "", "type": "text"} - ] # we do not record InputJSONDelta because it could contain PII + {"text": "{'location': 'San Francisco, CA'}", "type": "text"} + ] else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] @@ -757,3 +758,59 @@ async def test_span_origin_async(sentry_init, capture_events): assert event["contexts"]["trace"]["origin"] == "manual" assert event["spans"][0]["origin"] == "auto.ai.anthropic" + + +@pytest.mark.skipif( + ANTHROPIC_VERSION < (0, 27), + reason="Versions <0.27.0 do not include InputJSONDelta.", +) +def test_collect_ai_data_with_input_json_delta(): + event = ContentBlockDeltaEvent( + delta=InputJSONDelta(partial_json="test", type="input_json_delta"), + index=0, + type="content_block_delta", + ) + + input_tokens = 10 + output_tokens = 20 + content_blocks = [] + + new_input_tokens, new_output_tokens, new_content_blocks = _collect_ai_data( + event, input_tokens, output_tokens, content_blocks + ) + + assert new_input_tokens == input_tokens + assert new_output_tokens == output_tokens + assert new_content_blocks == ["test"] + + +@pytest.mark.skipif( + ANTHROPIC_VERSION < (0, 27), + reason="Versions <0.27.0 do not include InputJSONDelta.", +) +def test_add_ai_data_to_span_with_input_json_delta(sentry_init): + sentry_init( + integrations=[AnthropicIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with start_transaction(name="test"): + span = start_span() + integration = AnthropicIntegration() + + _add_ai_data_to_span( + span, + integration, + input_tokens=10, + output_tokens=20, + content_blocks=["{'test': 'data',", "'more': 'json'}"], + ) + + assert span._data.get(SPANDATA.AI_RESPONSES) == [ + {"type": "text", "text": "{'test': 'data','more': 'json'}"} + ] + assert span._data.get("ai.streaming") is True + assert span._measurements.get("ai_prompt_tokens_used")["value"] == 10 + assert span._measurements.get("ai_completion_tokens_used")["value"] == 20 + assert span._measurements.get("ai_total_tokens_used")["value"] == 30 From c557b56d7c7d0d256f59567a2a2a1e9c701aa44f Mon Sep 17 00:00:00 2001 From: Andrew Liu <159852527+aliu39@users.noreply.github.com> Date: Fri, 21 Feb 2025 13:32:29 -0800 Subject: [PATCH 158/591] ref(flags): add LRU update/dedupe test coverage (#4082) --- tests/test_feature_flags.py | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) diff --git a/tests/test_feature_flags.py b/tests/test_feature_flags.py index 4469b5c2ca..0df30bd0ea 100644 --- a/tests/test_feature_flags.py +++ b/tests/test_feature_flags.py @@ -170,6 +170,25 @@ def test_flag_tracking(): {"flag": "f", "result": False}, ] + # Test updates + buffer.set("e", True) + buffer.set("e", False) + buffer.set("e", True) + flags = buffer.get() + assert flags == [ + {"flag": "d", "result": False}, + {"flag": "f", "result": False}, + {"flag": "e", "result": True}, + ] + + buffer.set("d", True) + flags = buffer.get() + assert flags == [ + {"flag": "f", "result": False}, + {"flag": "e", "result": True}, + {"flag": "d", "result": True}, + ] + def test_flag_buffer_concurrent_access(): buffer = FlagBuffer(capacity=100) From eeedd11c1b0908c8bc68f999433b625508d979fa Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 24 Feb 2025 10:13:11 +0100 Subject: [PATCH 159/591] Fix ClickHouse in test suite (#4087) Use new version of the ClickHouse Github action. This works with newest ClickHouse and also now prints ClickHouse details. --- .github/workflows/test-integrations-dbs.yml | 6 ++++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 3 ++- 2 files changed, 6 insertions(+), 3 deletions(-) diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index d525e353ed..1fb0aa0715 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -59,7 +59,8 @@ jobs: with: python-version: ${{ matrix.python-version }} allow-prereleases: true - - uses: getsentry/action-clickhouse-in-ci@v1.1 + - name: "Setup ClickHouse Server" + uses: getsentry/action-clickhouse-in-ci@v1.5 - name: Setup Test Env run: | pip install "coverage[toml]" tox @@ -154,7 +155,8 @@ jobs: with: python-version: ${{ matrix.python-version }} allow-prereleases: true - - uses: getsentry/action-clickhouse-in-ci@v1.1 + - name: "Setup ClickHouse Server" + uses: getsentry/action-clickhouse-in-ci@v1.5 - name: Setup Test Env run: | pip install "coverage[toml]" tox diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 66e346511d..01f9cd56ec 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -51,7 +51,8 @@ python-version: {% raw %}${{ matrix.python-version }}{% endraw %} allow-prereleases: true {% if needs_clickhouse %} - - uses: getsentry/action-clickhouse-in-ci@v1.1 + - name: "Setup ClickHouse Server" + uses: getsentry/action-clickhouse-in-ci@v1.5 {% endif %} {% if needs_redis %} From 189e4a912ef922f400ef422d0827deac1fe1bab5 Mon Sep 17 00:00:00 2001 From: Marcelo Galigniana Date: Mon, 24 Feb 2025 06:29:15 -0300 Subject: [PATCH 160/591] ref(tracing): Move `TRANSACTION_SOURCE_*` constants to `Enum` (#3889) Change the `TRANSACTION_SOURCE_*` constants defined in `tracing.py` to be enums, for better developer experience. Fixes GH-2696 --------- Co-authored-by: Anton Pirker --- CHANGELOG.md | 8 ++-- sentry_sdk/integrations/aiohttp.py | 4 +- sentry_sdk/integrations/arq.py | 4 +- sentry_sdk/integrations/asgi.py | 17 ++++----- sentry_sdk/integrations/aws_lambda.py | 4 +- sentry_sdk/integrations/celery/__init__.py | 4 +- sentry_sdk/integrations/chalice.py | 4 +- sentry_sdk/integrations/django/__init__.py | 4 +- sentry_sdk/integrations/fastapi.py | 4 +- sentry_sdk/integrations/gcp.py | 4 +- sentry_sdk/integrations/grpc/aio/server.py | 4 +- sentry_sdk/integrations/grpc/server.py | 4 +- sentry_sdk/integrations/huey.py | 4 +- sentry_sdk/integrations/litestar.py | 4 +- sentry_sdk/integrations/ray.py | 4 +- sentry_sdk/integrations/rq.py | 4 +- sentry_sdk/integrations/sanic.py | 10 ++--- sentry_sdk/integrations/starlette.py | 9 ++--- sentry_sdk/integrations/starlite.py | 4 +- sentry_sdk/integrations/strawberry.py | 4 +- sentry_sdk/integrations/tornado.py | 9 ++--- sentry_sdk/integrations/wsgi.py | 4 +- sentry_sdk/metrics.py | 15 +++----- sentry_sdk/tracing.py | 43 +++++++++++++--------- tests/integrations/asgi/test_asgi.py | 5 ++- tests/integrations/sanic/test_sanic.py | 8 ++-- tests/test_metrics.py | 6 +-- 27 files changed, 99 insertions(+), 99 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index e6857c34ae..939a612bc0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2328,7 +2328,7 @@ By: @mgaligniana (#1773) import sentry_sdk from sentry_sdk.integrations.arq import ArqIntegration - from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT + from sentry_sdk.tracing import TransactionSource sentry_sdk.init( dsn="...", @@ -2348,7 +2348,7 @@ By: @mgaligniana (#1773) await ctx['session'].aclose() async def main(): - with sentry_sdk.start_transaction(name="testing_arq_tasks", source=TRANSACTION_SOURCE_COMPONENT): + with sentry_sdk.start_transaction(name="testing_arq_tasks", source=TransactionSource.COMPONENT): redis = await create_pool(RedisSettings()) for url in ('https://facebook.com/', 'https://microsoft.com/', 'https://github.com/', "asdf" ): @@ -2422,7 +2422,7 @@ By: @mgaligniana (#1773) import sentry_sdk from sentry_sdk.integrations.huey import HueyIntegration - from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT, Transaction + from sentry_sdk.tracing import TransactionSource, Transaction def main(): @@ -2434,7 +2434,7 @@ By: @mgaligniana (#1773) traces_sample_rate=1.0, ) - with sentry_sdk.start_transaction(name="testing_huey_tasks", source=TRANSACTION_SOURCE_COMPONENT): + with sentry_sdk.start_transaction(name="testing_huey_tasks", source=TransactionSource.COMPONENT): r = add_numbers(1, 2) if __name__ == "__main__": diff --git a/sentry_sdk/integrations/aiohttp.py b/sentry_sdk/integrations/aiohttp.py index 47c1272ae1..ad3202bf2c 100644 --- a/sentry_sdk/integrations/aiohttp.py +++ b/sentry_sdk/integrations/aiohttp.py @@ -20,7 +20,7 @@ from sentry_sdk.tracing import ( BAGGAGE_HEADER_NAME, SOURCE_FOR_STYLE, - TRANSACTION_SOURCE_ROUTE, + TransactionSource, ) from sentry_sdk.tracing_utils import should_propagate_trace from sentry_sdk.utils import ( @@ -129,7 +129,7 @@ async def sentry_app_handle(self, request, *args, **kwargs): # If this transaction name makes it to the UI, AIOHTTP's # URL resolver did not find a route or died trying. name="generic AIOHTTP request", - source=TRANSACTION_SOURCE_ROUTE, + source=TransactionSource.ROUTE, origin=AioHttpIntegration.origin, ) with sentry_sdk.start_transaction( diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index a2cce8e0ff..c356347dad 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -5,7 +5,7 @@ from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_TASK +from sentry_sdk.tracing import Transaction, TransactionSource from sentry_sdk.utils import ( capture_internal_exceptions, ensure_integration_enabled, @@ -102,7 +102,7 @@ async def _sentry_run_job(self, job_id, score): name="unknown arq task", status="ok", op=OP.QUEUE_TASK_ARQ, - source=TRANSACTION_SOURCE_TASK, + source=TransactionSource.TASK, origin=ArqIntegration.origin, ) diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index f5e8665b4f..733aa2b3fe 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -25,10 +25,7 @@ from sentry_sdk.sessions import track_session from sentry_sdk.tracing import ( SOURCE_FOR_STYLE, - TRANSACTION_SOURCE_ROUTE, - TRANSACTION_SOURCE_URL, - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_CUSTOM, + TransactionSource, ) from sentry_sdk.utils import ( ContextVar, @@ -273,9 +270,9 @@ def event_processor(self, event, hint, asgi_scope): already_set = event["transaction"] != _DEFAULT_TRANSACTION_NAME and event[ "transaction_info" ].get("source") in [ - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_ROUTE, - TRANSACTION_SOURCE_CUSTOM, + TransactionSource.COMPONENT, + TransactionSource.ROUTE, + TransactionSource.CUSTOM, ] if not already_set: name, source = self._get_transaction_name_and_source( @@ -313,7 +310,7 @@ def _get_transaction_name_and_source(self, transaction_style, asgi_scope): name = transaction_from_function(endpoint) or "" else: name = _get_url(https://github.com/asgi_scope,%20%22http%22%20if%20ty%20==%20%22http%22%20else%20%22ws%22,%20host=None) - source = TRANSACTION_SOURCE_URL + source = TransactionSource.URL elif transaction_style == "url": # FastAPI includes the route object in the scope to let Sentry extract the @@ -325,11 +322,11 @@ def _get_transaction_name_and_source(self, transaction_style, asgi_scope): name = path else: name = _get_url(https://github.com/asgi_scope,%20%22http%22%20if%20ty%20==%20%22http%22%20else%20%22ws%22,%20host=None) - source = TRANSACTION_SOURCE_URL + source = TransactionSource.URL if name is None: name = _DEFAULT_TRANSACTION_NAME - source = TRANSACTION_SOURCE_ROUTE + source = TransactionSource.ROUTE return name, source return name, source diff --git a/sentry_sdk/integrations/aws_lambda.py b/sentry_sdk/integrations/aws_lambda.py index c232094256..4990fd6e6a 100644 --- a/sentry_sdk/integrations/aws_lambda.py +++ b/sentry_sdk/integrations/aws_lambda.py @@ -10,7 +10,7 @@ from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( AnnotatedValue, capture_internal_exceptions, @@ -153,7 +153,7 @@ def sentry_handler(aws_event, aws_context, *args, **kwargs): headers, op=OP.FUNCTION_AWS, name=aws_context.function_name, - source=TRANSACTION_SOURCE_COMPONENT, + source=TransactionSource.COMPONENT, origin=AwsLambdaIntegration.origin, ) with sentry_sdk.start_transaction( diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index dc48aac0e6..e8811d767e 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -14,7 +14,7 @@ ) from sentry_sdk.integrations.celery.utils import _now_seconds_since_epoch from sentry_sdk.integrations.logging import ignore_logger -from sentry_sdk.tracing import BAGGAGE_HEADER_NAME, TRANSACTION_SOURCE_TASK +from sentry_sdk.tracing import BAGGAGE_HEADER_NAME, TransactionSource from sentry_sdk.tracing_utils import Baggage from sentry_sdk.utils import ( capture_internal_exceptions, @@ -319,7 +319,7 @@ def _inner(*args, **kwargs): headers, op=OP.QUEUE_TASK_CELERY, name="unknown celery task", - source=TRANSACTION_SOURCE_TASK, + source=TransactionSource.TASK, origin=CeleryIntegration.origin, ) transaction.name = task.name diff --git a/sentry_sdk/integrations/chalice.py b/sentry_sdk/integrations/chalice.py index 0754d1f13b..947e41ebf7 100644 --- a/sentry_sdk/integrations/chalice.py +++ b/sentry_sdk/integrations/chalice.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.integrations.aws_lambda import _make_request_event_processor -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, @@ -67,7 +67,7 @@ def wrapped_view_function(**function_args): configured_time = app.lambda_context.get_remaining_time_in_millis() scope.set_transaction_name( app.lambda_context.function_name, - source=TRANSACTION_SOURCE_COMPONENT, + source=TransactionSource.COMPONENT, ) scope.add_event_processor( diff --git a/sentry_sdk/integrations/django/__init__.py b/sentry_sdk/integrations/django/__init__.py index 54bc25675d..a9477d9954 100644 --- a/sentry_sdk/integrations/django/__init__.py +++ b/sentry_sdk/integrations/django/__init__.py @@ -8,7 +8,7 @@ from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.scope import add_global_event_processor, should_send_default_pii from sentry_sdk.serializer import add_global_repr_processor -from sentry_sdk.tracing import SOURCE_FOR_STYLE, TRANSACTION_SOURCE_URL +from sentry_sdk.tracing import SOURCE_FOR_STYLE, TransactionSource from sentry_sdk.tracing_utils import add_query_source, record_sql_queries from sentry_sdk.utils import ( AnnotatedValue, @@ -398,7 +398,7 @@ def _set_transaction_name_and_source(scope, transaction_style, request): if transaction_name is None: transaction_name = request.path_info - source = TRANSACTION_SOURCE_URL + source = TransactionSource.URL else: source = SOURCE_FOR_STYLE[transaction_style] diff --git a/sentry_sdk/integrations/fastapi.py b/sentry_sdk/integrations/fastapi.py index 8877925a36..76c6adee0f 100644 --- a/sentry_sdk/integrations/fastapi.py +++ b/sentry_sdk/integrations/fastapi.py @@ -5,7 +5,7 @@ import sentry_sdk from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import SOURCE_FOR_STYLE, TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing import SOURCE_FOR_STYLE, TransactionSource from sentry_sdk.utils import ( transaction_from_function, logger, @@ -61,7 +61,7 @@ def _set_transaction_name_and_source(scope, transaction_style, request): if not name: name = _DEFAULT_TRANSACTION_NAME - source = TRANSACTION_SOURCE_ROUTE + source = TransactionSource.ROUTE else: source = SOURCE_FOR_STYLE[transaction_style] diff --git a/sentry_sdk/integrations/gcp.py b/sentry_sdk/integrations/gcp.py index 3983f550d3..c637b7414a 100644 --- a/sentry_sdk/integrations/gcp.py +++ b/sentry_sdk/integrations/gcp.py @@ -10,7 +10,7 @@ from sentry_sdk.integrations import Integration from sentry_sdk.integrations._wsgi_common import _filter_headers from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( AnnotatedValue, capture_internal_exceptions, @@ -88,7 +88,7 @@ def sentry_func(functionhandler, gcp_event, *args, **kwargs): headers, op=OP.FUNCTION_GCP, name=environ.get("FUNCTION_NAME", ""), - source=TRANSACTION_SOURCE_COMPONENT, + source=TransactionSource.COMPONENT, origin=GcpIntegration.origin, ) sampling_context = { diff --git a/sentry_sdk/integrations/grpc/aio/server.py b/sentry_sdk/integrations/grpc/aio/server.py index addc6bee36..381c63103e 100644 --- a/sentry_sdk/integrations/grpc/aio/server.py +++ b/sentry_sdk/integrations/grpc/aio/server.py @@ -2,7 +2,7 @@ from sentry_sdk.consts import OP from sentry_sdk.integrations import DidNotEnable from sentry_sdk.integrations.grpc.consts import SPAN_ORIGIN -from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_CUSTOM +from sentry_sdk.tracing import Transaction, TransactionSource from sentry_sdk.utils import event_from_exception from typing import TYPE_CHECKING @@ -48,7 +48,7 @@ async def wrapped(request, context): dict(context.invocation_metadata()), op=OP.GRPC_SERVER, name=name, - source=TRANSACTION_SOURCE_CUSTOM, + source=TransactionSource.CUSTOM, origin=SPAN_ORIGIN, ) diff --git a/sentry_sdk/integrations/grpc/server.py b/sentry_sdk/integrations/grpc/server.py index a640df5e11..0d2792d1b7 100644 --- a/sentry_sdk/integrations/grpc/server.py +++ b/sentry_sdk/integrations/grpc/server.py @@ -2,7 +2,7 @@ from sentry_sdk.consts import OP from sentry_sdk.integrations import DidNotEnable from sentry_sdk.integrations.grpc.consts import SPAN_ORIGIN -from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_CUSTOM +from sentry_sdk.tracing import Transaction, TransactionSource from typing import TYPE_CHECKING @@ -42,7 +42,7 @@ def behavior(request, context): metadata, op=OP.GRPC_SERVER, name=name, - source=TRANSACTION_SOURCE_CUSTOM, + source=TransactionSource.CUSTOM, origin=SPAN_ORIGIN, ) diff --git a/sentry_sdk/integrations/huey.py b/sentry_sdk/integrations/huey.py index 7db57680f6..f0aff4c0dd 100644 --- a/sentry_sdk/integrations/huey.py +++ b/sentry_sdk/integrations/huey.py @@ -9,7 +9,7 @@ from sentry_sdk.tracing import ( BAGGAGE_HEADER_NAME, SENTRY_TRACE_HEADER_NAME, - TRANSACTION_SOURCE_TASK, + TransactionSource, ) from sentry_sdk.utils import ( capture_internal_exceptions, @@ -159,7 +159,7 @@ def _sentry_execute(self, task, timestamp=None): sentry_headers or {}, name=task.name, op=OP.QUEUE_TASK_HUEY, - source=TRANSACTION_SOURCE_TASK, + source=TransactionSource.TASK, origin=HueyIntegration.origin, ) transaction.set_status(SPANSTATUS.OK) diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 841c8a5cce..5f0b32b04e 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -9,7 +9,7 @@ from sentry_sdk.integrations.asgi import SentryAsgiMiddleware from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import SOURCE_FOR_STYLE, TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing import TransactionSource, SOURCE_FOR_STYLE from sentry_sdk.utils import ( ensure_integration_enabled, event_from_exception, @@ -249,7 +249,7 @@ def event_processor(event, _): if not tx_name: tx_name = _DEFAULT_TRANSACTION_NAME - tx_info = {"source": TRANSACTION_SOURCE_ROUTE} + tx_info = {"source": TransactionSource.ROUTE} event.update( { diff --git a/sentry_sdk/integrations/ray.py b/sentry_sdk/integrations/ray.py index 24a28c307f..0842b92265 100644 --- a/sentry_sdk/integrations/ray.py +++ b/sentry_sdk/integrations/ray.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANSTATUS from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration -from sentry_sdk.tracing import TRANSACTION_SOURCE_TASK +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( event_from_exception, logger, @@ -63,7 +63,7 @@ def _f(*f_args, _tracing=None, **f_kwargs): op=OP.QUEUE_TASK_RAY, name=qualname_from_function(f), origin=RayIntegration.origin, - source=TRANSACTION_SOURCE_TASK, + source=TransactionSource.TASK, ) with sentry_sdk.start_transaction(transaction) as transaction: diff --git a/sentry_sdk/integrations/rq.py b/sentry_sdk/integrations/rq.py index d4fca6a33b..6d7fcf723b 100644 --- a/sentry_sdk/integrations/rq.py +++ b/sentry_sdk/integrations/rq.py @@ -5,7 +5,7 @@ from sentry_sdk.api import continue_trace from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.integrations.logging import ignore_logger -from sentry_sdk.tracing import TRANSACTION_SOURCE_TASK +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( capture_internal_exceptions, ensure_integration_enabled, @@ -57,7 +57,7 @@ def sentry_patched_perform_job(self, job, *args, **kwargs): job.meta.get("_sentry_trace_headers") or {}, op=OP.QUEUE_TASK_RQ, name="unknown RQ task", - source=TRANSACTION_SOURCE_TASK, + source=TransactionSource.TASK, origin=RqIntegration.origin, ) diff --git a/sentry_sdk/integrations/sanic.py b/sentry_sdk/integrations/sanic.py index dfcc299d42..bd8f1f329b 100644 --- a/sentry_sdk/integrations/sanic.py +++ b/sentry_sdk/integrations/sanic.py @@ -9,7 +9,7 @@ from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations._wsgi_common import RequestExtractor, _filter_headers from sentry_sdk.integrations.logging import ignore_logger -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT, TRANSACTION_SOURCE_URL +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( capture_internal_exceptions, ensure_integration_enabled, @@ -192,7 +192,7 @@ async def _context_enter(request): op=OP.HTTP_SERVER, # Unless the request results in a 404 error, the name and source will get overwritten in _set_transaction name=request.path, - source=TRANSACTION_SOURCE_URL, + source=TransactionSource.URL, origin=SanicIntegration.origin, ) request.ctx._sentry_transaction = sentry_sdk.start_transaction( @@ -229,7 +229,7 @@ async def _set_transaction(request, route, **_): with capture_internal_exceptions(): scope = sentry_sdk.get_current_scope() route_name = route.name.replace(request.app.name, "").strip(".") - scope.set_transaction_name(route_name, source=TRANSACTION_SOURCE_COMPONENT) + scope.set_transaction_name(route_name, source=TransactionSource.COMPONENT) def _sentry_error_handler_lookup(self, exception, *args, **kwargs): @@ -304,11 +304,11 @@ def _legacy_router_get(self, *args): sanic_route = sanic_route[len(sanic_app_name) + 1 :] scope.set_transaction_name( - sanic_route, source=TRANSACTION_SOURCE_COMPONENT + sanic_route, source=TransactionSource.COMPONENT ) else: scope.set_transaction_name( - rv[0].__name__, source=TRANSACTION_SOURCE_COMPONENT + rv[0].__name__, source=TransactionSource.COMPONENT ) return rv diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index d9db8bd6b8..687a428203 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -21,8 +21,7 @@ from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing import ( SOURCE_FOR_STYLE, - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_ROUTE, + TransactionSource, ) from sentry_sdk.utils import ( AnnotatedValue, @@ -714,7 +713,7 @@ def _set_transaction_name_and_source(scope, transaction_style, request): if name is None: name = _DEFAULT_TRANSACTION_NAME - source = TRANSACTION_SOURCE_ROUTE + source = TransactionSource.ROUTE scope.set_transaction_name(name, source=source) logger.debug( @@ -729,9 +728,9 @@ def _get_transaction_from_middleware(app, asgi_scope, integration): if integration.transaction_style == "endpoint": name = transaction_from_function(app.__class__) - source = TRANSACTION_SOURCE_COMPONENT + source = TransactionSource.COMPONENT elif integration.transaction_style == "url": name = _transaction_name_from_router(asgi_scope) - source = TRANSACTION_SOURCE_ROUTE + source = TransactionSource.ROUTE return name, source diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index 8714ee2f08..24707a18b1 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -3,7 +3,7 @@ from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.integrations.asgi import SentryAsgiMiddleware from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import SOURCE_FOR_STYLE, TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing import SOURCE_FOR_STYLE, TransactionSource from sentry_sdk.utils import ( ensure_integration_enabled, event_from_exception, @@ -235,7 +235,7 @@ def event_processor(event, _): if not tx_name: tx_name = _DEFAULT_TRANSACTION_NAME - tx_info = {"source": TRANSACTION_SOURCE_ROUTE} + tx_info = {"source": TransactionSource.ROUTE} event.update( { diff --git a/sentry_sdk/integrations/strawberry.py b/sentry_sdk/integrations/strawberry.py index f12019cd60..ae7d273079 100644 --- a/sentry_sdk/integrations/strawberry.py +++ b/sentry_sdk/integrations/strawberry.py @@ -7,7 +7,7 @@ from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations.logging import ignore_logger from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( capture_internal_exceptions, ensure_integration_enabled, @@ -208,7 +208,7 @@ def on_operation(self): transaction = self.graphql_span.containing_transaction if transaction and self.execution_context.operation_name: transaction.name = self.execution_context.operation_name - transaction.source = TRANSACTION_SOURCE_COMPONENT + transaction.source = TransactionSource.COMPONENT transaction.op = op self.graphql_span.finish() diff --git a/sentry_sdk/integrations/tornado.py b/sentry_sdk/integrations/tornado.py index 0f0f64d1a1..3cd087524a 100644 --- a/sentry_sdk/integrations/tornado.py +++ b/sentry_sdk/integrations/tornado.py @@ -6,10 +6,7 @@ from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import ( - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_ROUTE, -) +from sentry_sdk.tracing import TransactionSource from sentry_sdk.utils import ( HAS_REAL_CONTEXTVARS, CONTEXTVARS_ERROR_MESSAGE, @@ -122,7 +119,7 @@ def _handle_request_impl(self): # sentry_urldispatcher_resolve is responsible for # setting a transaction name later. name="generic Tornado request", - source=TRANSACTION_SOURCE_ROUTE, + source=TransactionSource.ROUTE, origin=TornadoIntegration.origin, ) @@ -160,7 +157,7 @@ def tornado_processor(event, hint): with capture_internal_exceptions(): method = getattr(handler, handler.request.method.lower()) event["transaction"] = transaction_from_function(method) or "" - event["transaction_info"] = {"source": TRANSACTION_SOURCE_COMPONENT} + event["transaction_info"] = {"source": TransactionSource.COMPONENT} with capture_internal_exceptions(): extractor = TornadoRequestExtractor(request) diff --git a/sentry_sdk/integrations/wsgi.py b/sentry_sdk/integrations/wsgi.py index 50deae10c5..e628e50e69 100644 --- a/sentry_sdk/integrations/wsgi.py +++ b/sentry_sdk/integrations/wsgi.py @@ -13,7 +13,7 @@ ) from sentry_sdk.sessions import track_session from sentry_sdk.scope import use_isolation_scope -from sentry_sdk.tracing import Transaction, TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing import Transaction, TransactionSource from sentry_sdk.utils import ( ContextVar, capture_internal_exceptions, @@ -115,7 +115,7 @@ def __call__(self, environ, start_response): environ, op=OP.HTTP_SERVER, name="generic WSGI request", - source=TRANSACTION_SOURCE_ROUTE, + source=TransactionSource.ROUTE, origin=self.span_origin, ) diff --git a/sentry_sdk/metrics.py b/sentry_sdk/metrics.py index f6e9fd6bde..4bdbc62253 100644 --- a/sentry_sdk/metrics.py +++ b/sentry_sdk/metrics.py @@ -22,12 +22,7 @@ json_dumps, ) from sentry_sdk.envelope import Envelope, Item -from sentry_sdk.tracing import ( - TRANSACTION_SOURCE_ROUTE, - TRANSACTION_SOURCE_VIEW, - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_TASK, -) +from sentry_sdk.tracing import TransactionSource from typing import TYPE_CHECKING @@ -68,10 +63,10 @@ GOOD_TRANSACTION_SOURCES = frozenset( [ - TRANSACTION_SOURCE_ROUTE, - TRANSACTION_SOURCE_VIEW, - TRANSACTION_SOURCE_COMPONENT, - TRANSACTION_SOURCE_TASK, + TransactionSource.ROUTE, + TransactionSource.VIEW, + TransactionSource.COMPONENT, + TransactionSource.TASK, ] ) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 9d50d38963..cf708b839e 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -2,6 +2,7 @@ import random import warnings from datetime import datetime, timedelta, timezone +from enum import Enum import sentry_sdk from sentry_sdk.consts import INSTRUMENTER, SPANSTATUS, SPANDATA @@ -16,6 +17,7 @@ from typing import TYPE_CHECKING + if TYPE_CHECKING: from collections.abc import Callable, Mapping, MutableMapping from typing import Any @@ -126,30 +128,37 @@ class TransactionKwargs(SpanKwargs, total=False): BAGGAGE_HEADER_NAME = "baggage" SENTRY_TRACE_HEADER_NAME = "sentry-trace" + # Transaction source # see https://develop.sentry.dev/sdk/event-payloads/transaction/#transaction-annotations -TRANSACTION_SOURCE_CUSTOM = "custom" -TRANSACTION_SOURCE_URL = "url" -TRANSACTION_SOURCE_ROUTE = "route" -TRANSACTION_SOURCE_VIEW = "view" -TRANSACTION_SOURCE_COMPONENT = "component" -TRANSACTION_SOURCE_TASK = "task" +class TransactionSource(str, Enum): + COMPONENT = "component" + CUSTOM = "custom" + ROUTE = "route" + TASK = "task" + URL = "url" + VIEW = "view" + + def __str__(self): + # type: () -> str + return self.value + # These are typically high cardinality and the server hates them LOW_QUALITY_TRANSACTION_SOURCES = [ - TRANSACTION_SOURCE_URL, + TransactionSource.URL, ] SOURCE_FOR_STYLE = { - "endpoint": TRANSACTION_SOURCE_COMPONENT, - "function_name": TRANSACTION_SOURCE_COMPONENT, - "handler_name": TRANSACTION_SOURCE_COMPONENT, - "method_and_path_pattern": TRANSACTION_SOURCE_ROUTE, - "path": TRANSACTION_SOURCE_URL, - "route_name": TRANSACTION_SOURCE_COMPONENT, - "route_pattern": TRANSACTION_SOURCE_ROUTE, - "uri_template": TRANSACTION_SOURCE_ROUTE, - "url": TRANSACTION_SOURCE_ROUTE, + "endpoint": TransactionSource.COMPONENT, + "function_name": TransactionSource.COMPONENT, + "handler_name": TransactionSource.COMPONENT, + "method_and_path_pattern": TransactionSource.ROUTE, + "path": TransactionSource.URL, + "route_name": TransactionSource.COMPONENT, + "route_pattern": TransactionSource.ROUTE, + "uri_template": TransactionSource.ROUTE, + "url": TransactionSource.ROUTE, } @@ -777,7 +786,7 @@ def __init__( # type: ignore[misc] name="", # type: str parent_sampled=None, # type: Optional[bool] baggage=None, # type: Optional[Baggage] - source=TRANSACTION_SOURCE_CUSTOM, # type: str + source=TransactionSource.CUSTOM, # type: str **kwargs, # type: Unpack[SpanKwargs] ): # type: (...) -> None diff --git a/tests/integrations/asgi/test_asgi.py b/tests/integrations/asgi/test_asgi.py index f3bc7147bf..f95ea14d01 100644 --- a/tests/integrations/asgi/test_asgi.py +++ b/tests/integrations/asgi/test_asgi.py @@ -3,6 +3,7 @@ import pytest import sentry_sdk from sentry_sdk import capture_message +from sentry_sdk.tracing import TransactionSource from sentry_sdk.integrations._asgi_common import _get_ip, _get_headers from sentry_sdk.integrations.asgi import SentryAsgiMiddleware, _looks_like_asgi3 @@ -129,7 +130,9 @@ async def app(scope, receive, send): @pytest.fixture def asgi3_custom_transaction_app(): async def app(scope, receive, send): - sentry_sdk.get_current_scope().set_transaction_name("foobar", source="custom") + sentry_sdk.get_current_scope().set_transaction_name( + "foobar", source=TransactionSource.CUSTOM + ) await send( { "type": "http.response.start", diff --git a/tests/integrations/sanic/test_sanic.py b/tests/integrations/sanic/test_sanic.py index 9d95907144..0419127239 100644 --- a/tests/integrations/sanic/test_sanic.py +++ b/tests/integrations/sanic/test_sanic.py @@ -10,7 +10,7 @@ import sentry_sdk from sentry_sdk import capture_message from sentry_sdk.integrations.sanic import SanicIntegration -from sentry_sdk.tracing import TRANSACTION_SOURCE_COMPONENT, TRANSACTION_SOURCE_URL +from sentry_sdk.tracing import TransactionSource from sanic import Sanic, request, response, __version__ as SANIC_VERSION_RAW from sanic.response import HTTPResponse @@ -370,7 +370,7 @@ def __init__( url="/message", expected_status=200, expected_transaction_name="hi", - expected_source=TRANSACTION_SOURCE_COMPONENT, + expected_source=TransactionSource.COMPONENT, ), TransactionTestConfig( # Transaction still recorded when we have an internal server error @@ -378,7 +378,7 @@ def __init__( url="/500", expected_status=500, expected_transaction_name="fivehundred", - expected_source=TRANSACTION_SOURCE_COMPONENT, + expected_source=TransactionSource.COMPONENT, ), TransactionTestConfig( # By default, no transaction when we have a 404 error @@ -393,7 +393,7 @@ def __init__( url="/404", expected_status=404, expected_transaction_name="/404", - expected_source=TRANSACTION_SOURCE_URL, + expected_source=TransactionSource.URL, ), TransactionTestConfig( # Transaction can be suppressed for other HTTP statuses, too, by passing config to the integration diff --git a/tests/test_metrics.py b/tests/test_metrics.py index 537f8a9646..c02f075288 100644 --- a/tests/test_metrics.py +++ b/tests/test_metrics.py @@ -7,7 +7,7 @@ import sentry_sdk from sentry_sdk import metrics -from sentry_sdk.tracing import TRANSACTION_SOURCE_ROUTE +from sentry_sdk.tracing import TransactionSource from sentry_sdk.envelope import parse_json try: @@ -539,7 +539,7 @@ def test_transaction_name( envelopes = capture_envelopes() sentry_sdk.get_current_scope().set_transaction_name( - "/user/{user_id}", source="route" + "/user/{user_id}", source=TransactionSource.ROUTE ) metrics.distribution("dist", 1.0, tags={"a": "b"}, timestamp=ts) metrics.distribution("dist", 2.0, tags={"a": "b"}, timestamp=ts) @@ -581,7 +581,7 @@ def test_metric_summaries( envelopes = capture_envelopes() with sentry_sdk.start_transaction( - op="stuff", name="/foo", source=TRANSACTION_SOURCE_ROUTE + op="stuff", name="/foo", source=TransactionSource.ROUTE ) as transaction: metrics.increment("root-counter", timestamp=ts) with metrics.timing("my-timer-metric", tags={"a": "b"}, timestamp=ts): From 07d2dce5b96594b867fd0f9cfd74ca953c811c71 Mon Sep 17 00:00:00 2001 From: Matthew T <20070360+mdtro@users.noreply.github.com> Date: Wed, 26 Feb 2025 03:01:56 -0600 Subject: [PATCH 161/591] security(gha): fix potential for shell injection (#4099) Running these workflows is gated pretty well, but this mitigates the potential for a script injection attack by passing the input to an intermediary environment variable first. See https://docs.github.com/en/actions/security-for-github-actions/security-guides/security-hardening-for-github-actions#example-of-a-script-injection-attack for more details. --- .github/workflows/release-comment-issues.yml | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/.github/workflows/release-comment-issues.yml b/.github/workflows/release-comment-issues.yml index d31c61dced..8870f25bc0 100644 --- a/.github/workflows/release-comment-issues.yml +++ b/.github/workflows/release-comment-issues.yml @@ -17,7 +17,10 @@ jobs: steps: - name: Get version id: get_version - run: echo "version=${{ github.event.inputs.version || github.event.release.tag_name }}" >> $GITHUB_OUTPUT + env: + INPUTS_VERSION: ${{ github.event.inputs.version }} + RELEASE_TAG_NAME: ${{ github.event.release.tag_name }} + run: echo "version=${$INPUTS_VERSION:-$RELEASE_TAG_NAME}" >> "$GITHUB_OUTPUT" - name: Comment on linked issues that are mentioned in release if: | @@ -28,4 +31,4 @@ jobs: uses: getsentry/release-comment-issues-gh-action@v1 with: github_token: ${{ secrets.GITHUB_TOKEN }} - version: ${{ steps.get_version.outputs.version }} \ No newline at end of file + version: ${{ steps.get_version.outputs.version }} From 5d26201b3809a55b8f4fed1b272329b30330e4d7 Mon Sep 17 00:00:00 2001 From: Kevin Ji <1146876+kevinji@users.noreply.github.com> Date: Wed, 26 Feb 2025 01:13:21 -0800 Subject: [PATCH 162/591] fix(asgi): Fix KeyError if transaction does not exist (#4095) When "transaction" does not exist on the event, it will raise `KeyError: "transaction"`. Ensure that this code handles "transaction" and "transaction_info" gracefully. --- sentry_sdk/integrations/asgi.py | 19 ++++++++++++------- 1 file changed, 12 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index 733aa2b3fe..3569336aae 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -267,13 +267,18 @@ def event_processor(self, event, hint, asgi_scope): event["request"] = deepcopy(request_data) # Only set transaction name if not already set by Starlette or FastAPI (or other frameworks) - already_set = event["transaction"] != _DEFAULT_TRANSACTION_NAME and event[ - "transaction_info" - ].get("source") in [ - TransactionSource.COMPONENT, - TransactionSource.ROUTE, - TransactionSource.CUSTOM, - ] + transaction = event.get("transaction") + transaction_source = (event.get("transaction_info") or {}).get("source") + already_set = ( + transaction is not None + and transaction != _DEFAULT_TRANSACTION_NAME + and transaction_source + in [ + TransactionSource.COMPONENT, + TransactionSource.ROUTE, + TransactionSource.CUSTOM, + ] + ) if not already_set: name, source = self._get_transaction_name_and_source( self.transaction_style, asgi_scope From 0d23b726b6b47b81acc2a1d2ba359d845467c71d Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Wed, 26 Feb 2025 16:00:06 +0100 Subject: [PATCH 163/591] feat(tracing): Backfill missing `sample_rand` on `PropagationContext` (#4038) Whenever the `PropagationContext` continues an incoming trace (i.e. whenever the `trace_id` is set, rather than being randomly generated as for a new trace), check if the `sample_rand` is present and valid in the incoming DSC. If the `sample_rand` is missing, generate it deterministically based on the `trace_id` and backfill it into the DSC on the `PropagationContext`. When generating the backfilled `sample_rand`, we ensure the generated value is consistent with the incoming trace's sampling decision and sample rate, if both of these are present. Otherwise, we generate a new value in the range [0, 1). Additionally, we propagate the `sample_rand` to transactions generated with `continue_trace` (allowing the `sample_rand` to be propagated on outgoing traces), and also allow `sample_rand` to be used for making sampling decisions. Ref #3998 --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/scope.py | 13 ++ sentry_sdk/tracing.py | 23 ++- sentry_sdk/tracing_utils.py | 141 +++++++++++++++++- sentry_sdk/utils.py | 17 +++ tests/integrations/aiohttp/test_aiohttp.py | 25 ++-- tests/integrations/celery/test_celery.py | 35 +++-- tests/integrations/httpx/test_httpx.py | 48 +++--- tests/integrations/stdlib/test_httplib.py | 13 +- tests/test_api.py | 11 +- tests/test_dsc.py | 3 +- tests/test_monitor.py | 12 +- tests/test_propagationcontext.py | 99 ++++++++++++ tests/tracing/test_integration_tests.py | 10 +- tests/tracing/test_sample_rand.py | 55 +++++++ tests/tracing/test_sample_rand_propagation.py | 43 ++++++ tests/tracing/test_sampling.py | 13 +- 16 files changed, 474 insertions(+), 87 deletions(-) create mode 100644 tests/tracing/test_sample_rand.py create mode 100644 tests/tracing/test_sample_rand_propagation.py diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index fbe97ddf44..6a5e70a6eb 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -43,6 +43,7 @@ logger, ) +import typing from typing import TYPE_CHECKING if TYPE_CHECKING: @@ -1146,8 +1147,20 @@ def continue_trace( """ self.generate_propagation_context(environ_or_headers) + # When we generate the propagation context, the sample_rand value is set + # if missing or invalid (we use the original value if it's valid). + # We want the transaction to use the same sample_rand value. Due to duplicated + # propagation logic in the transaction, we pass it in to avoid recomputing it + # in the transaction. + # TYPE SAFETY: self.generate_propagation_context() ensures that self._propagation_context + # is not None. + sample_rand = typing.cast( + PropagationContext, self._propagation_context + )._sample_rand() + transaction = Transaction.continue_from_headers( normalize_incoming_data(environ_or_headers), + _sample_rand=sample_rand, op=op, origin=origin, name=name, diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index cf708b839e..866609a66e 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1,5 +1,4 @@ import uuid -import random import warnings from datetime import datetime, timedelta, timezone from enum import Enum @@ -477,6 +476,8 @@ def continue_from_environ( def continue_from_headers( cls, headers, # type: Mapping[str, str] + *, + _sample_rand=None, # type: Optional[str] **kwargs, # type: Any ): # type: (...) -> Transaction @@ -485,6 +486,8 @@ def continue_from_headers( the ``sentry-trace`` and ``baggage`` headers). :param headers: The dictionary with the HTTP headers to pull information from. + :param _sample_rand: If provided, we override the sample_rand value from the + incoming headers with this value. (internal use only) """ # TODO move this to the Transaction class if cls is Span: @@ -495,7 +498,9 @@ def continue_from_headers( # TODO-neel move away from this kwargs stuff, it's confusing and opaque # make more explicit - baggage = Baggage.from_incoming_header(headers.get(BAGGAGE_HEADER_NAME)) + baggage = Baggage.from_incoming_header( + headers.get(BAGGAGE_HEADER_NAME), _sample_rand=_sample_rand + ) kwargs.update({BAGGAGE_HEADER_NAME: baggage}) sentrytrace_kwargs = extract_sentrytrace_data( @@ -779,6 +784,7 @@ class Transaction(Span): "_profile", "_continuous_profile", "_baggage", + "_sample_rand", ) def __init__( # type: ignore[misc] @@ -803,6 +809,14 @@ def __init__( # type: ignore[misc] self._continuous_profile = None # type: Optional[ContinuousProfile] self._baggage = baggage + baggage_sample_rand = ( + None if self._baggage is None else self._baggage._sample_rand() + ) + if baggage_sample_rand is not None: + self._sample_rand = baggage_sample_rand + else: + self._sample_rand = _generate_sample_rand(self.trace_id) + def __repr__(self): # type: () -> str return ( @@ -1173,10 +1187,10 @@ def _set_initial_sampling_decision(self, sampling_context): self.sampled = False return - # Now we roll the dice. random.random is inclusive of 0, but not of 1, + # Now we roll the dice. self._sample_rand is inclusive of 0, but not of 1, # so strict < is safe here. In case sample_rate is a boolean, cast it # to a float (True becomes 1.0 and False becomes 0.0) - self.sampled = random.random() < self.sample_rate + self.sampled = self._sample_rand < self.sample_rate if self.sampled: logger.debug( @@ -1333,6 +1347,7 @@ async def my_async_function(): Baggage, EnvironHeaders, extract_sentrytrace_data, + _generate_sample_rand, has_tracing_enabled, maybe_create_breadcrumbs_from_span, ) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index ae72b8cce9..b1e2050708 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -5,7 +5,9 @@ import sys from collections.abc import Mapping from datetime import timedelta +from decimal import ROUND_DOWN, Decimal from functools import wraps +from random import Random from urllib.parse import quote, unquote import uuid @@ -19,6 +21,7 @@ match_regex_list, qualname_from_function, to_string, + try_convert, is_sentry_url, _is_external_source, _is_in_project_root, @@ -45,6 +48,7 @@ "[ \t]*$" # whitespace ) + # This is a normal base64 regex, modified to reflect that fact that we strip the # trailing = or == off base64_stripped = ( @@ -418,6 +422,9 @@ def from_incoming_data(cls, incoming_data): propagation_context = PropagationContext() propagation_context.update(sentrytrace_data) + if propagation_context is not None: + propagation_context._fill_sample_rand() + return propagation_context @property @@ -425,6 +432,7 @@ def trace_id(self): # type: () -> str """The trace id of the Sentry trace.""" if not self._trace_id: + # New trace, don't fill in sample_rand self._trace_id = uuid.uuid4().hex return self._trace_id @@ -469,6 +477,68 @@ def __repr__(self): self.dynamic_sampling_context, ) + def _fill_sample_rand(self): + # type: () -> None + """ + Ensure that there is a valid sample_rand value in the dynamic_sampling_context. + + If there is a valid sample_rand value in the dynamic_sampling_context, we keep it. + Otherwise, we generate a sample_rand value according to the following: + + - If we have a parent_sampled value and a sample_rate in the DSC, we compute + a sample_rand value randomly in the range: + - [0, sample_rate) if parent_sampled is True, + - or, in the range [sample_rate, 1) if parent_sampled is False. + + - If either parent_sampled or sample_rate is missing, we generate a random + value in the range [0, 1). + + The sample_rand is deterministically generated from the trace_id, if present. + + This function does nothing if there is no dynamic_sampling_context. + """ + if self.dynamic_sampling_context is None: + return + + sample_rand = try_convert( + Decimal, self.dynamic_sampling_context.get("sample_rand") + ) + if sample_rand is not None and 0 <= sample_rand < 1: + # sample_rand is present and valid, so don't overwrite it + return + + # Get the sample rate and compute the transformation that will map the random value + # to the desired range: [0, 1), [0, sample_rate), or [sample_rate, 1). + sample_rate = try_convert( + float, self.dynamic_sampling_context.get("sample_rate") + ) + lower, upper = _sample_rand_range(self.parent_sampled, sample_rate) + + try: + sample_rand = _generate_sample_rand(self.trace_id, interval=(lower, upper)) + except ValueError: + # ValueError is raised if the interval is invalid, i.e. lower >= upper. + # lower >= upper might happen if the incoming trace's sampled flag + # and sample_rate are inconsistent, e.g. sample_rate=0.0 but sampled=True. + # We cannot generate a sensible sample_rand value in this case. + logger.debug( + f"Could not backfill sample_rand, since parent_sampled={self.parent_sampled} " + f"and sample_rate={sample_rate}." + ) + return + + self.dynamic_sampling_context["sample_rand"] = ( + f"{sample_rand:.6f}" # noqa: E231 + ) + + def _sample_rand(self): + # type: () -> Optional[str] + """Convenience method to get the sample_rand value from the dynamic_sampling_context.""" + if self.dynamic_sampling_context is None: + return None + + return self.dynamic_sampling_context.get("sample_rand") + class Baggage: """ @@ -491,8 +561,13 @@ def __init__( self.mutable = mutable @classmethod - def from_incoming_header(cls, header): - # type: (Optional[str]) -> Baggage + def from_incoming_header( + cls, + header, # type: Optional[str] + *, + _sample_rand=None, # type: Optional[str] + ): + # type: (...) -> Baggage """ freeze if incoming header already has sentry baggage """ @@ -515,6 +590,10 @@ def from_incoming_header(cls, header): else: third_party_items += ("," if third_party_items else "") + item + if _sample_rand is not None: + sentry_items["sample_rand"] = str(_sample_rand) + mutable = False + return Baggage(sentry_items, third_party_items, mutable) @classmethod @@ -566,6 +645,7 @@ def populate_from_transaction(cls, transaction): options = client.options or {} sentry_items["trace_id"] = transaction.trace_id + sentry_items["sample_rand"] = str(transaction._sample_rand) if options.get("environment"): sentry_items["environment"] = options["environment"] @@ -638,6 +718,20 @@ def strip_sentry_baggage(header): ) ) + def _sample_rand(self): + # type: () -> Optional[Decimal] + """Convenience method to get the sample_rand value from the sentry_items. + + We validate the value and parse it as a Decimal before returning it. The value is considered + valid if it is a Decimal in the range [0, 1). + """ + sample_rand = try_convert(Decimal, self.sentry_items.get("sample_rand")) + + if sample_rand is not None and Decimal(0) <= sample_rand < Decimal(1): + return sample_rand + + return None + def __repr__(self): # type: () -> str return f'' @@ -748,6 +842,49 @@ def get_current_span(scope=None): return current_span +def _generate_sample_rand( + trace_id, # type: Optional[str] + *, + interval=(0.0, 1.0), # type: tuple[float, float] +): + # type: (...) -> Decimal + """Generate a sample_rand value from a trace ID. + + The generated value will be pseudorandomly chosen from the provided + interval. Specifically, given (lower, upper) = interval, the generated + value will be in the range [lower, upper). The value has 6-digit precision, + so when printing with .6f, the value will never be rounded up. + + The pseudorandom number generator is seeded with the trace ID. + """ + lower, upper = interval + if not lower < upper: # using `if lower >= upper` would handle NaNs incorrectly + raise ValueError("Invalid interval: lower must be less than upper") + + rng = Random(trace_id) + sample_rand = upper + while sample_rand >= upper: + sample_rand = rng.uniform(lower, upper) + + # Round down to exactly six decimal-digit precision. + return Decimal(sample_rand).quantize(Decimal("0.000001"), rounding=ROUND_DOWN) + + +def _sample_rand_range(parent_sampled, sample_rate): + # type: (Optional[bool], Optional[float]) -> tuple[float, float] + """ + Compute the lower (inclusive) and upper (exclusive) bounds of the range of values + that a generated sample_rand value must fall into, given the parent_sampled and + sample_rate values. + """ + if parent_sampled is None or sample_rate is None: + return 0.0, 1.0 + elif parent_sampled is True: + return 0.0, sample_rate + else: # parent_sampled is False + return sample_rate, 1.0 + + # Circular imports from sentry_sdk.tracing import ( BAGGAGE_HEADER_NAME, diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index b2a39b7af1..89b2354c52 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1888,3 +1888,20 @@ def should_be_treated_as_error(ty, value): return False return True + + +if TYPE_CHECKING: + T = TypeVar("T") + + +def try_convert(convert_func, value): + # type: (Callable[[Any], T], Any) -> Optional[T] + """ + Attempt to convert from an unknown type to a specific type, using the + given function. Return None if the conversion fails, i.e. if the function + raises an exception. + """ + try: + return convert_func(value) + except Exception: + return None diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index 83dc021844..ef7c04e90a 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -626,18 +626,19 @@ async def handler(request): raw_server = await aiohttp_raw_server(handler) - with start_transaction( - name="/interactions/other-dogs/new-dog", - op="greeting.sniff", - trace_id="0123456789012345678901234567890", - ): - client = await aiohttp_client(raw_server) - resp = await client.get("/", headers={"bagGage": "custom=value"}) - - assert ( - resp.request_info.headers["baggage"] - == "custom=value,sentry-trace_id=0123456789012345678901234567890,sentry-environment=production,sentry-release=d08ebdb9309e1b004c6f52202de58a09c2268e42,sentry-transaction=/interactions/other-dogs/new-dog,sentry-sample_rate=1.0,sentry-sampled=true" - ) + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + with start_transaction( + name="/interactions/other-dogs/new-dog", + op="greeting.sniff", + trace_id="0123456789012345678901234567890", + ): + client = await aiohttp_client(raw_server) + resp = await client.get("/", headers={"bagGage": "custom=value"}) + + assert ( + resp.request_info.headers["baggage"] + == "custom=value,sentry-trace_id=0123456789012345678901234567890,sentry-sample_rand=0.500000,sentry-environment=production,sentry-release=d08ebdb9309e1b004c6f52202de58a09c2268e42,sentry-transaction=/interactions/other-dogs/new-dog,sentry-sample_rate=1.0,sentry-sampled=true" + ) @pytest.mark.asyncio diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index e51341599f..8c794bd5ff 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -509,22 +509,25 @@ def test_baggage_propagation(init_celery): def dummy_task(self, x, y): return _get_headers(self) - with start_transaction() as transaction: - result = dummy_task.apply_async( - args=(1, 0), - headers={"baggage": "custom=value"}, - ).get() - - assert sorted(result["baggage"].split(",")) == sorted( - [ - "sentry-release=abcdef", - "sentry-trace_id={}".format(transaction.trace_id), - "sentry-environment=production", - "sentry-sample_rate=1.0", - "sentry-sampled=true", - "custom=value", - ] - ) + # patch random.uniform to return a predictable sample_rand value + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + with start_transaction() as transaction: + result = dummy_task.apply_async( + args=(1, 0), + headers={"baggage": "custom=value"}, + ).get() + + assert sorted(result["baggage"].split(",")) == sorted( + [ + "sentry-release=abcdef", + "sentry-trace_id={}".format(transaction.trace_id), + "sentry-environment=production", + "sentry-sample_rand=0.500000", + "sentry-sample_rate=1.0", + "sentry-sampled=true", + "custom=value", + ] + ) def test_sentry_propagate_traces_override(init_celery): diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index d37e1fddf2..5a35b68076 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -170,30 +170,32 @@ def test_outgoing_trace_headers_append_to_baggage( url = "http://example.com/" - with start_transaction( - name="/interactions/other-dogs/new-dog", - op="greeting.sniff", - trace_id="01234567890123456789012345678901", - ) as transaction: - if asyncio.iscoroutinefunction(httpx_client.get): - response = asyncio.get_event_loop().run_until_complete( - httpx_client.get(url, headers={"baGGage": "custom=data"}) + # patch random.uniform to return a predictable sample_rand value + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + with start_transaction( + name="/interactions/other-dogs/new-dog", + op="greeting.sniff", + trace_id="01234567890123456789012345678901", + ) as transaction: + if asyncio.iscoroutinefunction(httpx_client.get): + response = asyncio.get_event_loop().run_until_complete( + httpx_client.get(url, headers={"baGGage": "custom=data"}) + ) + else: + response = httpx_client.get(url, headers={"baGGage": "custom=data"}) + + request_span = transaction._span_recorder.spans[-1] + assert response.request.headers[ + "sentry-trace" + ] == "{trace_id}-{parent_span_id}-{sampled}".format( + trace_id=transaction.trace_id, + parent_span_id=request_span.span_id, + sampled=1, + ) + assert ( + response.request.headers["baggage"] + == "custom=data,sentry-trace_id=01234567890123456789012345678901,sentry-sample_rand=0.500000,sentry-environment=production,sentry-release=d08ebdb9309e1b004c6f52202de58a09c2268e42,sentry-transaction=/interactions/other-dogs/new-dog,sentry-sample_rate=1.0,sentry-sampled=true" ) - else: - response = httpx_client.get(url, headers={"baGGage": "custom=data"}) - - request_span = transaction._span_recorder.spans[-1] - assert response.request.headers[ - "sentry-trace" - ] == "{trace_id}-{parent_span_id}-{sampled}".format( - trace_id=transaction.trace_id, - parent_span_id=request_span.span_id, - sampled=1, - ) - assert ( - response.request.headers["baggage"] - == "custom=data,sentry-trace_id=01234567890123456789012345678901,sentry-environment=production,sentry-release=d08ebdb9309e1b004c6f52202de58a09c2268e42,sentry-transaction=/interactions/other-dogs/new-dog,sentry-sample_rate=1.0,sentry-sampled=true" - ) @pytest.mark.parametrize( diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 227a24336c..892e07980b 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -1,4 +1,3 @@ -import random from http.client import HTTPConnection, HTTPSConnection from socket import SocketIO from urllib.error import HTTPError @@ -189,7 +188,7 @@ def test_outgoing_trace_headers(sentry_init, monkeypatch): "baggage": ( "other-vendor-value-1=foo;bar;baz, sentry-trace_id=771a43a4192642f0b136d5159a501700, " "sentry-public_key=49d0f7386ad645858ae85020e393bef3, sentry-sample_rate=0.01337, " - "sentry-user_id=Am%C3%A9lie, other-vendor-value-2=foo;bar;" + "sentry-user_id=Am%C3%A9lie, sentry-sample_rand=0.132521102938283, other-vendor-value-2=foo;bar;" ), } @@ -222,7 +221,8 @@ def test_outgoing_trace_headers(sentry_init, monkeypatch): "sentry-trace_id=771a43a4192642f0b136d5159a501700," "sentry-public_key=49d0f7386ad645858ae85020e393bef3," "sentry-sample_rate=1.0," - "sentry-user_id=Am%C3%A9lie" + "sentry-user_id=Am%C3%A9lie," + "sentry-sample_rand=0.132521102938283" ) assert request_headers["baggage"] == expected_outgoing_baggage @@ -235,11 +235,9 @@ def test_outgoing_trace_headers_head_sdk(sentry_init, monkeypatch): mock_send = mock.Mock() monkeypatch.setattr(HTTPSConnection, "send", mock_send) - # make sure transaction is always sampled - monkeypatch.setattr(random, "random", lambda: 0.1) - sentry_init(traces_sample_rate=0.5, release="foo") - transaction = Transaction.continue_from_headers({}) + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.25): + transaction = Transaction.continue_from_headers({}) with start_transaction(transaction=transaction, name="Head SDK tx") as transaction: HTTPSConnection("www.squirrelchasers.com").request("GET", "/top-chasers") @@ -261,6 +259,7 @@ def test_outgoing_trace_headers_head_sdk(sentry_init, monkeypatch): expected_outgoing_baggage = ( "sentry-trace_id=%s," + "sentry-sample_rand=0.250000," "sentry-environment=production," "sentry-release=foo," "sentry-sample_rate=0.5," diff --git a/tests/test_api.py b/tests/test_api.py index 3b2a9c8fb7..08c295a5c4 100644 --- a/tests/test_api.py +++ b/tests/test_api.py @@ -1,4 +1,6 @@ import pytest + +import re from unittest import mock import sentry_sdk @@ -95,10 +97,10 @@ def test_baggage_with_tracing_disabled(sentry_init): def test_baggage_with_tracing_enabled(sentry_init): sentry_init(traces_sample_rate=1.0, release="1.0.0", environment="dev") with start_transaction() as transaction: - expected_baggage = "sentry-trace_id={},sentry-environment=dev,sentry-release=1.0.0,sentry-sample_rate=1.0,sentry-sampled={}".format( + expected_baggage_re = r"^sentry-trace_id={},sentry-sample_rand=0\.\d{{6}},sentry-environment=dev,sentry-release=1\.0\.0,sentry-sample_rate=1\.0,sentry-sampled={}$".format( transaction.trace_id, "true" if transaction.sampled else "false" ) - assert get_baggage() == expected_baggage + assert re.match(expected_baggage_re, get_baggage()) @pytest.mark.forked @@ -111,7 +113,7 @@ def test_continue_trace(sentry_init): transaction = continue_trace( { "sentry-trace": "{}-{}-{}".format(trace_id, parent_span_id, parent_sampled), - "baggage": "sentry-trace_id=566e3688a61d4bc888951642d6f14a19", + "baggage": "sentry-trace_id=566e3688a61d4bc888951642d6f14a19,sentry-sample_rand=0.123456", }, name="some name", ) @@ -123,7 +125,8 @@ def test_continue_trace(sentry_init): assert propagation_context.parent_span_id == parent_span_id assert propagation_context.parent_sampled == parent_sampled assert propagation_context.dynamic_sampling_context == { - "trace_id": "566e3688a61d4bc888951642d6f14a19" + "trace_id": "566e3688a61d4bc888951642d6f14a19", + "sample_rand": "0.123456", } diff --git a/tests/test_dsc.py b/tests/test_dsc.py index 4837384a8e..8e549d0cf8 100644 --- a/tests/test_dsc.py +++ b/tests/test_dsc.py @@ -8,7 +8,6 @@ This is not tested in this file. """ -import random from unittest import mock import pytest @@ -176,7 +175,7 @@ def my_traces_sampler(sampling_context): } # We continue the incoming trace and start a new transaction - with mock.patch.object(random, "random", return_value=0.2): + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.125): transaction = sentry_sdk.continue_trace(incoming_http_headers) with sentry_sdk.start_transaction(transaction, name="foo"): pass diff --git a/tests/test_monitor.py b/tests/test_monitor.py index 03e415b5cc..b48d9f6282 100644 --- a/tests/test_monitor.py +++ b/tests/test_monitor.py @@ -1,4 +1,3 @@ -import random from collections import Counter from unittest import mock @@ -68,17 +67,16 @@ def test_transaction_uses_downsampled_rate( monitor = sentry_sdk.get_client().monitor monitor.interval = 0.1 - # make sure rng doesn't sample - monkeypatch.setattr(random, "random", lambda: 0.9) - assert monitor.is_healthy() is True monitor.run() assert monitor.is_healthy() is False assert monitor.downsample_factor == 1 - with sentry_sdk.start_transaction(name="foobar") as transaction: - assert transaction.sampled is False - assert transaction.sample_rate == 0.5 + # make sure we don't sample the transaction + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.75): + with sentry_sdk.start_transaction(name="foobar") as transaction: + assert transaction.sampled is False + assert transaction.sample_rate == 0.5 assert Counter(record_lost_event_calls) == Counter( [ diff --git a/tests/test_propagationcontext.py b/tests/test_propagationcontext.py index 85f82913f8..a0ce1094fa 100644 --- a/tests/test_propagationcontext.py +++ b/tests/test_propagationcontext.py @@ -1,6 +1,19 @@ +from unittest import mock +from unittest.mock import Mock + +import pytest + from sentry_sdk.tracing_utils import PropagationContext +SAMPLED_FLAG = { + None: "", + False: "-0", + True: "-1", +} +"""Maps the `sampled` value to the flag appended to the sentry-trace header.""" + + def test_empty_context(): ctx = PropagationContext() @@ -51,6 +64,7 @@ def test_lazy_uuids(): def test_property_setters(): ctx = PropagationContext() + ctx.trace_id = "X234567890abcdef1234567890abcdef" ctx.span_id = "X234567890abcdef" @@ -58,6 +72,7 @@ def test_property_setters(): assert ctx.trace_id == "X234567890abcdef1234567890abcdef" assert ctx._span_id == "X234567890abcdef" assert ctx.span_id == "X234567890abcdef" + assert ctx.dynamic_sampling_context is None def test_update(): @@ -81,3 +96,87 @@ def test_update(): assert ctx.dynamic_sampling_context is None assert not hasattr(ctx, "foo") + + +def test_existing_sample_rand_kept(): + ctx = PropagationContext( + trace_id="00000000000000000000000000000000", + dynamic_sampling_context={"sample_rand": "0.5"}, + ) + + # If sample_rand was regenerated, the value would be 0.919221 based on the trace_id + assert ctx.dynamic_sampling_context["sample_rand"] == "0.5" + + +@pytest.mark.parametrize( + ("parent_sampled", "sample_rate", "expected_interval"), + ( + # Note that parent_sampled and sample_rate do not scale the + # sample_rand value, only determine the range of the value. + # Expected values are determined by parent_sampled, sample_rate, + # and the trace_id. + (None, None, (0.0, 1.0)), + (None, "0.5", (0.0, 1.0)), + (False, None, (0.0, 1.0)), + (True, None, (0.0, 1.0)), + (False, "0.0", (0.0, 1.0)), + (False, "0.01", (0.01, 1.0)), + (True, "0.01", (0.0, 0.01)), + (False, "0.1", (0.1, 1.0)), + (True, "0.1", (0.0, 0.1)), + (False, "0.5", (0.5, 1.0)), + (True, "0.5", (0.0, 0.5)), + (True, "1.0", (0.0, 1.0)), + ), +) +def test_sample_rand_filled(parent_sampled, sample_rate, expected_interval): + """When continuing a trace, we want to fill in the sample_rand value if it's missing.""" + if sample_rate is not None: + sample_rate_str = f",sentry-sample_rate={sample_rate}" # noqa: E231 + else: + sample_rate_str = "" + + # for convenience, we'll just return the lower bound of the interval + mock_uniform = mock.Mock(return_value=expected_interval[0]) + + def mock_random_class(seed): + assert seed == "00000000000000000000000000000000", "seed should be the trace_id" + rv = Mock() + rv.uniform = mock_uniform + return rv + + with mock.patch("sentry_sdk.tracing_utils.Random", mock_random_class): + ctx = PropagationContext().from_incoming_data( + { + "sentry-trace": f"00000000000000000000000000000000-0000000000000000{SAMPLED_FLAG[parent_sampled]}", + # Placeholder is needed, since we only add sample_rand if sentry items are present in baggage + "baggage": f"sentry-placeholder=asdf{sample_rate_str}", + } + ) + + assert ( + ctx.dynamic_sampling_context["sample_rand"] + == f"{expected_interval[0]:.6f}" # noqa: E231 + ) + assert mock_uniform.call_count == 1 + assert mock_uniform.call_args[0] == expected_interval + + +def test_sample_rand_rounds_down(): + # Mock value that should round down to 0.999_999 + mock_uniform = mock.Mock(return_value=0.999_999_9) + + def mock_random_class(_): + rv = Mock() + rv.uniform = mock_uniform + return rv + + with mock.patch("sentry_sdk.tracing_utils.Random", mock_random_class): + ctx = PropagationContext().from_incoming_data( + { + "sentry-trace": "00000000000000000000000000000000-0000000000000000", + "baggage": "sentry-placeholder=asdf", + } + ) + + assert ctx.dynamic_sampling_context["sample_rand"] == "0.999999" diff --git a/tests/tracing/test_integration_tests.py b/tests/tracing/test_integration_tests.py index 13d1a7a77b..61ef14b7d0 100644 --- a/tests/tracing/test_integration_tests.py +++ b/tests/tracing/test_integration_tests.py @@ -1,8 +1,8 @@ import gc -import random import re import sys import weakref +from unittest import mock import pytest @@ -169,9 +169,8 @@ def test_dynamic_sampling_head_sdk_creates_dsc( envelopes = capture_envelopes() # make sure transaction is sampled for both cases - monkeypatch.setattr(random, "random", lambda: 0.1) - - transaction = Transaction.continue_from_headers({}, name="Head SDK tx") + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.25): + transaction = Transaction.continue_from_headers({}, name="Head SDK tx") # will create empty mutable baggage baggage = transaction._baggage @@ -196,12 +195,14 @@ def test_dynamic_sampling_head_sdk_creates_dsc( "release": "foo", "sample_rate": str(sample_rate), "sampled": "true" if transaction.sampled else "false", + "sample_rand": "0.250000", "transaction": "Head SDK tx", "trace_id": trace_id, } expected_baggage = ( "sentry-trace_id=%s," + "sentry-sample_rand=0.250000," "sentry-environment=production," "sentry-release=foo," "sentry-transaction=Head%%20SDK%%20tx," @@ -217,6 +218,7 @@ def test_dynamic_sampling_head_sdk_creates_dsc( "environment": "production", "release": "foo", "sample_rate": str(sample_rate), + "sample_rand": "0.250000", "sampled": "true" if transaction.sampled else "false", "transaction": "Head SDK tx", "trace_id": trace_id, diff --git a/tests/tracing/test_sample_rand.py b/tests/tracing/test_sample_rand.py new file mode 100644 index 0000000000..b8f5c042ed --- /dev/null +++ b/tests/tracing/test_sample_rand.py @@ -0,0 +1,55 @@ +from unittest import mock + +import pytest + +import sentry_sdk +from sentry_sdk.tracing_utils import Baggage + + +@pytest.mark.parametrize("sample_rand", (0.0, 0.25, 0.5, 0.75)) +@pytest.mark.parametrize("sample_rate", (0.0, 0.25, 0.5, 0.75, 1.0)) +def test_deterministic_sampled(sentry_init, capture_events, sample_rate, sample_rand): + """ + Test that sample_rand is generated on new traces, that it is used to + make the sampling decision, and that it is included in the transaction's + baggage. + """ + sentry_init(traces_sample_rate=sample_rate) + events = capture_events() + + with mock.patch( + "sentry_sdk.tracing_utils.Random.uniform", return_value=sample_rand + ): + with sentry_sdk.start_transaction() as transaction: + assert ( + transaction.get_baggage().sentry_items["sample_rand"] + == f"{sample_rand:.6f}" # noqa: E231 + ) + + # Transaction event captured if sample_rand < sample_rate, indicating that + # sample_rand is used to make the sampling decision. + assert len(events) == int(sample_rand < sample_rate) + + +@pytest.mark.parametrize("sample_rand", (0.0, 0.25, 0.5, 0.75)) +@pytest.mark.parametrize("sample_rate", (0.0, 0.25, 0.5, 0.75, 1.0)) +def test_transaction_uses_incoming_sample_rand( + sentry_init, capture_events, sample_rate, sample_rand +): + """ + Test that the transaction uses the sample_rand value from the incoming baggage. + """ + baggage = Baggage(sentry_items={"sample_rand": f"{sample_rand:.6f}"}) # noqa: E231 + + sentry_init(traces_sample_rate=sample_rate) + events = capture_events() + + with sentry_sdk.start_transaction(baggage=baggage) as transaction: + assert ( + transaction.get_baggage().sentry_items["sample_rand"] + == f"{sample_rand:.6f}" # noqa: E231 + ) + + # Transaction event captured if sample_rand < sample_rate, indicating that + # sample_rand is used to make the sampling decision. + assert len(events) == int(sample_rand < sample_rate) diff --git a/tests/tracing/test_sample_rand_propagation.py b/tests/tracing/test_sample_rand_propagation.py new file mode 100644 index 0000000000..ea3ea548ff --- /dev/null +++ b/tests/tracing/test_sample_rand_propagation.py @@ -0,0 +1,43 @@ +""" +These tests exist to verify that Scope.continue_trace() correctly propagates the +sample_rand value onto the transaction's baggage. + +We check both the case where there is an incoming sample_rand, as well as the case +where we need to compute it because it is missing. +""" + +from unittest import mock +from unittest.mock import Mock + +import sentry_sdk + + +def test_continue_trace_with_sample_rand(): + """ + Test that an incoming sample_rand is propagated onto the transaction's baggage. + """ + headers = { + "sentry-trace": "00000000000000000000000000000000-0000000000000000-0", + "baggage": "sentry-sample_rand=0.1,sentry-sample_rate=0.5", + } + + transaction = sentry_sdk.continue_trace(headers) + assert transaction.get_baggage().sentry_items["sample_rand"] == "0.1" + + +def test_continue_trace_missing_sample_rand(): + """ + Test that a missing sample_rand is filled in onto the transaction's baggage. + """ + + headers = { + "sentry-trace": "00000000000000000000000000000000-0000000000000000", + "baggage": "sentry-placeholder=asdf", + } + + mock_uniform = Mock(return_value=0.5) + + with mock.patch("sentry_sdk.tracing_utils.Random.uniform", mock_uniform): + transaction = sentry_sdk.continue_trace(headers) + + assert transaction.get_baggage().sentry_items["sample_rand"] == "0.500000" diff --git a/tests/tracing/test_sampling.py b/tests/tracing/test_sampling.py index 1ad08ecec2..1761a3dbac 100644 --- a/tests/tracing/test_sampling.py +++ b/tests/tracing/test_sampling.py @@ -7,6 +7,7 @@ import sentry_sdk from sentry_sdk import start_span, start_transaction, capture_exception from sentry_sdk.tracing import Transaction +from sentry_sdk.tracing_utils import Baggage from sentry_sdk.utils import logger @@ -73,9 +74,9 @@ def test_uses_traces_sample_rate_correctly( ): sentry_init(traces_sample_rate=traces_sample_rate) - with mock.patch.object(random, "random", return_value=0.5): - transaction = start_transaction(name="dogpark") - assert transaction.sampled is expected_decision + baggage = Baggage(sentry_items={"sample_rand": "0.500000"}) + transaction = start_transaction(name="dogpark", baggage=baggage) + assert transaction.sampled is expected_decision @pytest.mark.parametrize( @@ -89,9 +90,9 @@ def test_uses_traces_sampler_return_value_correctly( ): sentry_init(traces_sampler=mock.Mock(return_value=traces_sampler_return_value)) - with mock.patch.object(random, "random", return_value=0.5): - transaction = start_transaction(name="dogpark") - assert transaction.sampled is expected_decision + baggage = Baggage(sentry_items={"sample_rand": "0.500000"}) + transaction = start_transaction(name="dogpark", baggage=baggage) + assert transaction.sampled is expected_decision @pytest.mark.parametrize("traces_sampler_return_value", [True, False]) From 8672dc1a5c98926b570977c31241fb6394aa975d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 4 Mar 2025 09:10:20 +0100 Subject: [PATCH 164/591] Fixed bug when `cron_jobs` is set to `None` in arq integration (#4115) Handle `None` values in arq configuration gracefully. Fixes #3827 --- sentry_sdk/integrations/arq.py | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index c356347dad..1ea8e32fb3 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -199,12 +199,13 @@ def _sentry_create_worker(*args, **kwargs): if isinstance(settings_cls, dict): if "functions" in settings_cls: settings_cls["functions"] = [ - _get_arq_function(func) for func in settings_cls["functions"] + _get_arq_function(func) + for func in settings_cls.get("functions", []) ] if "cron_jobs" in settings_cls: settings_cls["cron_jobs"] = [ _get_arq_cron_job(cron_job) - for cron_job in settings_cls["cron_jobs"] + for cron_job in settings_cls.get("cron_jobs", []) ] if hasattr(settings_cls, "functions"): @@ -218,11 +219,11 @@ def _sentry_create_worker(*args, **kwargs): if "functions" in kwargs: kwargs["functions"] = [ - _get_arq_function(func) for func in kwargs["functions"] + _get_arq_function(func) for func in kwargs.get("functions", []) ] if "cron_jobs" in kwargs: kwargs["cron_jobs"] = [ - _get_arq_cron_job(cron_job) for cron_job in kwargs["cron_jobs"] + _get_arq_cron_job(cron_job) for cron_job in kwargs.get("cron_jobs", []) ] return old_create_worker(*args, **kwargs) From 7b54cfb63e683d79642d05fc92f65d7af2a18949 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 10 Mar 2025 13:14:35 +0100 Subject: [PATCH 165/591] chore(tests): Regenerate tox.ini (#4108) Run `generate-test-files.sh` (this will be automated at some point) --- tox.ini | 52 +++++++++++++++++++++++++++++----------------------- 1 file changed, 29 insertions(+), 23 deletions(-) diff --git a/tox.ini b/tox.ini index 360d16342e..f176c70f1a 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-02-19T12:41:15.689786+00:00 +# Last generated: 2025-03-10T11:46:25.287445+00:00 [tox] requires = @@ -181,7 +181,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.11.1 + {py3.9,py3.12,py3.13}-pymongo-v4.11.2 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -202,28 +202,30 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 + {py3.7,py3.12,py3.13}-statsig-v0.57.1 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 + {py3.8,py3.12,py3.13}-unleash-v6.2.0 # ~~~ GraphQL ~~~ {py3.8,py3.10,py3.11}-ariadne-v0.20.1 {py3.8,py3.11,py3.12}-ariadne-v0.22 {py3.8,py3.11,py3.12}-ariadne-v0.24.0 - {py3.9,py3.12,py3.13}-ariadne-v0.26.0 + {py3.9,py3.12,py3.13}-ariadne-v0.26.1 {py3.6,py3.9,py3.10}-gql-v3.4.1 - {py3.7,py3.11,py3.12}-gql-v3.5.0 + {py3.7,py3.11,py3.12}-gql-v3.5.2 {py3.9,py3.12,py3.13}-gql-v3.6.0b4 {py3.6,py3.9,py3.10}-graphene-v3.3 {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.226.2 - {py3.8,py3.11,py3.12}-strawberry-v0.243.1 - {py3.9,py3.12,py3.13}-strawberry-v0.260.2 + {py3.8,py3.11,py3.12}-strawberry-v0.227.7 + {py3.8,py3.11,py3.12}-strawberry-v0.245.0 + {py3.9,py3.12,py3.13}-strawberry-v0.262.1 # ~~~ Network ~~~ @@ -231,13 +233,14 @@ envlist = {py3.7,py3.9,py3.10}-grpc-v1.44.0 {py3.7,py3.10,py3.11}-grpc-v1.58.3 {py3.8,py3.12,py3.13}-grpc-v1.70.0 + {py3.9,py3.12,py3.13}-grpc-v1.71.0rc2 # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.11,py3.12}-celery-v5.4.0 - {py3.8,py3.12,py3.13}-celery-v5.5.0rc4 + {py3.8,py3.12,py3.13}-celery-v5.5.0rc5 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -247,7 +250,7 @@ envlist = {py3.8,py3.9}-spark-v3.0.3 {py3.8,py3.9}-spark-v3.2.4 {py3.8,py3.10,py3.11}-spark-v3.4.4 - {py3.8,py3.10,py3.11}-spark-v3.5.4 + {py3.8,py3.10,py3.11}-spark-v3.5.5 # ~~~ Web 1 ~~~ @@ -259,7 +262,7 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.45.3 + {py3.9,py3.12,py3.13}-starlette-v0.46.1 # ~~~ Web 2 ~~~ @@ -294,9 +297,9 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.9 - {py3.8,py3.11,py3.12}-trytond-v7.4.6 + {py3.8,py3.11,py3.12}-trytond-v7.4.7 - {py3.7,py3.11,py3.12}-typer-v0.15.1 + {py3.7,py3.12,py3.13}-typer-v0.15.2 @@ -562,7 +565,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.11.1: pymongo==4.11.1 + pymongo-v4.11.2: pymongo==4.11.2 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -584,23 +587,25 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 + statsig-v0.57.1: statsig==0.57.1 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 unleash-v6.1.0: UnleashClient==6.1.0 + unleash-v6.2.0: UnleashClient==6.2.0 # ~~~ GraphQL ~~~ ariadne-v0.20.1: ariadne==0.20.1 ariadne-v0.22: ariadne==0.22 ariadne-v0.24.0: ariadne==0.24.0 - ariadne-v0.26.0: ariadne==0.26.0 + ariadne-v0.26.1: ariadne==0.26.1 ariadne: fastapi ariadne: flask ariadne: httpx gql-v3.4.1: gql[all]==3.4.1 - gql-v3.5.0: gql[all]==3.5.0 + gql-v3.5.2: gql[all]==3.5.2 gql-v3.6.0b4: gql[all]==3.6.0b4 graphene-v3.3: graphene==3.3 @@ -612,9 +617,9 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.226.2: strawberry-graphql[fastapi,flask]==0.226.2 - strawberry-v0.243.1: strawberry-graphql[fastapi,flask]==0.243.1 - strawberry-v0.260.2: strawberry-graphql[fastapi,flask]==0.260.2 + strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 + strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 + strawberry-v0.262.1: strawberry-graphql[fastapi,flask]==0.262.1 strawberry: httpx @@ -623,6 +628,7 @@ deps = grpc-v1.44.0: grpcio==1.44.0 grpc-v1.58.3: grpcio==1.58.3 grpc-v1.70.0: grpcio==1.70.0 + grpc-v1.71.0rc2: grpcio==1.71.0rc2 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -633,7 +639,7 @@ deps = celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.4.0: celery==5.4.0 - celery-v5.5.0rc4: celery==5.5.0rc4 + celery-v5.5.0rc5: celery==5.5.0rc5 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 @@ -646,7 +652,7 @@ deps = spark-v3.0.3: pyspark==3.0.3 spark-v3.2.4: pyspark==3.2.4 spark-v3.4.4: pyspark==3.4.4 - spark-v3.5.4: pyspark==3.5.4 + spark-v3.5.5: pyspark==3.5.5 # ~~~ Web 1 ~~~ @@ -662,7 +668,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.45.3: starlette==0.45.3 + starlette-v0.46.1: starlette==0.46.1 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -720,12 +726,12 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.9: trytond==7.0.9 - trytond-v7.4.6: trytond==7.4.6 + trytond-v7.4.7: trytond==7.4.7 trytond: werkzeug trytond-v4.6.9: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 - typer-v0.15.1: typer==0.15.1 + typer-v0.15.2: typer==0.15.2 From 9e89c3054f6289b544f84d20bae605c520728b2d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 10 Mar 2025 13:42:41 +0100 Subject: [PATCH 166/591] fix(typing): Set correct type for set_context everywhere (#4123) --- sentry_sdk/tracing.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 866609a66e..13d9f63d5e 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1052,7 +1052,7 @@ def set_measurement(self, name, value, unit=""): self._measurements[name] = {"value": value, "unit": unit} def set_context(self, key, value): - # type: (str, Any) -> None + # type: (str, dict[str, Any]) -> None """Sets a context. Transactions can have multiple contexts and they should follow the format described in the "Contexts Interface" documentation. @@ -1287,7 +1287,7 @@ def set_measurement(self, name, value, unit=""): pass def set_context(self, key, value): - # type: (str, Any) -> None + # type: (str, dict[str, Any]) -> None pass def init_span_recorder(self, maxlen): From 7deebf0883750823953e84c29e96840319e95f60 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 10 Mar 2025 14:50:15 +0100 Subject: [PATCH 167/591] Fix FastAPI/Starlette middleware with positional arguments. (#4118) Fixes #3246 --- sentry_sdk/integrations/starlette.py | 8 +++---- .../integrations/starlette/test_starlette.py | 23 ++++++++++++++++++- 2 files changed, 26 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index 687a428203..deb05059d5 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -362,13 +362,13 @@ def patch_middlewares(): if not_yet_patched: - def _sentry_middleware_init(self, cls, **options): - # type: (Any, Any, Any) -> None + def _sentry_middleware_init(self, cls, *args, **kwargs): + # type: (Any, Any, Any, Any) -> None if cls == SentryAsgiMiddleware: - return old_middleware_init(self, cls, **options) + return old_middleware_init(self, cls, *args, **kwargs) span_enabled_cls = _enable_span_for_middleware(cls) - old_middleware_init(self, span_enabled_cls, **options) + old_middleware_init(self, span_enabled_cls, *args, **kwargs) if cls == AuthenticationMiddleware: patch_authentication_middleware(cls) diff --git a/tests/integrations/starlette/test_starlette.py b/tests/integrations/starlette/test_starlette.py index 93da0420aa..3289f69ed6 100644 --- a/tests/integrations/starlette/test_starlette.py +++ b/tests/integrations/starlette/test_starlette.py @@ -31,7 +31,6 @@ from starlette.middleware.authentication import AuthenticationMiddleware from starlette.middleware.trustedhost import TrustedHostMiddleware from starlette.testclient import TestClient - from tests.integrations.conftest import parametrize_test_configurable_status_codes @@ -238,6 +237,12 @@ async def do_stuff(message): await self.app(scope, receive, do_stuff) +class SampleMiddlewareWithArgs(Middleware): + def __init__(self, app, bla=None): + self.app = app + self.bla = bla + + class SampleReceiveSendMiddleware: def __init__(self, app): self.app = app @@ -862,6 +867,22 @@ def test_middleware_partial_receive_send(sentry_init, capture_events): idx += 1 +@pytest.mark.skipif( + STARLETTE_VERSION < (0, 35), + reason="Positional args for middleware have been introduced in Starlette >= 0.35", +) +def test_middleware_positional_args(sentry_init): + sentry_init( + traces_sample_rate=1.0, + integrations=[StarletteIntegration()], + ) + _ = starlette_app_factory(middleware=[Middleware(SampleMiddlewareWithArgs, "bla")]) + + # Only creating the App with an Middleware with args + # should not raise an error + # So as long as test passes, we are good + + def test_legacy_setup( sentry_init, capture_events, From a97c53ca697c1fd3132e5b3d5e67887d63187963 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 10 Mar 2025 14:59:05 +0100 Subject: [PATCH 168/591] Added timeout to HTTP requests in CloudResourceContextIntegration (#4120) The URL that works in EC2 does not work in ECS, this can lead to the HTTP request getting stuck. Fixes #2376 --- .../integrations/cloud_resource_context.py | 36 +++++++++++++++---- 1 file changed, 29 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/cloud_resource_context.py b/sentry_sdk/integrations/cloud_resource_context.py index 8d080899f3..ca5ae47e6b 100644 --- a/sentry_sdk/integrations/cloud_resource_context.py +++ b/sentry_sdk/integrations/cloud_resource_context.py @@ -13,6 +13,8 @@ CONTEXT_TYPE = "cloud_resource" +HTTP_TIMEOUT = 2.0 + AWS_METADATA_HOST = "169.254.169.254" AWS_TOKEN_URL = "http://{}/latest/api/token".format(AWS_METADATA_HOST) AWS_METADATA_URL = "http://{}/latest/dynamic/instance-identity/document".format( @@ -59,7 +61,7 @@ class CloudResourceContextIntegration(Integration): cloud_provider = "" aws_token = "" - http = urllib3.PoolManager() + http = urllib3.PoolManager(timeout=HTTP_TIMEOUT) gcp_metadata = None @@ -83,7 +85,13 @@ def _is_aws(cls): cls.aws_token = r.data.decode() return True - except Exception: + except urllib3.exceptions.TimeoutError: + logger.debug( + "AWS metadata service timed out after %s seconds", HTTP_TIMEOUT + ) + return False + except Exception as e: + logger.debug("Error checking AWS metadata service: %s", str(e)) return False @classmethod @@ -131,8 +139,12 @@ def _get_aws_context(cls): except Exception: pass - except Exception: - pass + except urllib3.exceptions.TimeoutError: + logger.debug( + "AWS metadata service timed out after %s seconds", HTTP_TIMEOUT + ) + except Exception as e: + logger.debug("Error fetching AWS metadata: %s", str(e)) return ctx @@ -152,7 +164,13 @@ def _is_gcp(cls): cls.gcp_metadata = json.loads(r.data.decode("utf-8")) return True - except Exception: + except urllib3.exceptions.TimeoutError: + logger.debug( + "GCP metadata service timed out after %s seconds", HTTP_TIMEOUT + ) + return False + except Exception as e: + logger.debug("Error checking GCP metadata service: %s", str(e)) return False @classmethod @@ -201,8 +219,12 @@ def _get_gcp_context(cls): except Exception: pass - except Exception: - pass + except urllib3.exceptions.TimeoutError: + logger.debug( + "GCP metadata service timed out after %s seconds", HTTP_TIMEOUT + ) + except Exception as e: + logger.debug("Error fetching GCP metadata: %s", str(e)) return ctx From d4f4130ad9e2c5c24c06c50855aa0b55fa407a11 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 12 Mar 2025 14:56:42 +0100 Subject: [PATCH 169/591] Run AWS Lambda tests locally (#3988) Test Sentry AWS Lambda integration locally instead of creating actual Lambda function in AWS: - Create a local AWS Lambda environment using AWS SAM and AWS CDK. (Docker based) - Start a local Sentry server that accepts envelopes. - Run the tests in the local AWS Lambda environment configured with a DSN that tells the SDK to send data to the local Sentry server. - Read the captured envelopes from the local Sentry server to assert their correctness. - Update CI configuration, so AWS tests are now handled the same as test suite matrices of other integrations. There is also a follow-up PR that removes obsolete code handling AWS authentication data: #4076 (This PR will also fix the one failing test) Fixes #2795 --- .github/PULL_REQUEST_TEMPLATE.md | 2 +- .../scripts/trigger_tests_on_label.py | 72 -- .github/workflows/test-integrations-aws.yml | 126 --- .github/workflows/test-integrations-cloud.yml | 18 +- .gitignore | 3 + requirements-testing.txt | 1 + scripts/aws-cleanup.sh | 18 - .../aws-attach-layer-to-lambda-function.sh | 0 .../aws-delete-lambda-layer-versions.sh | 1 + scripts/{ => aws}/aws-deploy-local-layer.sh | 3 +- scripts/aws_lambda_functions/README.md | 4 - .../sentryPythonDeleteTestFunctions/README.md | 13 - .../lambda_function.py | 55 -- scripts/populate_tox/tox.jinja | 12 +- .../split_tox_gh_actions.py | 17 +- .../split_tox_gh_actions/templates/base.jinja | 22 - .../templates/check_permissions.jinja | 30 - .../templates/test_group.jinja | 14 +- tests/integrations/aws_lambda/__init__.py | 2 + tests/integrations/aws_lambda/client.py | 408 -------- .../lambda_functions/BasicException/index.py | 6 + .../lambda_functions/BasicOk/index.py | 4 + .../lambda_functions/InitError/index.py | 3 + .../lambda_functions/TimeoutError/index.py | 8 + .../RaiseErrorPerformanceDisabled/.gitignore | 11 + .../RaiseErrorPerformanceDisabled/index.py | 14 + .../RaiseErrorPerformanceEnabled/.gitignore | 11 + .../RaiseErrorPerformanceEnabled/index.py | 14 + .../TracesSampler/.gitignore | 11 + .../TracesSampler/index.py | 49 + tests/integrations/aws_lambda/test_aws.py | 898 ------------------ .../aws_lambda/test_aws_lambda.py | 550 +++++++++++ tests/integrations/aws_lambda/utils.py | 294 ++++++ tox.ini | 12 +- 34 files changed, 1021 insertions(+), 1685 deletions(-) delete mode 100644 .github/workflows/scripts/trigger_tests_on_label.py delete mode 100644 .github/workflows/test-integrations-aws.yml delete mode 100755 scripts/aws-cleanup.sh rename scripts/{ => aws}/aws-attach-layer-to-lambda-function.sh (100%) rename scripts/{ => aws}/aws-delete-lambda-layer-versions.sh (95%) rename scripts/{ => aws}/aws-deploy-local-layer.sh (81%) delete mode 100644 scripts/aws_lambda_functions/README.md delete mode 100644 scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/README.md delete mode 100644 scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/lambda_function.py delete mode 100644 scripts/split_tox_gh_actions/templates/check_permissions.jinja delete mode 100644 tests/integrations/aws_lambda/client.py create mode 100644 tests/integrations/aws_lambda/lambda_functions/BasicException/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions/BasicOk/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions/InitError/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions/TimeoutError/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/index.py create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/index.py delete mode 100644 tests/integrations/aws_lambda/test_aws.py create mode 100644 tests/integrations/aws_lambda/test_aws_lambda.py create mode 100644 tests/integrations/aws_lambda/utils.py diff --git a/.github/PULL_REQUEST_TEMPLATE.md b/.github/PULL_REQUEST_TEMPLATE.md index f0002fe486..12db62315a 100644 --- a/.github/PULL_REQUEST_TEMPLATE.md +++ b/.github/PULL_REQUEST_TEMPLATE.md @@ -4,4 +4,4 @@ Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. -Running the test suite on your PR might require maintainer approval. The AWS Lambda tests additionally require a maintainer to add a special label, and they will fail until this label is added. +Running the test suite on your PR might require maintainer approval. \ No newline at end of file diff --git a/.github/workflows/scripts/trigger_tests_on_label.py b/.github/workflows/scripts/trigger_tests_on_label.py deleted file mode 100644 index f6039fd16a..0000000000 --- a/.github/workflows/scripts/trigger_tests_on_label.py +++ /dev/null @@ -1,72 +0,0 @@ -#!/usr/bin/env python3 -import argparse -import json -import os -from urllib.parse import quote -from urllib.request import Request, urlopen - -LABEL = "Trigger: tests using secrets" - - -def _has_write(repo_id: int, username: str, *, token: str) -> bool: - req = Request( - f"https://api.github.com/repositories/%7Brepo_id%7D/collaborators/%7Busername%7D/permission", - headers={"Authorization": f"token {token}"}, - ) - contents = json.load(urlopen(req, timeout=10)) - - return contents["permission"] in {"admin", "write"} - - -def _remove_label(repo_id: int, pr: int, label: str, *, token: str) -> None: - quoted_label = quote(label) - req = Request( - f"https://api.github.com/repositories/%7Brepo_id%7D/issues/%7Bpr%7D/labels/%7Bquoted_label%7D", - method="DELETE", - headers={"Authorization": f"token {token}"}, - ) - urlopen(req) - - -def main() -> int: - parser = argparse.ArgumentParser() - parser.add_argument("--repo-id", type=int, required=True) - parser.add_argument("--pr", type=int, required=True) - parser.add_argument("--event", required=True) - parser.add_argument("--username", required=True) - parser.add_argument("--label-names", type=json.loads, required=True) - args = parser.parse_args() - - token = os.environ["GITHUB_TOKEN"] - - write_permission = _has_write(args.repo_id, args.username, token=token) - - if ( - not write_permission - # `reopened` is included here due to close => push => reopen - and args.event in {"synchronize", "reopened"} - and LABEL in args.label_names - ): - print(f"Invalidating label [{LABEL}] due to code change...") - _remove_label(args.repo_id, args.pr, LABEL, token=token) - args.label_names.remove(LABEL) - - if write_permission or LABEL in args.label_names: - print("Permissions passed!") - print(f"- has write permission: {write_permission}") - print(f"- has [{LABEL}] label: {LABEL in args.label_names}") - return 0 - else: - print("Permissions failed!") - print(f"- has write permission: {write_permission}") - print(f"- has [{LABEL}] label: {LABEL in args.label_names}") - print(f"- args.label_names: {args.label_names}") - print( - f"Please have a collaborator add the [{LABEL}] label once they " - f"have reviewed the code to trigger tests." - ) - return 1 - - -if __name__ == "__main__": - raise SystemExit(main()) diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml deleted file mode 100644 index 21171f7843..0000000000 --- a/.github/workflows/test-integrations-aws.yml +++ /dev/null @@ -1,126 +0,0 @@ -# Do not edit this YAML file. This file is generated automatically by executing -# python scripts/split_tox_gh_actions/split_tox_gh_actions.py -# The template responsible for it is in -# scripts/split_tox_gh_actions/templates/base.jinja -name: Test AWS -on: - push: - branches: - - master - - release/** - - potel-base - # XXX: We are using `pull_request_target` instead of `pull_request` because we want - # this to run on forks with access to the secrets necessary to run the test suite. - # Prefer to use `pull_request` when possible. - pull_request_target: - types: [labeled, opened, reopened, synchronize] -# Cancel in progress workflows on pull_requests. -# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value -concurrency: - group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} - cancel-in-progress: true -permissions: - contents: read - # `write` is needed to remove the `Trigger: tests using secrets` label - pull-requests: write -env: - SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID: ${{ secrets.SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID }} - SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY: ${{ secrets.SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY }} - BUILD_CACHE_KEY: ${{ github.sha }} - CACHED_BUILD_PATHS: | - ${{ github.workspace }}/dist-serverless -jobs: - check-permissions: - name: permissions check - runs-on: ubuntu-20.04 - steps: - - uses: actions/checkout@v4.2.2 - with: - persist-credentials: false - - name: Check permissions on PR - if: github.event_name == 'pull_request_target' - run: | - python3 -uS .github/workflows/scripts/trigger_tests_on_label.py \ - --repo-id ${{ github.event.repository.id }} \ - --pr ${{ github.event.number }} \ - --event ${{ github.event.action }} \ - --username "$ARG_USERNAME" \ - --label-names "$ARG_LABEL_NAMES" - env: - GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} - # these can contain special characters - ARG_USERNAME: ${{ github.event.pull_request.user.login }} - ARG_LABEL_NAMES: ${{ toJSON(github.event.pull_request.labels.*.name) }} - - name: Check permissions on repo branch - if: github.event_name == 'push' - run: true - test-aws-pinned: - name: AWS (pinned) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.9"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] - needs: check-permissions - steps: - - uses: actions/checkout@v4.2.2 - with: - ref: ${{ github.event.pull_request.head.sha || github.ref }} - - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test aws_lambda pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aws_lambda" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true - check_required_tests: - name: All pinned AWS tests passed - needs: test-aws-pinned - # Always run this, even if a dependent job failed - if: always() - runs-on: ubuntu-20.04 - steps: - - name: Check for failures - if: contains(needs.test-aws-pinned.result, 'failure') || contains(needs.test-aws-pinned.result, 'skipped') - run: | - echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index b929b8d899..efa71c8e0c 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -35,6 +35,10 @@ jobs: # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] + services: + docker: + image: docker:dind # Required for Docker network management + options: --privileged # Required for Docker-in-Docker operations steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 @@ -47,6 +51,10 @@ jobs: - name: Erase coverage run: | coverage erase + - name: Test aws_lambda latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-aws_lambda-latest" - name: Test boto3 latest run: | set -x # print commands that are executed @@ -97,12 +105,16 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.9","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] + services: + docker: + image: docker:dind # Required for Docker network management + options: --privileged # Required for Docker-in-Docker operations steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 @@ -115,6 +127,10 @@ jobs: - name: Erase coverage run: | coverage erase + - name: Test aws_lambda pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aws_lambda" - name: Test boto3 pinned run: | set -x # print commands that are executed diff --git a/.gitignore b/.gitignore index 8c7a5f2174..0dad53b2f4 100644 --- a/.gitignore +++ b/.gitignore @@ -28,3 +28,6 @@ relay pip-wheel-metadata .mypy_cache .vscode/ + +# for running AWS Lambda tests using AWS SAM +sam.template.yaml diff --git a/requirements-testing.txt b/requirements-testing.txt index dfbd821845..503ab5de68 100644 --- a/requirements-testing.txt +++ b/requirements-testing.txt @@ -14,3 +14,4 @@ socksio httpcore[http2] setuptools Brotli +docker \ No newline at end of file diff --git a/scripts/aws-cleanup.sh b/scripts/aws-cleanup.sh deleted file mode 100755 index 982835c283..0000000000 --- a/scripts/aws-cleanup.sh +++ /dev/null @@ -1,18 +0,0 @@ -#!/bin/sh -# -# Helper script to clean up AWS Lambda functions created -# by the test suite (tests/integrations/aws_lambda/test_aws.py). -# -# This will delete all Lambda functions named `test_function_*`. -# - -export AWS_DEFAULT_REGION="us-east-1" -export AWS_ACCESS_KEY_ID="$SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID" -export AWS_SECRET_ACCESS_KEY="$SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY" - -for func in $(aws lambda list-functions --output text --query 'Functions[?starts_with(FunctionName, `test_`) == `true`].FunctionName'); do - echo "Deleting $func" - aws lambda delete-function --function-name "$func" -done - -echo "All done! Have a nice day!" diff --git a/scripts/aws-attach-layer-to-lambda-function.sh b/scripts/aws/aws-attach-layer-to-lambda-function.sh similarity index 100% rename from scripts/aws-attach-layer-to-lambda-function.sh rename to scripts/aws/aws-attach-layer-to-lambda-function.sh diff --git a/scripts/aws-delete-lambda-layer-versions.sh b/scripts/aws/aws-delete-lambda-layer-versions.sh similarity index 95% rename from scripts/aws-delete-lambda-layer-versions.sh rename to scripts/aws/aws-delete-lambda-layer-versions.sh index f467f9398b..dcbd2f9c65 100755 --- a/scripts/aws-delete-lambda-layer-versions.sh +++ b/scripts/aws/aws-delete-lambda-layer-versions.sh @@ -1,6 +1,7 @@ #!/usr/bin/env bash # # Deletes all versions of the layer specified in LAYER_NAME in one region. +# Use with caution! # set -euo pipefail diff --git a/scripts/aws-deploy-local-layer.sh b/scripts/aws/aws-deploy-local-layer.sh similarity index 81% rename from scripts/aws-deploy-local-layer.sh rename to scripts/aws/aws-deploy-local-layer.sh index 56f2087596..ee7b3e45c0 100755 --- a/scripts/aws-deploy-local-layer.sh +++ b/scripts/aws/aws-deploy-local-layer.sh @@ -1,9 +1,8 @@ #!/usr/bin/env bash # -# Builds and deploys the Sentry AWS Lambda layer (including the Sentry SDK and the Sentry Lambda Extension) +# Builds and deploys the `SentryPythonServerlessSDK-local-dev` AWS Lambda layer (containing the Sentry SDK) # # The currently checked out version of the SDK in your local directory is used. -# The latest version of the Lambda Extension is fetched from the Sentry Release Registry. # set -euo pipefail diff --git a/scripts/aws_lambda_functions/README.md b/scripts/aws_lambda_functions/README.md deleted file mode 100644 index e07b445d5b..0000000000 --- a/scripts/aws_lambda_functions/README.md +++ /dev/null @@ -1,4 +0,0 @@ -aws_lambda_functions -==================== - -In this directory you can place AWS Lambda functions that are used for administrative tasks (or whatever) \ No newline at end of file diff --git a/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/README.md b/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/README.md deleted file mode 100644 index de1120a026..0000000000 --- a/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/README.md +++ /dev/null @@ -1,13 +0,0 @@ -sentryPythonDeleteTestFunctions -=============================== - -This AWS Lambda function deletes all AWS Lambda functions in the current AWS account that are prefixed with `test_`. -The functions that are deleted are created by the Google Actions CI checks running on every PR of the `sentry-python` repository. - -The Lambda function has been deployed here: -- AWS Account ID: `943013980633` -- Region: `us-east-1` -- Function ARN: `arn:aws:lambda:us-east-1:943013980633:function:sentryPythonDeleteTestFunctions` - -This function also emits Sentry Metrics and Sentry Crons checkins to the `sentry-python` project in the `Sentry SDKs` organisation on Sentry.io: -https://sentry-sdks.sentry.io/projects/sentry-python/?project=5461230 \ No newline at end of file diff --git a/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/lambda_function.py b/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/lambda_function.py deleted file mode 100644 index ce7afb6aa4..0000000000 --- a/scripts/aws_lambda_functions/sentryPythonDeleteTestFunctions/lambda_function.py +++ /dev/null @@ -1,55 +0,0 @@ -import boto3 -import sentry_sdk - - -monitor_slug = "python-sdk-aws-lambda-tests-cleanup" -monitor_config = { - "schedule": { - "type": "crontab", - "value": "0 12 * * 0", # 12 o'clock on Sunday - }, - "timezone": "UTC", - "checkin_margin": 2, - "max_runtime": 20, - "failure_issue_threshold": 1, - "recovery_threshold": 1, -} - - -@sentry_sdk.crons.monitor(monitor_slug=monitor_slug) -def delete_lambda_functions(prefix="test_"): - """ - Delete all AWS Lambda functions in the current account - where the function name matches the prefix - """ - client = boto3.client("lambda", region_name="us-east-1") - functions_deleted = 0 - - functions_paginator = client.get_paginator("list_functions") - for functions_page in functions_paginator.paginate(): - for func in functions_page["Functions"]: - function_name = func["FunctionName"] - if function_name.startswith(prefix): - try: - response = client.delete_function( - FunctionName=func["FunctionArn"], - ) - functions_deleted += 1 - except Exception as ex: - print(f"Got exception: {ex}") - - return functions_deleted - - -def lambda_handler(event, context): - functions_deleted = delete_lambda_functions() - - sentry_sdk.metrics.gauge( - key="num_aws_functions_deleted", - value=functions_deleted, - ) - - return { - "statusCode": 200, - "body": f"{functions_deleted} AWS Lambda functions deleted successfully.", - } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 81ab17c919..9da986a35a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -57,10 +57,7 @@ envlist = {py3.8,py3.11,py3.12}-asyncpg-latest # AWS Lambda - # The aws_lambda tests deploy to the real AWS and have their own - # matrix of Python versions to run the test lambda function in. - # see `lambda_runtime` fixture in tests/integrations/aws_lambda.py - {py3.9}-aws_lambda + {py3.8,py3.9,py3.11,py3.13}-aws_lambda # Beam {py3.7}-beam-v{2.12} @@ -250,7 +247,12 @@ deps = asyncpg: pytest-asyncio # AWS Lambda + aws_lambda: aws-cdk-lib + aws_lambda: aws-sam-cli aws_lambda: boto3 + aws_lambda: fastapi + aws_lambda: requests + aws_lambda: uvicorn # Beam beam-v2.12: apache-beam~=2.12.0 @@ -528,8 +530,6 @@ setenv = socket: TESTPATH=tests/integrations/socket passenv = - SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID - SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY SENTRY_PYTHON_TEST_POSTGRES_HOST SENTRY_PYTHON_TEST_POSTGRES_USER SENTRY_PYTHON_TEST_POSTGRES_PASSWORD diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 5218b0675f..293af897c9 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -43,11 +43,7 @@ "clickhouse_driver", } -FRAMEWORKS_NEEDING_AWS = { - "aws_lambda", -} - -FRAMEWORKS_NEEDING_GITHUB_SECRETS = { +FRAMEWORKS_NEEDING_DOCKER = { "aws_lambda", } @@ -65,12 +61,8 @@ "openai", "huggingface_hub", ], - "AWS": [ - # this is separate from Cloud Computing because only this one test suite - # needs to run with access to GitHub secrets - "aws_lambda", - ], "Cloud": [ + "aws_lambda", "boto3", "chalice", "cloud_resource_context", @@ -292,13 +284,10 @@ def render_template(group, frameworks, py_versions_pinned, py_versions_latest): "group": group, "frameworks": frameworks, "categories": sorted(categories), - "needs_aws_credentials": bool(set(frameworks) & FRAMEWORKS_NEEDING_AWS), "needs_clickhouse": bool(set(frameworks) & FRAMEWORKS_NEEDING_CLICKHOUSE), + "needs_docker": bool(set(frameworks) & FRAMEWORKS_NEEDING_DOCKER), "needs_postgres": bool(set(frameworks) & FRAMEWORKS_NEEDING_POSTGRES), "needs_redis": bool(set(frameworks) & FRAMEWORKS_NEEDING_REDIS), - "needs_github_secrets": bool( - set(frameworks) & FRAMEWORKS_NEEDING_GITHUB_SECRETS - ), "py_versions": { category: [f'"{version}"' for version in _normalize_py_versions(versions)] for category, versions in py_versions.items() diff --git a/scripts/split_tox_gh_actions/templates/base.jinja b/scripts/split_tox_gh_actions/templates/base.jinja index e69b6f9134..75c988e32a 100644 --- a/scripts/split_tox_gh_actions/templates/base.jinja +++ b/scripts/split_tox_gh_actions/templates/base.jinja @@ -13,15 +13,7 @@ on: - release/** - potel-base - {% if needs_github_secrets %} - # XXX: We are using `pull_request_target` instead of `pull_request` because we want - # this to run on forks with access to the secrets necessary to run the test suite. - # Prefer to use `pull_request` when possible. - pull_request_target: - types: [labeled, opened, reopened, synchronize] - {% else %} pull_request: - {% endif %} # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value @@ -31,27 +23,13 @@ concurrency: permissions: contents: read - {% if needs_github_secrets %} - # `write` is needed to remove the `Trigger: tests using secrets` label - pull-requests: write - {% endif %} env: -{% if needs_aws_credentials %} -{% raw %} - SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID: ${{ secrets.SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID }} - SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY: ${{ secrets.SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY }} -{% endraw %} -{% endif %} BUILD_CACHE_KEY: {% raw %}${{ github.sha }}{% endraw %} CACHED_BUILD_PATHS: | {% raw %}${{ github.workspace }}/dist-serverless{% endraw %} jobs: -{% if needs_github_secrets %} -{% include "check_permissions.jinja" %} -{% endif %} - {% for category in categories %} {% include "test_group.jinja" %} {% endfor %} diff --git a/scripts/split_tox_gh_actions/templates/check_permissions.jinja b/scripts/split_tox_gh_actions/templates/check_permissions.jinja deleted file mode 100644 index 390f447856..0000000000 --- a/scripts/split_tox_gh_actions/templates/check_permissions.jinja +++ /dev/null @@ -1,30 +0,0 @@ - check-permissions: - name: permissions check - runs-on: ubuntu-20.04 - steps: - - uses: actions/checkout@v4.2.2 - with: - persist-credentials: false - - - name: Check permissions on PR - if: github.event_name == 'pull_request_target' - run: | - {% raw %} - python3 -uS .github/workflows/scripts/trigger_tests_on_label.py \ - --repo-id ${{ github.event.repository.id }} \ - --pr ${{ github.event.number }} \ - --event ${{ github.event.action }} \ - --username "$ARG_USERNAME" \ - --label-names "$ARG_LABEL_NAMES" - {% endraw %} - env: - {% raw %} - GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} - # these can contain special characters - ARG_USERNAME: ${{ github.event.pull_request.user.login }} - ARG_LABEL_NAMES: ${{ toJSON(github.event.pull_request.labels.*.name) }} - {% endraw %} - - - name: Check permissions on repo branch - if: github.event_name == 'push' - run: true diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 01f9cd56ec..9fcc0b1527 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -12,10 +12,12 @@ # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 os: [ubuntu-20.04] - {% if needs_github_secrets %} - needs: check-permissions + {% if needs_docker %} + services: + docker: + image: docker:dind # Required for Docker network management + options: --privileged # Required for Docker-in-Docker operations {% endif %} - {% if needs_postgres %} services: postgres: @@ -40,12 +42,6 @@ steps: - uses: actions/checkout@v4.2.2 - {% if needs_github_secrets %} - {% raw %} - with: - ref: ${{ github.event.pull_request.head.sha || github.ref }} - {% endraw %} - {% endif %} - uses: actions/setup-python@v5 with: python-version: {% raw %}${{ matrix.python-version }}{% endraw %} diff --git a/tests/integrations/aws_lambda/__init__.py b/tests/integrations/aws_lambda/__init__.py index 71eb245353..449f4dc95d 100644 --- a/tests/integrations/aws_lambda/__init__.py +++ b/tests/integrations/aws_lambda/__init__.py @@ -1,3 +1,5 @@ import pytest pytest.importorskip("boto3") +pytest.importorskip("fastapi") +pytest.importorskip("uvicorn") diff --git a/tests/integrations/aws_lambda/client.py b/tests/integrations/aws_lambda/client.py deleted file mode 100644 index afacf6fc42..0000000000 --- a/tests/integrations/aws_lambda/client.py +++ /dev/null @@ -1,408 +0,0 @@ -import base64 -import boto3 -import glob -import hashlib -import os -import subprocess -import sys -import tempfile - -from sentry_sdk.consts import VERSION as SDK_VERSION -from sentry_sdk.utils import get_git_revision - -AWS_REGION_NAME = "us-east-1" -AWS_CREDENTIALS = { - "aws_access_key_id": os.environ["SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID"], - "aws_secret_access_key": os.environ["SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY"], -} -AWS_LAMBDA_EXECUTION_ROLE_NAME = "lambda-ex" -AWS_LAMBDA_EXECUTION_ROLE_ARN = None - - -def _install_dependencies(base_dir, subprocess_kwargs): - """ - Installs dependencies for AWS Lambda function - """ - setup_cfg = os.path.join(base_dir, "setup.cfg") - with open(setup_cfg, "w") as f: - f.write("[install]\nprefix=") - - # Install requirements for Lambda Layer (these are more limited than the SDK requirements, - # because Lambda does not support the newest versions of some packages) - subprocess.check_call( - [ - sys.executable, - "-m", - "pip", - "install", - "-r", - "requirements-aws-lambda-layer.txt", - "--target", - base_dir, - ], - **subprocess_kwargs, - ) - # Install requirements used for testing - subprocess.check_call( - [ - sys.executable, - "-m", - "pip", - "install", - "mock==3.0.0", - "funcsigs", - "--target", - base_dir, - ], - **subprocess_kwargs, - ) - # Create a source distribution of the Sentry SDK (in parent directory of base_dir) - subprocess.check_call( - [ - sys.executable, - "setup.py", - "sdist", - "--dist-dir", - os.path.dirname(base_dir), - ], - **subprocess_kwargs, - ) - # Install the created Sentry SDK source distribution into the target directory - # Do not install the dependencies of the SDK, because they where installed by requirements-aws-lambda-layer.txt above - source_distribution_archive = glob.glob( - "{}/*.tar.gz".format(os.path.dirname(base_dir)) - )[0] - subprocess.check_call( - [ - sys.executable, - "-m", - "pip", - "install", - source_distribution_archive, - "--no-deps", - "--target", - base_dir, - ], - **subprocess_kwargs, - ) - - -def _create_lambda_function_zip(base_dir): - """ - Zips the given base_dir omitting Python cache files - """ - subprocess.run( - [ - "zip", - "-q", - "-x", - "**/__pycache__/*", - "-r", - "lambda-function-package.zip", - "./", - ], - cwd=base_dir, - check=True, - ) - - -def _create_lambda_package( - base_dir, code, initial_handler, layer, syntax_check, subprocess_kwargs -): - """ - Creates deployable packages (as zip files) for AWS Lambda function - and optional the accompanying Sentry Lambda layer - """ - if initial_handler: - # If Initial handler value is provided i.e. it is not the default - # `test_lambda.test_handler`, then create another dir level so that our path is - # test_dir.test_lambda.test_handler - test_dir_path = os.path.join(base_dir, "test_dir") - python_init_file = os.path.join(test_dir_path, "__init__.py") - os.makedirs(test_dir_path) - with open(python_init_file, "w"): - # Create __init__ file to make it a python package - pass - - test_lambda_py = os.path.join(base_dir, "test_dir", "test_lambda.py") - else: - test_lambda_py = os.path.join(base_dir, "test_lambda.py") - - with open(test_lambda_py, "w") as f: - f.write(code) - - if syntax_check: - # Check file for valid syntax first, and that the integration does not - # crash when not running in Lambda (but rather a local deployment tool - # such as chalice's) - subprocess.check_call([sys.executable, test_lambda_py]) - - if layer is None: - _install_dependencies(base_dir, subprocess_kwargs) - _create_lambda_function_zip(base_dir) - - else: - _create_lambda_function_zip(base_dir) - - # Create Lambda layer zip package - from scripts.build_aws_lambda_layer import build_packaged_zip - - build_packaged_zip( - base_dir=base_dir, - make_dist=True, - out_zip_filename="lambda-layer-package.zip", - ) - - -def _get_or_create_lambda_execution_role(): - global AWS_LAMBDA_EXECUTION_ROLE_ARN - - policy = """{ - "Version": "2012-10-17", - "Statement": [ - { - "Effect": "Allow", - "Principal": { - "Service": "lambda.amazonaws.com" - }, - "Action": "sts:AssumeRole" - } - ] - } - """ - iam_client = boto3.client( - "iam", - region_name=AWS_REGION_NAME, - **AWS_CREDENTIALS, - ) - - try: - response = iam_client.get_role(RoleName=AWS_LAMBDA_EXECUTION_ROLE_NAME) - AWS_LAMBDA_EXECUTION_ROLE_ARN = response["Role"]["Arn"] - except iam_client.exceptions.NoSuchEntityException: - # create role for lambda execution - response = iam_client.create_role( - RoleName=AWS_LAMBDA_EXECUTION_ROLE_NAME, - AssumeRolePolicyDocument=policy, - ) - AWS_LAMBDA_EXECUTION_ROLE_ARN = response["Role"]["Arn"] - - # attach policy to role - iam_client.attach_role_policy( - RoleName=AWS_LAMBDA_EXECUTION_ROLE_NAME, - PolicyArn="arn:aws:iam::aws:policy/service-role/AWSLambdaBasicExecutionRole", - ) - - -def get_boto_client(): - _get_or_create_lambda_execution_role() - - return boto3.client( - "lambda", - region_name=AWS_REGION_NAME, - **AWS_CREDENTIALS, - ) - - -def run_lambda_function( - client, - runtime, - code, - payload, - add_finalizer, - syntax_check=True, - timeout=30, - layer=None, - initial_handler=None, - subprocess_kwargs=(), -): - """ - Creates a Lambda function with the given code, and invokes it. - - If the same code is run multiple times the function will NOT be - created anew each time but the existing function will be reused. - """ - subprocess_kwargs = dict(subprocess_kwargs) - - # Making a unique function name depending on all the code that is run in it (function code plus SDK version) - # The name needs to be short so the generated event/envelope json blobs are small enough to be output - # in the log result of the Lambda function. - rev = get_git_revision() or SDK_VERSION - function_hash = hashlib.shake_256((code + rev).encode("utf-8")).hexdigest(6) - fn_name = "test_{}".format(function_hash) - full_fn_name = "{}_{}".format( - fn_name, runtime.replace(".", "").replace("python", "py") - ) - - function_exists_in_aws = True - try: - client.get_function( - FunctionName=full_fn_name, - ) - print( - "Lambda function in AWS already existing, taking it (and do not create a local one)" - ) - except client.exceptions.ResourceNotFoundException: - function_exists_in_aws = False - - if not function_exists_in_aws: - tmp_base_dir = tempfile.gettempdir() - base_dir = os.path.join(tmp_base_dir, fn_name) - dir_already_existing = os.path.isdir(base_dir) - - if dir_already_existing: - print("Local Lambda function directory already exists, skipping creation") - - if not dir_already_existing: - os.mkdir(base_dir) - _create_lambda_package( - base_dir, code, initial_handler, layer, syntax_check, subprocess_kwargs - ) - - @add_finalizer - def clean_up(): - # this closes the web socket so we don't get a - # ResourceWarning: unclosed - # warning on every test - # based on https://github.com/boto/botocore/pull/1810 - # (if that's ever merged, this can just become client.close()) - session = client._endpoint.http_session - managers = [session._manager] + list(session._proxy_managers.values()) - for manager in managers: - manager.clear() - - layers = [] - environment = {} - handler = initial_handler or "test_lambda.test_handler" - - if layer is not None: - with open( - os.path.join(base_dir, "lambda-layer-package.zip"), "rb" - ) as lambda_layer_zip: - response = client.publish_layer_version( - LayerName="python-serverless-sdk-test", - Description="Created as part of testsuite for getsentry/sentry-python", - Content={"ZipFile": lambda_layer_zip.read()}, - ) - - layers = [response["LayerVersionArn"]] - handler = ( - "sentry_sdk.integrations.init_serverless_sdk.sentry_lambda_handler" - ) - environment = { - "Variables": { - "SENTRY_INITIAL_HANDLER": initial_handler - or "test_lambda.test_handler", - "SENTRY_DSN": "https://123abc@example.com/123", - "SENTRY_TRACES_SAMPLE_RATE": "1.0", - } - } - - try: - with open( - os.path.join(base_dir, "lambda-function-package.zip"), "rb" - ) as lambda_function_zip: - client.create_function( - Description="Created as part of testsuite for getsentry/sentry-python", - FunctionName=full_fn_name, - Runtime=runtime, - Timeout=timeout, - Role=AWS_LAMBDA_EXECUTION_ROLE_ARN, - Handler=handler, - Code={"ZipFile": lambda_function_zip.read()}, - Environment=environment, - Layers=layers, - ) - - waiter = client.get_waiter("function_active_v2") - waiter.wait(FunctionName=full_fn_name) - except client.exceptions.ResourceConflictException: - print( - "Lambda function already exists, this is fine, we will just invoke it." - ) - - response = client.invoke( - FunctionName=full_fn_name, - InvocationType="RequestResponse", - LogType="Tail", - Payload=payload, - ) - - assert 200 <= response["StatusCode"] < 300, response - return response - - -# This is for inspecting new Python runtime environments in AWS Lambda -# If you need to debug a new runtime, use this REPL to run arbitrary Python or bash commands -# in that runtime in a Lambda function: -# -# pip3 install click -# python3 tests/integrations/aws_lambda/client.py --runtime=python4.0 -# - - -_REPL_CODE = """ -import os - -def test_handler(event, context): - line = {line!r} - if line.startswith(">>> "): - exec(line[4:]) - elif line.startswith("$ "): - os.system(line[2:]) - else: - print("Start a line with $ or >>>") - - return b"" -""" - -try: - import click -except ImportError: - pass -else: - - @click.command() - @click.option( - "--runtime", required=True, help="name of the runtime to use, eg python3.11" - ) - @click.option("--verbose", is_flag=True, default=False) - def repl(runtime, verbose): - """ - Launch a "REPL" against AWS Lambda to inspect their runtime. - """ - - cleanup = [] - client = get_boto_client() - - print("Start a line with `$ ` to run shell commands, or `>>> ` to run Python") - - while True: - line = input() - - response = run_lambda_function( - client, - runtime, - _REPL_CODE.format(line=line), - b"", - cleanup.append, - subprocess_kwargs=( - { - "stdout": subprocess.DEVNULL, - "stderr": subprocess.DEVNULL, - } - if not verbose - else {} - ), - ) - - for line in base64.b64decode(response["LogResult"]).splitlines(): - print(line.decode("utf8")) - - for f in cleanup: - f() - - cleanup = [] - - if __name__ == "__main__": - repl() diff --git a/tests/integrations/aws_lambda/lambda_functions/BasicException/index.py b/tests/integrations/aws_lambda/lambda_functions/BasicException/index.py new file mode 100644 index 0000000000..875b984e2a --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions/BasicException/index.py @@ -0,0 +1,6 @@ +def handler(event, context): + raise RuntimeError("Oh!") + + return { + "event": event, + } diff --git a/tests/integrations/aws_lambda/lambda_functions/BasicOk/index.py b/tests/integrations/aws_lambda/lambda_functions/BasicOk/index.py new file mode 100644 index 0000000000..257fea04f0 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions/BasicOk/index.py @@ -0,0 +1,4 @@ +def handler(event, context): + return { + "event": event, + } diff --git a/tests/integrations/aws_lambda/lambda_functions/InitError/index.py b/tests/integrations/aws_lambda/lambda_functions/InitError/index.py new file mode 100644 index 0000000000..20b4fcc111 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions/InitError/index.py @@ -0,0 +1,3 @@ +# We have no handler() here and try to call a non-existing function. + +func() # noqa: F821 diff --git a/tests/integrations/aws_lambda/lambda_functions/TimeoutError/index.py b/tests/integrations/aws_lambda/lambda_functions/TimeoutError/index.py new file mode 100644 index 0000000000..01334bbfbc --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions/TimeoutError/index.py @@ -0,0 +1,8 @@ +import time + + +def handler(event, context): + time.sleep(15) + return { + "event": event, + } diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore new file mode 100644 index 0000000000..ee0b7b9305 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore @@ -0,0 +1,11 @@ +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. + +# Ignore everything +* + +# But not index.py +!index.py + +# And not .gitignore itself +!.gitignore \ No newline at end of file diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/index.py b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/index.py new file mode 100644 index 0000000000..12f43f0009 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/index.py @@ -0,0 +1,14 @@ +import os +import sentry_sdk +from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration + + +sentry_sdk.init( + dsn=os.environ.get("SENTRY_DSN"), + traces_sample_rate=None, # this is the default, just added for clarity + integrations=[AwsLambdaIntegration()], +) + + +def handler(event, context): + raise Exception("Oh!") diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore new file mode 100644 index 0000000000..ee0b7b9305 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore @@ -0,0 +1,11 @@ +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. + +# Ignore everything +* + +# But not index.py +!index.py + +# And not .gitignore itself +!.gitignore \ No newline at end of file diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/index.py b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/index.py new file mode 100644 index 0000000000..c694299682 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/index.py @@ -0,0 +1,14 @@ +import os +import sentry_sdk +from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration + + +sentry_sdk.init( + dsn=os.environ.get("SENTRY_DSN"), + traces_sample_rate=1.0, + integrations=[AwsLambdaIntegration()], +) + + +def handler(event, context): + raise Exception("Oh!") diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore new file mode 100644 index 0000000000..ee0b7b9305 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore @@ -0,0 +1,11 @@ +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. + +# Ignore everything +* + +# But not index.py +!index.py + +# And not .gitignore itself +!.gitignore \ No newline at end of file diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/index.py b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/index.py new file mode 100644 index 0000000000..ce797faf71 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/index.py @@ -0,0 +1,49 @@ +import json +import os +import sentry_sdk +from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration + +# Global variables to store sampling context for verification +sampling_context_data = { + "aws_event_present": False, + "aws_context_present": False, + "event_data": None, +} + + +def trace_sampler(sampling_context): + # Store the sampling context for verification + global sampling_context_data + + # Check if aws_event and aws_context are in the sampling_context + if "aws_event" in sampling_context: + sampling_context_data["aws_event_present"] = True + sampling_context_data["event_data"] = sampling_context["aws_event"] + + if "aws_context" in sampling_context: + sampling_context_data["aws_context_present"] = True + + print("Sampling context data:", sampling_context_data) + return 1.0 # Always sample + + +sentry_sdk.init( + dsn=os.environ.get("SENTRY_DSN"), + traces_sample_rate=1.0, + traces_sampler=trace_sampler, + integrations=[AwsLambdaIntegration()], +) + + +def handler(event, context): + # Return the sampling context data for verification + return { + "statusCode": 200, + "body": json.dumps( + { + "message": "Hello from Lambda with embedded Sentry SDK!", + "event": event, + "sampling_context_data": sampling_context_data, + } + ), + } diff --git a/tests/integrations/aws_lambda/test_aws.py b/tests/integrations/aws_lambda/test_aws.py deleted file mode 100644 index 8bbd33505b..0000000000 --- a/tests/integrations/aws_lambda/test_aws.py +++ /dev/null @@ -1,898 +0,0 @@ -""" -# AWS Lambda System Tests - -This testsuite uses boto3 to upload actual Lambda functions to AWS Lambda and invoke them. - -For running test locally you need to set these env vars: -(You can find the values in the Sentry password manager by searching for "AWS Lambda for Python SDK Tests"). - - export SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID="..." - export SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY="..." - - -You can use `scripts/aws-cleanup.sh` to delete all files generated by this test suite. - - -If you need to debug a new runtime, use this REPL to run arbitrary Python or bash commands -in that runtime in a Lambda function: (see the bottom of client.py for more information.) - - pip3 install click - python3 tests/integrations/aws_lambda/client.py --runtime=python4.0 - -IMPORTANT: - -During running of this test suite temporary folders will be created for compiling the Lambda functions. -This temporary folders will not be cleaned up. This is because in CI generated files have to be shared -between tests and thus the folders can not be deleted right after use. - -If you run your tests locally, you need to clean up the temporary folders manually. The location of -the temporary folders is printed when running a test. -""" - -import base64 -import json -import re -from textwrap import dedent - -import pytest - -RUNTIMES_TO_TEST = [ - "python3.8", - "python3.10", - "python3.12", - "python3.13", -] - -LAMBDA_PRELUDE = """ -from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration, get_lambda_bootstrap -import sentry_sdk -import json -import time - -from sentry_sdk.transport import Transport - -def truncate_data(data): - # AWS Lambda truncates the log output to 4kb, which is small enough to miss - # parts of even a single error-event/transaction-envelope pair if considered - # in full, so only grab the data we need. - - cleaned_data = {} - - if data.get("type") is not None: - cleaned_data["type"] = data["type"] - - if data.get("contexts") is not None: - cleaned_data["contexts"] = {} - - if data["contexts"].get("trace") is not None: - cleaned_data["contexts"]["trace"] = data["contexts"].get("trace") - - if data.get("transaction") is not None: - cleaned_data["transaction"] = data.get("transaction") - - if data.get("request") is not None: - cleaned_data["request"] = data.get("request") - - if data.get("tags") is not None: - cleaned_data["tags"] = data.get("tags") - - if data.get("exception") is not None: - cleaned_data["exception"] = data.get("exception") - - for value in cleaned_data["exception"]["values"]: - for frame in value.get("stacktrace", {}).get("frames", []): - del frame["vars"] - del frame["pre_context"] - del frame["context_line"] - del frame["post_context"] - - if data.get("extra") is not None: - cleaned_data["extra"] = {} - - for key in data["extra"].keys(): - if key == "lambda": - for lambda_key in data["extra"]["lambda"].keys(): - if lambda_key in ["function_name"]: - cleaned_data["extra"].setdefault("lambda", {})[lambda_key] = data["extra"]["lambda"][lambda_key] - elif key == "cloudwatch logs": - for cloudwatch_key in data["extra"]["cloudwatch logs"].keys(): - if cloudwatch_key in ["url", "log_group", "log_stream"]: - cleaned_data["extra"].setdefault("cloudwatch logs", {})[cloudwatch_key] = data["extra"]["cloudwatch logs"][cloudwatch_key].split("=")[0] - - if data.get("level") is not None: - cleaned_data["level"] = data.get("level") - - if data.get("message") is not None: - cleaned_data["message"] = data.get("message") - - if "contexts" not in cleaned_data: - raise Exception(json.dumps(data)) - - return cleaned_data - -def event_processor(event): - return truncate_data(event) - -def envelope_processor(envelope): - (item,) = envelope.items - item_json = json.loads(item.get_bytes()) - - return truncate_data(item_json) - - -class TestTransport(Transport): - def capture_envelope(self, envelope): - envelope_items = envelope_processor(envelope) - print("\\nENVELOPE: {}\\n".format(json.dumps(envelope_items))) - -def init_sdk(timeout_warning=False, **extra_init_args): - sentry_sdk.init( - dsn="https://123abc@example.com/123", - transport=TestTransport, - integrations=[AwsLambdaIntegration(timeout_warning=timeout_warning)], - shutdown_timeout=10, - **extra_init_args - ) -""" - - -@pytest.fixture -def lambda_client(): - from tests.integrations.aws_lambda.client import get_boto_client - - return get_boto_client() - - -@pytest.fixture(params=RUNTIMES_TO_TEST) -def lambda_runtime(request): - return request.param - - -@pytest.fixture -def run_lambda_function(request, lambda_client, lambda_runtime): - def inner( - code, payload, timeout=30, syntax_check=True, layer=None, initial_handler=None - ): - from tests.integrations.aws_lambda.client import run_lambda_function - - response = run_lambda_function( - client=lambda_client, - runtime=lambda_runtime, - code=code, - payload=payload, - add_finalizer=request.addfinalizer, - timeout=timeout, - syntax_check=syntax_check, - layer=layer, - initial_handler=initial_handler, - ) - - # Make sure the "ENVELOPE:" and "EVENT:" log entries are always starting a new line. (Sometimes they don't.) - response["LogResult"] = ( - base64.b64decode(response["LogResult"]) - .replace(b"EVENT:", b"\nEVENT:") - .replace(b"ENVELOPE:", b"\nENVELOPE:") - .splitlines() - ) - response["Payload"] = json.loads(response["Payload"].read().decode("utf-8")) - del response["ResponseMetadata"] - - envelope_items = [] - - for line in response["LogResult"]: - print("AWS:", line) - if line.startswith(b"ENVELOPE: "): - line = line[len(b"ENVELOPE: ") :] - envelope_items.append(json.loads(line.decode("utf-8"))) - else: - continue - - return envelope_items, response - - return inner - - -def test_basic(run_lambda_function): - envelope_items, response = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk() - - def test_handler(event, context): - raise Exception("Oh!") - """ - ), - b'{"foo": "bar"}', - ) - - assert response["FunctionError"] == "Unhandled" - - (event,) = envelope_items - assert event["level"] == "error" - (exception,) = event["exception"]["values"] - assert exception["type"] == "Exception" - assert exception["value"] == "Oh!" - - (frame1,) = exception["stacktrace"]["frames"] - assert frame1["filename"] == "test_lambda.py" - assert frame1["abs_path"] == "/var/task/test_lambda.py" - assert frame1["function"] == "test_handler" - - assert frame1["in_app"] is True - - assert exception["mechanism"]["type"] == "aws_lambda" - assert not exception["mechanism"]["handled"] - - assert event["extra"]["lambda"]["function_name"].startswith("test_") - - logs_url = event["extra"]["cloudwatch logs"]["url"] - assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region") - assert not re.search("(=;|=$)", logs_url) - assert event["extra"]["cloudwatch logs"]["log_group"].startswith( - "/aws/lambda/test_" - ) - - log_stream_re = "^[0-9]{4}/[0-9]{2}/[0-9]{2}/\\[[^\\]]+][a-f0-9]+$" - log_stream = event["extra"]["cloudwatch logs"]["log_stream"] - - assert re.match(log_stream_re, log_stream) - - -def test_initialization_order(run_lambda_function): - """Zappa lazily imports our code, so by the time we monkeypatch the handler - as seen by AWS already runs. At this point at least draining the queue - should work.""" - - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - def test_handler(event, context): - init_sdk() - sentry_sdk.capture_exception(Exception("Oh!")) - """ - ), - b'{"foo": "bar"}', - ) - - (event,) = envelope_items - - assert event["level"] == "error" - (exception,) = event["exception"]["values"] - assert exception["type"] == "Exception" - assert exception["value"] == "Oh!" - - -def test_request_data(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk() - def test_handler(event, context): - sentry_sdk.capture_message("hi") - return "ok" - """ - ), - payload=b""" - { - "resource": "/asd", - "path": "/asd", - "httpMethod": "GET", - "headers": { - "Host": "iwsz2c7uwi.execute-api.us-east-1.amazonaws.com", - "User-Agent": "custom", - "X-Forwarded-Proto": "https" - }, - "queryStringParameters": { - "bonkers": "true" - }, - "pathParameters": null, - "stageVariables": null, - "requestContext": { - "identity": { - "sourceIp": "213.47.147.207", - "userArn": "42" - } - }, - "body": null, - "isBase64Encoded": false - } - """, - ) - - (event,) = envelope_items - - assert event["request"] == { - "headers": { - "Host": "iwsz2c7uwi.execute-api.us-east-1.amazonaws.com", - "User-Agent": "custom", - "X-Forwarded-Proto": "https", - }, - "method": "GET", - "query_string": {"bonkers": "true"}, - "url": "https://iwsz2c7uwi.execute-api.us-east-1.amazonaws.com/asd", - } - - -def test_init_error(run_lambda_function, lambda_runtime): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk() - func() - """ - ), - b'{"foo": "bar"}', - syntax_check=False, - ) - - # We just take the last one, because it could be that in the output of the Lambda - # invocation there is still the envelope of the previous invocation of the function. - event = envelope_items[-1] - assert event["exception"]["values"][0]["value"] == "name 'func' is not defined" - - -def test_timeout_error(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(timeout_warning=True) - - def test_handler(event, context): - time.sleep(10) - return 0 - """ - ), - b'{"foo": "bar"}', - timeout=2, - ) - - (event,) = envelope_items - assert event["level"] == "error" - (exception,) = event["exception"]["values"] - assert exception["type"] == "ServerlessTimeoutWarning" - assert exception["value"] in ( - "WARNING : Function is expected to get timed out. Configured timeout duration = 3 seconds.", - "WARNING : Function is expected to get timed out. Configured timeout duration = 2 seconds.", - ) - - assert exception["mechanism"]["type"] == "threading" - assert not exception["mechanism"]["handled"] - - assert event["extra"]["lambda"]["function_name"].startswith("test_") - - logs_url = event["extra"]["cloudwatch logs"]["url"] - assert logs_url.startswith("https://console.aws.amazon.com/cloudwatch/home?region") - assert not re.search("(=;|=$)", logs_url) - assert event["extra"]["cloudwatch logs"]["log_group"].startswith( - "/aws/lambda/test_" - ) - - log_stream_re = "^[0-9]{4}/[0-9]{2}/[0-9]{2}/\\[[^\\]]+][a-f0-9]+$" - log_stream = event["extra"]["cloudwatch logs"]["log_stream"] - - assert re.match(log_stream_re, log_stream) - - -def test_performance_no_error(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - return "test_string" - """ - ), - b'{"foo": "bar"}', - ) - - (envelope,) = envelope_items - - assert envelope["type"] == "transaction" - assert envelope["contexts"]["trace"]["op"] == "function.aws" - assert envelope["transaction"].startswith("test_") - assert envelope["transaction"] in envelope["request"]["url"] - - -def test_performance_error(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - raise Exception("Oh!") - """ - ), - b'{"foo": "bar"}', - ) - - ( - error_event, - transaction_event, - ) = envelope_items - - assert error_event["level"] == "error" - (exception,) = error_event["exception"]["values"] - assert exception["type"] == "Exception" - assert exception["value"] == "Oh!" - - assert transaction_event["type"] == "transaction" - assert transaction_event["contexts"]["trace"]["op"] == "function.aws" - assert transaction_event["transaction"].startswith("test_") - assert transaction_event["transaction"] in transaction_event["request"]["url"] - - -@pytest.mark.parametrize( - "aws_event, has_request_data, batch_size", - [ - (b"1231", False, 1), - (b"11.21", False, 1), - (b'"Good dog!"', False, 1), - (b"true", False, 1), - ( - b""" - [ - {"good dog": "Maisey"}, - {"good dog": "Charlie"}, - {"good dog": "Cory"}, - {"good dog": "Bodhi"} - ] - """, - False, - 4, - ), - ( - b""" - [ - { - "headers": { - "Host": "x1.io", - "X-Forwarded-Proto": "https" - }, - "httpMethod": "GET", - "path": "/1", - "queryStringParameters": { - "done": "f" - }, - "d": "D1" - }, - { - "headers": { - "Host": "x2.io", - "X-Forwarded-Proto": "http" - }, - "httpMethod": "POST", - "path": "/2", - "queryStringParameters": { - "done": "t" - }, - "d": "D2" - } - ] - """, - True, - 2, - ), - (b"[]", False, 1), - ], -) -def test_non_dict_event( - run_lambda_function, - aws_event, - has_request_data, - batch_size, - DictionaryContaining, # noqa:N803 -): - envelope_items, response = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - raise Exception("Oh?") - """ - ), - aws_event, - ) - - assert response["FunctionError"] == "Unhandled" - - ( - error_event, - transaction_event, - ) = envelope_items - assert error_event["level"] == "error" - assert error_event["contexts"]["trace"]["op"] == "function.aws" - - function_name = error_event["extra"]["lambda"]["function_name"] - assert function_name.startswith("test_") - assert error_event["transaction"] == function_name - - exception = error_event["exception"]["values"][0] - assert exception["type"] == "Exception" - assert exception["value"] == "Oh?" - assert exception["mechanism"]["type"] == "aws_lambda" - - assert transaction_event["type"] == "transaction" - assert transaction_event["contexts"]["trace"] == DictionaryContaining( - error_event["contexts"]["trace"] - ) - assert transaction_event["contexts"]["trace"]["status"] == "internal_error" - assert transaction_event["transaction"] == error_event["transaction"] - assert transaction_event["request"]["url"] == error_event["request"]["url"] - - if has_request_data: - request_data = { - "headers": {"Host": "x1.io", "X-Forwarded-Proto": "https"}, - "method": "GET", - "url": "https://x1.io/1", - "query_string": { - "done": "f", - }, - } - else: - request_data = {"url": "awslambda:///{}".format(function_name)} - - assert error_event["request"] == request_data - assert transaction_event["request"] == request_data - - if batch_size > 1: - assert error_event["tags"]["batch_size"] == batch_size - assert error_event["tags"]["batch_request"] is True - assert transaction_event["tags"]["batch_size"] == batch_size - assert transaction_event["tags"]["batch_request"] is True - - -def test_traces_sampler_gets_correct_values_in_sampling_context( - run_lambda_function, - DictionaryContaining, # noqa: N803 - ObjectDescribedBy, # noqa: N803 - StringContaining, # noqa: N803 -): - # TODO: This whole thing is a little hacky, specifically around the need to - # get `conftest.py` code into the AWS runtime, which is why there's both - # `inspect.getsource` and a copy of `_safe_is_equal` included directly in - # the code below. Ideas which have been discussed to fix this: - - # - Include the test suite as a module installed in the package which is - # shot up to AWS - # - In client.py, copy `conftest.py` (or wherever the necessary code lives) - # from the test suite into the main SDK directory so it gets included as - # "part of the SDK" - - # It's also worth noting why it's necessary to run the assertions in the AWS - # runtime rather than asserting on side effects the way we do with events - # and envelopes. The reasons are two-fold: - - # - We're testing against the `LambdaContext` class, which only exists in - # the AWS runtime - # - If we were to transmit call args data they way we transmit event and - # envelope data (through JSON), we'd quickly run into the problem that all - # sorts of stuff isn't serializable by `json.dumps` out of the box, up to - # and including `datetime` objects (so anything with a timestamp is - # automatically out) - - # Perhaps these challenges can be solved in a cleaner and more systematic - # way if we ever decide to refactor the entire AWS testing apparatus. - - import inspect - - _, response = run_lambda_function( - LAMBDA_PRELUDE - + dedent(inspect.getsource(StringContaining)) - + dedent(inspect.getsource(DictionaryContaining)) - + dedent(inspect.getsource(ObjectDescribedBy)) - + dedent( - """ - from unittest import mock - - def _safe_is_equal(x, y): - # copied from conftest.py - see docstring and comments there - try: - is_equal = x.__eq__(y) - except AttributeError: - is_equal = NotImplemented - - if is_equal == NotImplemented: - # using == smoothes out weird variations exposed by raw __eq__ - return x == y - - return is_equal - - def test_handler(event, context): - # this runs after the transaction has started, which means we - # can make assertions about traces_sampler - try: - traces_sampler.assert_any_call( - DictionaryContaining( - { - "aws_event": DictionaryContaining({ - "httpMethod": "GET", - "path": "/sit/stay/rollover", - "headers": {"Host": "x.io", "X-Forwarded-Proto": "http"}, - }), - "aws_context": ObjectDescribedBy( - type=get_lambda_bootstrap().LambdaContext, - attrs={ - 'function_name': StringContaining("test_"), - 'function_version': '$LATEST', - } - ) - } - ) - ) - except AssertionError: - # catch the error and return it because the error itself will - # get swallowed by the SDK as an "internal exception" - return {"AssertionError raised": True,} - - return {"AssertionError raised": False,} - - - traces_sampler = mock.Mock(return_value=True) - - init_sdk( - traces_sampler=traces_sampler, - ) - """ - ), - b'{"httpMethod": "GET", "path": "/sit/stay/rollover", "headers": {"Host": "x.io", "X-Forwarded-Proto": "http"}}', - ) - - assert response["Payload"]["AssertionError raised"] is False - - -@pytest.mark.xfail( - reason="The limited log output we depend on is being clogged by a new warning" -) -def test_serverless_no_code_instrumentation(run_lambda_function): - """ - Test that ensures that just by adding a lambda layer containing the - python sdk, with no code changes sentry is able to capture errors - """ - - for initial_handler in [ - None, - "test_dir/test_lambda.test_handler", - "test_dir.test_lambda.test_handler", - ]: - print("Testing Initial Handler ", initial_handler) - _, response = run_lambda_function( - dedent( - """ - import sentry_sdk - - def test_handler(event, context): - current_client = sentry_sdk.get_client() - - assert current_client.is_active() - - assert len(current_client.options['integrations']) == 1 - assert isinstance(current_client.options['integrations'][0], - sentry_sdk.integrations.aws_lambda.AwsLambdaIntegration) - - raise Exception("Oh!") - """ - ), - b'{"foo": "bar"}', - layer=True, - initial_handler=initial_handler, - ) - assert response["FunctionError"] == "Unhandled" - assert response["StatusCode"] == 200 - - assert response["Payload"]["errorType"] != "AssertionError" - - assert response["Payload"]["errorType"] == "Exception" - assert response["Payload"]["errorMessage"] == "Oh!" - - assert "sentry_handler" in response["LogResult"][3].decode("utf-8") - - -@pytest.mark.xfail( - reason="The limited log output we depend on is being clogged by a new warning" -) -def test_error_has_new_trace_context_performance_enabled(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - sentry_sdk.capture_message("hi") - raise Exception("Oh!") - """ - ), - payload=b'{"foo": "bar"}', - ) - - (msg_event, error_event, transaction_event) = envelope_items - - assert "trace" in msg_event["contexts"] - assert "trace_id" in msg_event["contexts"]["trace"] - - assert "trace" in error_event["contexts"] - assert "trace_id" in error_event["contexts"]["trace"] - - assert "trace" in transaction_event["contexts"] - assert "trace_id" in transaction_event["contexts"]["trace"] - - assert ( - msg_event["contexts"]["trace"]["trace_id"] - == error_event["contexts"]["trace"]["trace_id"] - == transaction_event["contexts"]["trace"]["trace_id"] - ) - - -def test_error_has_new_trace_context_performance_disabled(run_lambda_function): - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=None) # this is the default, just added for clarity - - def test_handler(event, context): - sentry_sdk.capture_message("hi") - raise Exception("Oh!") - """ - ), - payload=b'{"foo": "bar"}', - ) - - (msg_event, error_event) = envelope_items - - assert "trace" in msg_event["contexts"] - assert "trace_id" in msg_event["contexts"]["trace"] - - assert "trace" in error_event["contexts"] - assert "trace_id" in error_event["contexts"]["trace"] - - assert ( - msg_event["contexts"]["trace"]["trace_id"] - == error_event["contexts"]["trace"]["trace_id"] - ) - - -@pytest.mark.xfail( - reason="The limited log output we depend on is being clogged by a new warning" -) -def test_error_has_existing_trace_context_performance_enabled(run_lambda_function): - trace_id = "471a43a4192642f0b136d5159a501701" - parent_span_id = "6e8f22c393e68f19" - parent_sampled = 1 - sentry_trace_header = "{}-{}-{}".format(trace_id, parent_span_id, parent_sampled) - - # We simulate here AWS Api Gateway's behavior of passing HTTP headers - # as the `headers` dict in the event passed to the Lambda function. - payload = { - "headers": { - "sentry-trace": sentry_trace_header, - } - } - - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - sentry_sdk.capture_message("hi") - raise Exception("Oh!") - """ - ), - payload=json.dumps(payload).encode(), - ) - - (msg_event, error_event, transaction_event) = envelope_items - - assert "trace" in msg_event["contexts"] - assert "trace_id" in msg_event["contexts"]["trace"] - - assert "trace" in error_event["contexts"] - assert "trace_id" in error_event["contexts"]["trace"] - - assert "trace" in transaction_event["contexts"] - assert "trace_id" in transaction_event["contexts"]["trace"] - - assert ( - msg_event["contexts"]["trace"]["trace_id"] - == error_event["contexts"]["trace"]["trace_id"] - == transaction_event["contexts"]["trace"]["trace_id"] - == "471a43a4192642f0b136d5159a501701" - ) - - -def test_error_has_existing_trace_context_performance_disabled(run_lambda_function): - trace_id = "471a43a4192642f0b136d5159a501701" - parent_span_id = "6e8f22c393e68f19" - parent_sampled = 1 - sentry_trace_header = "{}-{}-{}".format(trace_id, parent_span_id, parent_sampled) - - # We simulate here AWS Api Gateway's behavior of passing HTTP headers - # as the `headers` dict in the event passed to the Lambda function. - payload = { - "headers": { - "sentry-trace": sentry_trace_header, - } - } - - envelope_items, _ = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=None) # this is the default, just added for clarity - - def test_handler(event, context): - sentry_sdk.capture_message("hi") - raise Exception("Oh!") - """ - ), - payload=json.dumps(payload).encode(), - ) - - (msg_event, error_event) = envelope_items - - assert "trace" in msg_event["contexts"] - assert "trace_id" in msg_event["contexts"]["trace"] - - assert "trace" in error_event["contexts"] - assert "trace_id" in error_event["contexts"]["trace"] - - assert ( - msg_event["contexts"]["trace"]["trace_id"] - == error_event["contexts"]["trace"]["trace_id"] - == "471a43a4192642f0b136d5159a501701" - ) - - -def test_basic_with_eventbridge_source(run_lambda_function): - envelope_items, response = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk() - - def test_handler(event, context): - raise Exception("Oh!") - """ - ), - b'[{"topic":"lps-ranges","partition":1,"offset":0,"timestamp":1701268939207,"timestampType":"CREATE_TIME","key":"REDACTED","value":"REDACTED","headers":[],"eventSourceArn":"REDACTED","bootstrapServers":"REDACTED","eventSource":"aws:kafka","eventSourceKey":"lps-ranges-1"}]', - ) - - assert response["FunctionError"] == "Unhandled" - - (event,) = envelope_items - assert event["level"] == "error" - (exception,) = event["exception"]["values"] - assert exception["type"] == "Exception" - assert exception["value"] == "Oh!" - - -def test_span_origin(run_lambda_function): - envelope_items, response = run_lambda_function( - LAMBDA_PRELUDE - + dedent( - """ - init_sdk(traces_sample_rate=1.0) - - def test_handler(event, context): - pass - """ - ), - b'{"foo": "bar"}', - ) - - (event,) = envelope_items - - assert event["contexts"]["trace"]["origin"] == "auto.function.aws_lambda" diff --git a/tests/integrations/aws_lambda/test_aws_lambda.py b/tests/integrations/aws_lambda/test_aws_lambda.py new file mode 100644 index 0000000000..85da7e0b14 --- /dev/null +++ b/tests/integrations/aws_lambda/test_aws_lambda.py @@ -0,0 +1,550 @@ +import boto3 +import docker +import json +import pytest +import subprocess +import tempfile +import time +import yaml + +from unittest import mock + +from aws_cdk import App + +from .utils import LocalLambdaStack, SentryServerForTesting, SAM_PORT + + +DOCKER_NETWORK_NAME = "lambda-test-network" +SAM_TEMPLATE_FILE = "sam.template.yaml" + + +@pytest.fixture(scope="session", autouse=True) +def test_environment(): + print("[test_environment fixture] Setting up AWS Lambda test infrastructure") + + # Create a Docker network + docker_client = docker.from_env() + docker_client.networks.prune() + docker_client.networks.create(DOCKER_NETWORK_NAME, driver="bridge") + + # Start Sentry server + server = SentryServerForTesting() + server.start() + time.sleep(1) # Give it a moment to start up + + # Create local AWS SAM stack + app = App() + stack = LocalLambdaStack(app, "LocalLambdaStack") + + # Write SAM template to file + template = app.synth().get_stack_by_name("LocalLambdaStack").template + with open(SAM_TEMPLATE_FILE, "w") as f: + yaml.dump(template, f) + + # Write SAM debug log to file + debug_log_file = tempfile.gettempdir() + "/sentry_aws_lambda_tests_sam_debug.log" + debug_log = open(debug_log_file, "w") + print("[test_environment fixture] Writing SAM debug log to: %s" % debug_log_file) + + # Start SAM local + process = subprocess.Popen( + [ + "sam", + "local", + "start-lambda", + "--debug", + "--template", + SAM_TEMPLATE_FILE, + "--warm-containers", + "EAGER", + "--docker-network", + DOCKER_NETWORK_NAME, + ], + stdout=debug_log, + stderr=debug_log, + text=True, # This makes stdout/stderr return strings instead of bytes + ) + + try: + # Wait for SAM to be ready + LocalLambdaStack.wait_for_stack() + + def before_test(): + server.clear_envelopes() + + yield { + "stack": stack, + "server": server, + "before_test": before_test, + } + + finally: + print("[test_environment fixture] Tearing down AWS Lambda test infrastructure") + + process.terminate() + process.wait(timeout=5) # Give it time to shut down gracefully + + # Force kill if still running + if process.poll() is None: + process.kill() + + +@pytest.fixture(autouse=True) +def clear_before_test(test_environment): + test_environment["before_test"]() + + +@pytest.fixture +def lambda_client(): + """ + Create a boto3 client configured to use the local AWS SAM instance. + """ + return boto3.client( + "lambda", + endpoint_url=f"http://127.0.0.1:{SAM_PORT}", # noqa: E231 + aws_access_key_id="dummy", + aws_secret_access_key="dummy", + region_name="us-east-1", + ) + + +def test_basic_no_exception(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="BasicOk", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + (transaction_event,) = envelopes + + assert transaction_event["type"] == "transaction" + assert transaction_event["transaction"] == "BasicOk" + assert transaction_event["sdk"]["name"] == "sentry.python.aws_lambda" + assert transaction_event["tags"] == {"aws_region": "us-east-1"} + + assert transaction_event["extra"]["cloudwatch logs"] == { + "log_group": mock.ANY, + "log_stream": mock.ANY, + "url": mock.ANY, + } + assert transaction_event["extra"]["lambda"] == { + "aws_request_id": mock.ANY, + "execution_duration_in_millis": mock.ANY, + "function_name": "BasicOk", + "function_version": "$LATEST", + "invoked_function_arn": "arn:aws:lambda:us-east-1:012345678912:function:BasicOk", + "remaining_time_in_millis": mock.ANY, + } + assert transaction_event["contexts"]["trace"] == { + "op": "function.aws", + "description": mock.ANY, + "span_id": mock.ANY, + "parent_span_id": mock.ANY, + "trace_id": mock.ANY, + "origin": "auto.function.aws_lambda", + "data": mock.ANY, + } + + +def test_basic_exception(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="BasicException", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + # The second envelope we ignore. + # It is the transaction that we test in test_basic_no_exception. + (error_event, _) = envelopes + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "RuntimeError" + assert error_event["exception"]["values"][0]["value"] == "Oh!" + assert error_event["sdk"]["name"] == "sentry.python.aws_lambda" + + assert error_event["tags"] == {"aws_region": "us-east-1"} + assert error_event["extra"]["cloudwatch logs"] == { + "log_group": mock.ANY, + "log_stream": mock.ANY, + "url": mock.ANY, + } + assert error_event["extra"]["lambda"] == { + "aws_request_id": mock.ANY, + "execution_duration_in_millis": mock.ANY, + "function_name": "BasicException", + "function_version": "$LATEST", + "invoked_function_arn": "arn:aws:lambda:us-east-1:012345678912:function:BasicException", + "remaining_time_in_millis": mock.ANY, + } + assert error_event["contexts"]["trace"] == { + "op": "function.aws", + "description": mock.ANY, + "span_id": mock.ANY, + "parent_span_id": mock.ANY, + "trace_id": mock.ANY, + "origin": "auto.function.aws_lambda", + "data": mock.ANY, + } + + +def test_init_error(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="InitError", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + (error_event, transaction_event) = envelopes + + assert ( + error_event["exception"]["values"][0]["value"] == "name 'func' is not defined" + ) + assert transaction_event["transaction"] == "InitError" + + +def test_timeout_error(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="TimeoutError", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + (error_event,) = envelopes + + assert error_event["level"] == "error" + assert error_event["extra"]["lambda"]["function_name"] == "TimeoutError" + + (exception,) = error_event["exception"]["values"] + assert not exception["mechanism"]["handled"] + assert exception["type"] == "ServerlessTimeoutWarning" + assert exception["value"].startswith( + "WARNING : Function is expected to get timed out. Configured timeout duration =" + ) + assert exception["mechanism"]["type"] == "threading" + + +@pytest.mark.parametrize( + "aws_event, has_request_data, batch_size", + [ + (b"1231", False, 1), + (b"11.21", False, 1), + (b'"Good dog!"', False, 1), + (b"true", False, 1), + ( + b""" + [ + {"good dog": "Maisey"}, + {"good dog": "Charlie"}, + {"good dog": "Cory"}, + {"good dog": "Bodhi"} + ] + """, + False, + 4, + ), + ( + b""" + [ + { + "headers": { + "Host": "x1.io", + "X-Forwarded-Proto": "https" + }, + "httpMethod": "GET", + "path": "/1", + "queryStringParameters": { + "done": "f" + }, + "d": "D1" + }, + { + "headers": { + "Host": "x2.io", + "X-Forwarded-Proto": "http" + }, + "httpMethod": "POST", + "path": "/2", + "queryStringParameters": { + "done": "t" + }, + "d": "D2" + } + ] + """, + True, + 2, + ), + (b"[]", False, 1), + ], + ids=[ + "event as integer", + "event as float", + "event as string", + "event as bool", + "event as list of dicts", + "event as dict", + "event as empty list", + ], +) +def test_non_dict_event( + lambda_client, test_environment, aws_event, has_request_data, batch_size +): + lambda_client.invoke( + FunctionName="BasicException", + Payload=aws_event, + ) + envelopes = test_environment["server"].envelopes + + (error_event, transaction_event) = envelopes + + assert transaction_event["type"] == "transaction" + assert transaction_event["transaction"] == "BasicException" + assert transaction_event["sdk"]["name"] == "sentry.python.aws_lambda" + assert transaction_event["contexts"]["trace"]["status"] == "internal_error" + + assert error_event["level"] == "error" + assert error_event["transaction"] == "BasicException" + assert error_event["sdk"]["name"] == "sentry.python.aws_lambda" + assert error_event["exception"]["values"][0]["type"] == "RuntimeError" + assert error_event["exception"]["values"][0]["value"] == "Oh!" + assert error_event["exception"]["values"][0]["mechanism"]["type"] == "aws_lambda" + + if has_request_data: + request_data = { + "headers": {"Host": "x1.io", "X-Forwarded-Proto": "https"}, + "method": "GET", + "url": "https://x1.io/1", + "query_string": { + "done": "f", + }, + } + else: + request_data = {"url": "awslambda:///BasicException"} + + assert error_event["request"] == request_data + assert transaction_event["request"] == request_data + + if batch_size > 1: + assert error_event["tags"]["batch_size"] == batch_size + assert error_event["tags"]["batch_request"] is True + assert transaction_event["tags"]["batch_size"] == batch_size + assert transaction_event["tags"]["batch_request"] is True + + +def test_request_data(lambda_client, test_environment): + payload = b""" + { + "resource": "/asd", + "path": "/asd", + "httpMethod": "GET", + "headers": { + "Host": "iwsz2c7uwi.execute-api.us-east-1.amazonaws.com", + "User-Agent": "custom", + "X-Forwarded-Proto": "https" + }, + "queryStringParameters": { + "bonkers": "true" + }, + "pathParameters": null, + "stageVariables": null, + "requestContext": { + "identity": { + "sourceIp": "213.47.147.207", + "userArn": "42" + } + }, + "body": null, + "isBase64Encoded": false + } + """ + + lambda_client.invoke( + FunctionName="BasicOk", + Payload=payload, + ) + envelopes = test_environment["server"].envelopes + + (transaction_event,) = envelopes + + assert transaction_event["request"] == { + "headers": { + "Host": "iwsz2c7uwi.execute-api.us-east-1.amazonaws.com", + "User-Agent": "custom", + "X-Forwarded-Proto": "https", + }, + "method": "GET", + "query_string": {"bonkers": "true"}, + "url": "https://iwsz2c7uwi.execute-api.us-east-1.amazonaws.com/asd", + } + + +def test_trace_continuation(lambda_client, test_environment): + trace_id = "471a43a4192642f0b136d5159a501701" + parent_span_id = "6e8f22c393e68f19" + parent_sampled = 1 + sentry_trace_header = "{}-{}-{}".format(trace_id, parent_span_id, parent_sampled) + + # We simulate here AWS Api Gateway's behavior of passing HTTP headers + # as the `headers` dict in the event passed to the Lambda function. + payload = { + "headers": { + "sentry-trace": sentry_trace_header, + } + } + + lambda_client.invoke( + FunctionName="BasicException", + Payload=json.dumps(payload), + ) + envelopes = test_environment["server"].envelopes + + (error_event, transaction_event) = envelopes + + assert ( + error_event["contexts"]["trace"]["trace_id"] + == transaction_event["contexts"]["trace"]["trace_id"] + == "471a43a4192642f0b136d5159a501701" + ) + + +@pytest.mark.parametrize( + "payload", + [ + {}, + {"headers": None}, + {"headers": ""}, + {"headers": {}}, + {"headers": []}, # EventBridge sends an empty list + ], + ids=[ + "no headers", + "none headers", + "empty string headers", + "empty dict headers", + "empty list headers", + ], +) +def test_headers(lambda_client, test_environment, payload): + lambda_client.invoke( + FunctionName="BasicException", + Payload=json.dumps(payload), + ) + envelopes = test_environment["server"].envelopes + + (error_event, _) = envelopes + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "RuntimeError" + assert error_event["exception"]["values"][0]["value"] == "Oh!" + + +def test_span_origin(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="BasicOk", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + (transaction_event,) = envelopes + + assert ( + transaction_event["contexts"]["trace"]["origin"] == "auto.function.aws_lambda" + ) + + +def test_traces_sampler_has_correct_sampling_context(lambda_client, test_environment): + """ + Test that aws_event and aws_context are passed in the custom_sampling_context + when using the AWS Lambda integration. + """ + test_payload = {"test_key": "test_value"} + response = lambda_client.invoke( + FunctionName="TracesSampler", + Payload=json.dumps(test_payload), + ) + response_payload = json.loads(response["Payload"].read().decode()) + sampling_context_data = json.loads(response_payload["body"])[ + "sampling_context_data" + ] + assert sampling_context_data.get("aws_event_present") is True + assert sampling_context_data.get("aws_context_present") is True + assert sampling_context_data.get("event_data", {}).get("test_key") == "test_value" + + +@pytest.mark.parametrize( + "lambda_function_name", + ["RaiseErrorPerformanceEnabled", "RaiseErrorPerformanceDisabled"], +) +def test_error_has_new_trace_context( + lambda_client, test_environment, lambda_function_name +): + lambda_client.invoke( + FunctionName=lambda_function_name, + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + if lambda_function_name == "RaiseErrorPerformanceEnabled": + (error_event, transaction_event) = envelopes + else: + (error_event,) = envelopes + transaction_event = None + + assert "trace" in error_event["contexts"] + assert "trace_id" in error_event["contexts"]["trace"] + + if transaction_event: + assert "trace" in transaction_event["contexts"] + assert "trace_id" in transaction_event["contexts"]["trace"] + assert ( + error_event["contexts"]["trace"]["trace_id"] + == transaction_event["contexts"]["trace"]["trace_id"] + ) + + +@pytest.mark.parametrize( + "lambda_function_name", + ["RaiseErrorPerformanceEnabled", "RaiseErrorPerformanceDisabled"], +) +def test_error_has_existing_trace_context( + lambda_client, test_environment, lambda_function_name +): + trace_id = "471a43a4192642f0b136d5159a501701" + parent_span_id = "6e8f22c393e68f19" + parent_sampled = 1 + sentry_trace_header = "{}-{}-{}".format(trace_id, parent_span_id, parent_sampled) + + # We simulate here AWS Api Gateway's behavior of passing HTTP headers + # as the `headers` dict in the event passed to the Lambda function. + payload = { + "headers": { + "sentry-trace": sentry_trace_header, + } + } + + lambda_client.invoke( + FunctionName=lambda_function_name, + Payload=json.dumps(payload), + ) + envelopes = test_environment["server"].envelopes + + if lambda_function_name == "RaiseErrorPerformanceEnabled": + (error_event, transaction_event) = envelopes + else: + (error_event,) = envelopes + transaction_event = None + + assert "trace" in error_event["contexts"] + assert "trace_id" in error_event["contexts"]["trace"] + assert ( + error_event["contexts"]["trace"]["trace_id"] + == "471a43a4192642f0b136d5159a501701" + ) + + if transaction_event: + assert "trace" in transaction_event["contexts"] + assert "trace_id" in transaction_event["contexts"]["trace"] + assert ( + transaction_event["contexts"]["trace"]["trace_id"] + == "471a43a4192642f0b136d5159a501701" + ) diff --git a/tests/integrations/aws_lambda/utils.py b/tests/integrations/aws_lambda/utils.py new file mode 100644 index 0000000000..d20c9352e7 --- /dev/null +++ b/tests/integrations/aws_lambda/utils.py @@ -0,0 +1,294 @@ +import gzip +import json +import os +import shutil +import subprocess +import requests +import sys +import time +import threading +import socket +import platform + +from aws_cdk import ( + CfnResource, + Stack, +) +from constructs import Construct +from fastapi import FastAPI, Request +import uvicorn + +from scripts.build_aws_lambda_layer import build_packaged_zip, DIST_PATH + + +LAMBDA_FUNCTION_DIR = "./tests/integrations/aws_lambda/lambda_functions/" +LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR = ( + "./tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/" +) +LAMBDA_FUNCTION_TIMEOUT = 10 +SAM_PORT = 3001 + +PYTHON_VERSION = f"python{sys.version_info.major}.{sys.version_info.minor}" + + +def get_host_ip(): + """ + Returns the IP address of the host we are running on. + """ + if os.environ.get("GITHUB_ACTIONS"): + # Running in GitHub Actions + hostname = socket.gethostname() + host = socket.gethostbyname(hostname) + else: + # Running locally + if platform.system() in ["Darwin", "Windows"]: + # Windows or MacOS + host = "host.docker.internal" + else: + # Linux + hostname = socket.gethostname() + host = socket.gethostbyname(hostname) + + return host + + +def get_project_root(): + """ + Returns the absolute path to the project root directory. + """ + # Start from the current file's directory + current_dir = os.path.dirname(os.path.abspath(__file__)) + + # Navigate up to the project root (4 levels up from tests/integrations/aws_lambda/) + # This is equivalent to the multiple dirname() calls + project_root = os.path.abspath(os.path.join(current_dir, "../../../")) + + return project_root + + +class LocalLambdaStack(Stack): + """ + Uses the AWS CDK to create a local SAM stack containing Lambda functions. + """ + + def __init__(self, scope: Construct, construct_id: str, **kwargs) -> None: + print("[LocalLambdaStack] Creating local SAM Lambda Stack") + super().__init__(scope, construct_id, **kwargs) + + # Override the template synthesis + self.template_options.template_format_version = "2010-09-09" + self.template_options.transforms = ["AWS::Serverless-2016-10-31"] + + print("[LocalLambdaStack] Create Sentry Lambda layer package") + filename = "sentry-sdk-lambda-layer.zip" + build_packaged_zip( + make_dist=True, + out_zip_filename=filename, + ) + + print( + "[LocalLambdaStack] Add Sentry Lambda layer containing the Sentry SDK to the SAM stack" + ) + self.sentry_layer = CfnResource( + self, + "SentryPythonServerlessSDK", + type="AWS::Serverless::LayerVersion", + properties={ + "ContentUri": os.path.join(DIST_PATH, filename), + "CompatibleRuntimes": [ + PYTHON_VERSION, + ], + }, + ) + + dsn = f"http://123@{get_host_ip()}:9999/0" # noqa: E231 + print("[LocalLambdaStack] Using Sentry DSN: %s" % dsn) + + print( + "[LocalLambdaStack] Add all Lambda functions defined in " + "/tests/integrations/aws_lambda/lambda_functions/ to the SAM stack" + ) + lambda_dirs = [ + d + for d in os.listdir(LAMBDA_FUNCTION_DIR) + if os.path.isdir(os.path.join(LAMBDA_FUNCTION_DIR, d)) + ] + for lambda_dir in lambda_dirs: + CfnResource( + self, + lambda_dir, + type="AWS::Serverless::Function", + properties={ + "CodeUri": os.path.join(LAMBDA_FUNCTION_DIR, lambda_dir), + "Handler": "sentry_sdk.integrations.init_serverless_sdk.sentry_lambda_handler", + "Runtime": PYTHON_VERSION, + "Timeout": LAMBDA_FUNCTION_TIMEOUT, + "Layers": [ + {"Ref": self.sentry_layer.logical_id} + ], # Add layer containing the Sentry SDK to function. + "Environment": { + "Variables": { + "SENTRY_DSN": dsn, + "SENTRY_INITIAL_HANDLER": "index.handler", + "SENTRY_TRACES_SAMPLE_RATE": "1.0", + } + }, + }, + ) + print( + "[LocalLambdaStack] - Created Lambda function: %s (%s)" + % ( + lambda_dir, + os.path.join(LAMBDA_FUNCTION_DIR, lambda_dir), + ) + ) + + print( + "[LocalLambdaStack] Add all Lambda functions defined in " + "/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/ to the SAM stack" + ) + lambda_dirs = [ + d + for d in os.listdir(LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR) + if os.path.isdir(os.path.join(LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR, d)) + ] + for lambda_dir in lambda_dirs: + # Copy the Sentry SDK into the function directory + sdk_path = os.path.join( + LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR, lambda_dir, "sentry_sdk" + ) + if not os.path.exists(sdk_path): + # Find the Sentry SDK in the current environment + import sentry_sdk as sdk_module + + sdk_source = os.path.dirname(sdk_module.__file__) + shutil.copytree(sdk_source, sdk_path) + + # Install the requirements of Sentry SDK into the function directory + requirements_file = os.path.join( + get_project_root(), "requirements-aws-lambda-layer.txt" + ) + + # Install the package using pip + subprocess.check_call( + [ + sys.executable, + "-m", + "pip", + "install", + "--upgrade", + "--target", + os.path.join(LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR, lambda_dir), + "-r", + requirements_file, + ] + ) + + CfnResource( + self, + lambda_dir, + type="AWS::Serverless::Function", + properties={ + "CodeUri": os.path.join( + LAMBDA_FUNCTION_WITH_EMBEDDED_SDK_DIR, lambda_dir + ), + "Handler": "index.handler", + "Runtime": PYTHON_VERSION, + "Timeout": LAMBDA_FUNCTION_TIMEOUT, + "Environment": { + "Variables": { + "SENTRY_DSN": dsn, + } + }, + }, + ) + print( + "[LocalLambdaStack] - Created Lambda function: %s (%s)" + % ( + lambda_dir, + os.path.join(LAMBDA_FUNCTION_DIR, lambda_dir), + ) + ) + + @classmethod + def wait_for_stack(cls, timeout=60, port=SAM_PORT): + """ + Wait for SAM to be ready, with timeout. + """ + start_time = time.time() + while True: + if time.time() - start_time > timeout: + raise TimeoutError( + "AWS SAM failed to start within %s seconds. (Maybe Docker is not running?)" + % timeout + ) + + try: + # Try to connect to SAM + response = requests.get(f"http://127.0.0.1:{port}/") # noqa: E231 + if response.status_code == 200 or response.status_code == 404: + return + + except requests.exceptions.ConnectionError: + time.sleep(1) + continue + + +class SentryServerForTesting: + """ + A simple Sentry.io style server that accepts envelopes and stores them in a list. + """ + + def __init__(self, host="0.0.0.0", port=9999, log_level="warning"): + self.envelopes = [] + self.host = host + self.port = port + self.log_level = log_level + self.app = FastAPI() + + @self.app.post("/api/0/envelope/") + async def envelope(request: Request): + print("[SentryServerForTesting] Received envelope") + try: + raw_body = await request.body() + except Exception: + return {"status": "no body received"} + + try: + body = gzip.decompress(raw_body).decode("utf-8") + except Exception: + # If decompression fails, assume it's plain text + body = raw_body.decode("utf-8") + + lines = body.split("\n") + + current_line = 1 # line 0 is envelope header + while current_line < len(lines): + # skip empty lines + if not lines[current_line].strip(): + current_line += 1 + continue + + # skip envelope item header + current_line += 1 + + # add envelope item to store + envelope_item = lines[current_line] + if envelope_item.strip(): + self.envelopes.append(json.loads(envelope_item)) + + return {"status": "ok"} + + def run_server(self): + uvicorn.run(self.app, host=self.host, port=self.port, log_level=self.log_level) + + def start(self): + print( + "[SentryServerForTesting] Starting server on %s:%s" % (self.host, self.port) + ) + server_thread = threading.Thread(target=self.run_server, daemon=True) + server_thread.start() + + def clear_envelopes(self): + print("[SentryServerForTesting] Clearing envelopes") + self.envelopes = [] diff --git a/tox.ini b/tox.ini index f176c70f1a..932ef256ab 100644 --- a/tox.ini +++ b/tox.ini @@ -57,10 +57,7 @@ envlist = {py3.8,py3.11,py3.12}-asyncpg-latest # AWS Lambda - # The aws_lambda tests deploy to the real AWS and have their own - # matrix of Python versions to run the test lambda function in. - # see `lambda_runtime` fixture in tests/integrations/aws_lambda.py - {py3.9}-aws_lambda + {py3.8,py3.9,py3.11,py3.13}-aws_lambda # Beam {py3.7}-beam-v{2.12} @@ -367,7 +364,12 @@ deps = asyncpg: pytest-asyncio # AWS Lambda + aws_lambda: aws-cdk-lib + aws_lambda: aws-sam-cli aws_lambda: boto3 + aws_lambda: fastapi + aws_lambda: requests + aws_lambda: uvicorn # Beam beam-v2.12: apache-beam~=2.12.0 @@ -803,8 +805,6 @@ setenv = socket: TESTPATH=tests/integrations/socket passenv = - SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID - SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY SENTRY_PYTHON_TEST_POSTGRES_HOST SENTRY_PYTHON_TEST_POSTGRES_USER SENTRY_PYTHON_TEST_POSTGRES_PASSWORD From 50b1919a9ddeb19138e9a8dc3510043d5cf00e41 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 12 Mar 2025 15:12:21 +0100 Subject: [PATCH 170/591] Improve asyncio integration error handling. (#4129) Instrumenting asyncio projects can be confusing. Here are two improvements: - If users try to init the Sentry SDK outside of an async loop, a warning message will now printed instructing them how to correctly call init() in async envrionments. Including a link to the docs. - During shutdown of Python unfinished async tasks emit an error `Task was destroyed but it is pending!`. This happens if you use Sentry or not. The error message is confusing and led people to believe the Sentry instrumentation caused this problem. This is now remediated by - The tasks is wrapped by Sentry, but we now **set the name of the wrapped task to include the original** and (and a hint that is has been wrapped by Sentry) to show that the original task is failing, not just some Sentry task unknown to the user. - When shutting down a **info message** is printed, informing that there could be `Task was destroyed but it is pending!` but that those are OK and not a problem with the users code or Sentry. Before this PR the users saw this during shutdown: ``` Exception ignored in: ._sentry_task_factory.._coro_creating_hub_and_span at 0x103ae84f0> Traceback (most recent call last): File "/Users/antonpirker/code/sentry-python/sentry_sdk/integrations/asyncio.py", line 46, in _coro_creating_hub_and_span with sentry_sdk.isolation_scope(): File "/Users/antonpirker/.pyenv/versions/3.12.3/lib/python3.12/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/Users/antonpirker/code/sentry-python/sentry_sdk/scope.py", line 1732, in isolation_scope _current_scope.reset(current_token) ValueError: at 0x103b1cfc0> was created in a different Context Task was destroyed but it is pending! task: ._sentry_task_factory.._coro_creating_hub_and_span() done, defined at /Users/antonpirker/code/sentry-python/sentry_sdk/integrations/asyncio.py:42> wait_for= cb=[gather.._done_callback() at /Users/antonpirker/.pyenv/versions/3.12.3/lib/python3.12/asyncio/tasks.py:767]> ``` With this PR the users will see this during shutdown: Note the INFO message on top and also the task name on the bottom. ``` [sentry] INFO: AsyncIO is shutting down. If you see 'Task was destroyed but it is pending!' errors with '_task_with_sentry_span_creation', these are normal during shutdown and not a problem with your code or Sentry. Exception ignored in: ._sentry_task_factory.._task_with_sentry_span_creation at 0x1028fc4f0> Traceback (most recent call last): File "/Users/antonpirker/code/sentry-python/sentry_sdk/integrations/asyncio.py", line 62, in _task_with_sentry_span_creation with sentry_sdk.isolation_scope(): File "/Users/antonpirker/.pyenv/versions/3.12.3/lib/python3.12/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/Users/antonpirker/code/sentry-python/sentry_sdk/scope.py", line 1732, in isolation_scope _current_scope.reset(current_token) ValueError: at 0x1029710c0> was created in a different Context Task was destroyed but it is pending! task: ._sentry_task_factory.._task_with_sentry_span_creation() done, defined at /Users/antonpirker/code/sentry-python/sentry_sdk/integrations/asyncio.py:58> wait_for= cb=[gather.._done_callback() at /Users/antonpirker/.pyenv/versions/3.12.3/lib/python3.12/asyncio/tasks.py:767]> ``` Fixes #2908 Improves #2333 --- sentry_sdk/integrations/asyncio.py | 69 +++++++++++++++++++++++------- 1 file changed, 53 insertions(+), 16 deletions(-) diff --git a/sentry_sdk/integrations/asyncio.py b/sentry_sdk/integrations/asyncio.py index 7021d7fceb..9326c16e9a 100644 --- a/sentry_sdk/integrations/asyncio.py +++ b/sentry_sdk/integrations/asyncio.py @@ -1,9 +1,10 @@ import sys +import signal import sentry_sdk from sentry_sdk.consts import OP from sentry_sdk.integrations import Integration, DidNotEnable -from sentry_sdk.utils import event_from_exception, reraise +from sentry_sdk.utils import event_from_exception, logger, reraise try: import asyncio @@ -11,7 +12,7 @@ except ImportError: raise DidNotEnable("asyncio not available") -from typing import TYPE_CHECKING +from typing import cast, TYPE_CHECKING if TYPE_CHECKING: from typing import Any @@ -36,10 +37,26 @@ def patch_asyncio(): loop = asyncio.get_running_loop() orig_task_factory = loop.get_task_factory() + # Add a shutdown handler to log a helpful message + def shutdown_handler(): + # type: () -> None + logger.info( + "AsyncIO is shutting down. If you see 'Task was destroyed but it is pending!' " + "errors with '_task_with_sentry_span_creation', these are normal during shutdown " + "and not a problem with your code or Sentry." + ) + + try: + loop.add_signal_handler(signal.SIGINT, shutdown_handler) + loop.add_signal_handler(signal.SIGTERM, shutdown_handler) + except (NotImplementedError, AttributeError): + # Signal handlers might not be supported on all platforms + pass + def _sentry_task_factory(loop, coro, **kwargs): # type: (asyncio.AbstractEventLoop, Coroutine[Any, Any, Any], Any) -> asyncio.Future[Any] - async def _coro_creating_hub_and_span(): + async def _task_with_sentry_span_creation(): # type: () -> Any result = None @@ -56,27 +73,47 @@ async def _coro_creating_hub_and_span(): return result + task = None + # Trying to use user set task factory (if there is one) if orig_task_factory: - return orig_task_factory(loop, _coro_creating_hub_and_span(), **kwargs) - - # The default task factory in `asyncio` does not have its own function - # but is just a couple of lines in `asyncio.base_events.create_task()` - # Those lines are copied here. - - # WARNING: - # If the default behavior of the task creation in asyncio changes, - # this will break! - task = Task(_coro_creating_hub_and_span(), loop=loop, **kwargs) - if task._source_traceback: # type: ignore - del task._source_traceback[-1] # type: ignore + task = orig_task_factory( + loop, _task_with_sentry_span_creation(), **kwargs + ) + + if task is None: + # The default task factory in `asyncio` does not have its own function + # but is just a couple of lines in `asyncio.base_events.create_task()` + # Those lines are copied here. + + # WARNING: + # If the default behavior of the task creation in asyncio changes, + # this will break! + task = Task(_task_with_sentry_span_creation(), loop=loop, **kwargs) + if task._source_traceback: # type: ignore + del task._source_traceback[-1] # type: ignore + + # Set the task name to include the original coroutine's name + try: + cast("asyncio.Task[Any]", task).set_name( + f"{get_name(coro)} (Sentry-wrapped)" + ) + except AttributeError: + # set_name might not be available in all Python versions + pass return task loop.set_task_factory(_sentry_task_factory) # type: ignore + except RuntimeError: # When there is no running loop, we have nothing to patch. - pass + logger.warning( + "There is no running asyncio loop so there is nothing Sentry can patch. " + "Please make sure you call sentry_sdk.init() within a running " + "asyncio loop for the AsyncioIntegration to work. " + "See https://docs.sentry.io/platforms/python/integrations/asyncio/" + ) def _capture_exception(): From e8be8edb56c7d96a35c40177e5286f788daf2af0 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 12 Mar 2025 15:14:56 +0100 Subject: [PATCH 171/591] fix(pyspark): Grab `attemptId` more defensively (#4130) Closes https://github.com/getsentry/sentry-python/issues/1099 --- sentry_sdk/integrations/spark/spark_driver.py | 28 ++++++++- tests/integrations/spark/test_spark.py | 60 +++++++++++++++++++ 2 files changed, 86 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/spark/spark_driver.py b/sentry_sdk/integrations/spark/spark_driver.py index a86f16344d..701ba12d89 100644 --- a/sentry_sdk/integrations/spark/spark_driver.py +++ b/sentry_sdk/integrations/spark/spark_driver.py @@ -260,7 +260,12 @@ def onStageSubmitted(self, stageSubmitted): # noqa: N802,N803 # type: (Any) -> None stage_info = stageSubmitted.stageInfo() message = "Stage {} Submitted".format(stage_info.stageId()) - data = {"attemptId": stage_info.attemptId(), "name": stage_info.name()} + + data = {"name": stage_info.name()} + attempt_id = _get_attempt_id(stage_info) + if attempt_id is not None: + data["attemptId"] = attempt_id + self._add_breadcrumb(level="info", message=message, data=data) _set_app_properties() @@ -271,7 +276,11 @@ def onStageCompleted(self, stageCompleted): # noqa: N802,N803 stage_info = stageCompleted.stageInfo() message = "" level = "" - data = {"attemptId": stage_info.attemptId(), "name": stage_info.name()} + + data = {"name": stage_info.name()} + attempt_id = _get_attempt_id(stage_info) + if attempt_id is not None: + data["attemptId"] = attempt_id # Have to Try Except because stageInfo.failureReason() is typed with Scala Option try: @@ -283,3 +292,18 @@ def onStageCompleted(self, stageCompleted): # noqa: N802,N803 level = "info" self._add_breadcrumb(level=level, message=message, data=data) + + +def _get_attempt_id(stage_info): + # type: (Any) -> Optional[int] + try: + return stage_info.attemptId() + except Exception: + pass + + try: + return stage_info.attemptNumber() + except Exception: + pass + + return None diff --git a/tests/integrations/spark/test_spark.py b/tests/integrations/spark/test_spark.py index 44ba9f8728..7eeab15dc4 100644 --- a/tests/integrations/spark/test_spark.py +++ b/tests/integrations/spark/test_spark.py @@ -14,6 +14,7 @@ from py4j.protocol import Py4JJavaError + ################ # DRIVER TESTS # ################ @@ -166,6 +167,65 @@ def stageInfo(self): # noqa: N802 assert mock_hub.kwargs["data"]["name"] == "run-job" +def test_sentry_listener_on_stage_submitted_no_attempt_id(sentry_listener): + listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: + + class StageInfo: + def stageId(self): # noqa: N802 + return "sample-stage-id-submit" + + def name(self): + return "run-job" + + def attemptNumber(self): # noqa: N802 + return 14 + + class MockStageSubmitted: + def stageInfo(self): # noqa: N802 + stageinf = StageInfo() + return stageinf + + mock_stage_submitted = MockStageSubmitted() + listener.onStageSubmitted(mock_stage_submitted) + + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args + + assert mock_hub.kwargs["level"] == "info" + assert "sample-stage-id-submit" in mock_hub.kwargs["message"] + assert mock_hub.kwargs["data"]["attemptId"] == 14 + assert mock_hub.kwargs["data"]["name"] == "run-job" + + +def test_sentry_listener_on_stage_submitted_no_attempt_id_or_number(sentry_listener): + listener = sentry_listener + with patch.object(listener, "_add_breadcrumb") as mock_add_breadcrumb: + + class StageInfo: + def stageId(self): # noqa: N802 + return "sample-stage-id-submit" + + def name(self): + return "run-job" + + class MockStageSubmitted: + def stageInfo(self): # noqa: N802 + stageinf = StageInfo() + return stageinf + + mock_stage_submitted = MockStageSubmitted() + listener.onStageSubmitted(mock_stage_submitted) + + mock_add_breadcrumb.assert_called_once() + mock_hub = mock_add_breadcrumb.call_args + + assert mock_hub.kwargs["level"] == "info" + assert "sample-stage-id-submit" in mock_hub.kwargs["message"] + assert "attemptId" not in mock_hub.kwargs["data"] + assert mock_hub.kwargs["data"]["name"] == "run-job" + + @pytest.fixture def get_mock_stage_completed(): def _inner(failure_reason): From 42ad8df79815cc6113d4106ce19c32a195a18cfb Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 12 Mar 2025 15:25:44 +0100 Subject: [PATCH 172/591] A way to locally run AWS Lambda functions (#4128) This gives us a way to locally run and test our AWS Lambda integration, without needing a real AWS Lambda account. This should make development of AWS Lambda support better. --------- Co-authored-by: Ivana Kellyer --- scripts/test-lambda-locally/.gitignore | 4 + scripts/test-lambda-locally/README.md | 28 + .../deploy-lambda-locally.sh | 25 + .../test-lambda-locally/lambda_function.py | 25 + scripts/test-lambda-locally/pyproject.toml | 8 + scripts/test-lambda-locally/template.yaml | 29 + scripts/test-lambda-locally/uv.lock | 1239 +++++++++++++++++ 7 files changed, 1358 insertions(+) create mode 100644 scripts/test-lambda-locally/.gitignore create mode 100644 scripts/test-lambda-locally/README.md create mode 100755 scripts/test-lambda-locally/deploy-lambda-locally.sh create mode 100644 scripts/test-lambda-locally/lambda_function.py create mode 100644 scripts/test-lambda-locally/pyproject.toml create mode 100644 scripts/test-lambda-locally/template.yaml create mode 100644 scripts/test-lambda-locally/uv.lock diff --git a/scripts/test-lambda-locally/.gitignore b/scripts/test-lambda-locally/.gitignore new file mode 100644 index 0000000000..f9b7f4de58 --- /dev/null +++ b/scripts/test-lambda-locally/.gitignore @@ -0,0 +1,4 @@ +.envrc +.venv/ +package/ +lambda_deployment_package.zip diff --git a/scripts/test-lambda-locally/README.md b/scripts/test-lambda-locally/README.md new file mode 100644 index 0000000000..115927cc2b --- /dev/null +++ b/scripts/test-lambda-locally/README.md @@ -0,0 +1,28 @@ +# Test AWS Lambda functions locally + +An easy way to run an AWS Lambda function with the Sentry SDK locally. + +This is a small helper to create a AWS Lambda function that includes the +currently checked out Sentry SDK and runs it in a local AWS Lambda environment. + +Currently only embedding the Sentry SDK into the Lambda function package +is supported. Adding the SDK as Lambda Layer is not possible at the moment. + +## Prerequisites + +- Set `SENTRY_DSN` environment variable. The Lambda function will use this DSN. +- You need to have Docker installed and running. + +## Run Lambda function + +- Update `lambda_function.py` to include your test code. +- Run `./deploy-lambda-locally.sh`. This will: + - Install [AWS SAM](https://aws.amazon.com/serverless/sam/) in a virtual Python environment + - Create a lambda function package in `package/` that includes + - The currently checked out Sentry SDK + - All dependencies of the Sentry SDK (certifi and urllib3) + - The actual function defined in `lamdba_function.py`. + - Zip everything together into lambda_deployment_package.zip + - Run a local Lambda environment that serves that Lambda function. +- Point your browser to `http://127.0.0.1:3000` to access your Lambda function. + - Currently GET and POST requests are possible. This is defined in `template.yaml`. \ No newline at end of file diff --git a/scripts/test-lambda-locally/deploy-lambda-locally.sh b/scripts/test-lambda-locally/deploy-lambda-locally.sh new file mode 100755 index 0000000000..495c1259dc --- /dev/null +++ b/scripts/test-lambda-locally/deploy-lambda-locally.sh @@ -0,0 +1,25 @@ +#!/usr/bin/env bash + +# exit on first error +set -xeuo pipefail + +# Setup local AWS Lambda environment + +# Install uv if it's not installed +if ! command -v uv &> /dev/null; then + curl -LsSf https://astral.sh/uv/install.sh | sh +fi + +uv sync + +# Create a deployment package of the lambda function in `lambda_function.py`. +rm -rf package && mkdir -p package +pip install ../../../sentry-python -t package/ --upgrade +cp lambda_function.py package/ +cd package && zip -r ../lambda_deployment_package.zip . && cd .. + +# Start the local Lambda server with the new function (defined in template.yaml) +uv run sam local start-api \ + --skip-pull-image \ + --force-image-build \ + --parameter-overrides SentryDsn=$SENTRY_DSN diff --git a/scripts/test-lambda-locally/lambda_function.py b/scripts/test-lambda-locally/lambda_function.py new file mode 100644 index 0000000000..ceab090499 --- /dev/null +++ b/scripts/test-lambda-locally/lambda_function.py @@ -0,0 +1,25 @@ +import logging +import os +import sentry_sdk + +from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration +from sentry_sdk.integrations.logging import LoggingIntegration + +def lambda_handler(event, context): + sentry_sdk.init( + dsn=os.environ.get("SENTRY_DSN"), + attach_stacktrace=True, + integrations=[ + LoggingIntegration(level=logging.INFO, event_level=logging.ERROR), + AwsLambdaIntegration(timeout_warning=True) + ], + traces_sample_rate=1.0, + debug=True, + ) + + try: + my_dict = {"a" : "test"} + value = my_dict["b"] # This should raise exception + except: + logging.exception("Key Does not Exists") + raise diff --git a/scripts/test-lambda-locally/pyproject.toml b/scripts/test-lambda-locally/pyproject.toml new file mode 100644 index 0000000000..522e9620e8 --- /dev/null +++ b/scripts/test-lambda-locally/pyproject.toml @@ -0,0 +1,8 @@ +[project] +name = "test-lambda-locally" +version = "0" +requires-python = ">=3.12" + +dependencies = [ + "aws-sam-cli>=1.135.0", +] diff --git a/scripts/test-lambda-locally/template.yaml b/scripts/test-lambda-locally/template.yaml new file mode 100644 index 0000000000..67b8f6e7da --- /dev/null +++ b/scripts/test-lambda-locally/template.yaml @@ -0,0 +1,29 @@ +AWSTemplateFormatVersion: '2010-09-09' +Transform: AWS::Serverless-2016-10-31 +Resources: + SentryLambdaFunction: + Type: AWS::Serverless::Function + Properties: + CodeUri: lambda_deployment_package.zip + Handler: lambda_function.lambda_handler + Runtime: python3.12 + Timeout: 30 + Environment: + Variables: + SENTRY_DSN: !Ref SentryDsn + Events: + ApiEventGet: + Type: Api + Properties: + Path: / + Method: get + ApiEventPost: + Type: Api + Properties: + Path: / + Method: post + +Parameters: + SentryDsn: + Type: String + Default: '' diff --git a/scripts/test-lambda-locally/uv.lock b/scripts/test-lambda-locally/uv.lock new file mode 100644 index 0000000000..889ca8e62f --- /dev/null +++ b/scripts/test-lambda-locally/uv.lock @@ -0,0 +1,1239 @@ +version = 1 +revision = 1 +requires-python = ">=3.12" + +[[package]] +name = "annotated-types" +version = "0.7.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/ee/67/531ea369ba64dcff5ec9c3402f9f51bf748cec26dde048a2f973a4eea7f5/annotated_types-0.7.0.tar.gz", hash = "sha256:aff07c09a53a08bc8cfccb9c85b05f1aa9a2a6f23728d790723543408344ce89", size = 16081 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl", hash = "sha256:1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53", size = 13643 }, +] + +[[package]] +name = "arrow" +version = "1.3.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "python-dateutil" }, + { name = "types-python-dateutil" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/2e/00/0f6e8fcdb23ea632c866620cc872729ff43ed91d284c866b515c6342b173/arrow-1.3.0.tar.gz", hash = "sha256:d4540617648cb5f895730f1ad8c82a65f2dad0166f57b75f3ca54759c4d67a85", size = 131960 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f8/ed/e97229a566617f2ae958a6b13e7cc0f585470eac730a73e9e82c32a3cdd2/arrow-1.3.0-py3-none-any.whl", hash = "sha256:c728b120ebc00eb84e01882a6f5e7927a53960aa990ce7dd2b10f39005a67f80", size = 66419 }, +] + +[[package]] +name = "attrs" +version = "25.1.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/49/7c/fdf464bcc51d23881d110abd74b512a42b3d5d376a55a831b44c603ae17f/attrs-25.1.0.tar.gz", hash = "sha256:1c97078a80c814273a76b2a298a932eb681c87415c11dee0a6921de7f1b02c3e", size = 810562 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/fc/30/d4986a882011f9df997a55e6becd864812ccfcd821d64aac8570ee39f719/attrs-25.1.0-py3-none-any.whl", hash = "sha256:c75a69e28a550a7e93789579c22aa26b0f5b83b75dc4e08fe092980051e1090a", size = 63152 }, +] + +[[package]] +name = "aws-lambda-builders" +version = "1.53.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "setuptools" }, + { name = "wheel" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/0b/0a/09a966ac588a3eb3333348a5e13892889fe9531a491359b35bc5b7b13818/aws_lambda_builders-1.53.0.tar.gz", hash = "sha256:d08bfa947fff590f1bedd16c2f4ec7722cbb8869aae80764d99215a41ff284a1", size = 95491 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/28/8c/9cf80784437059db1999655a943eb950a0587793c3fddb56aee3c0f60ae3/aws_lambda_builders-1.53.0-py3-none-any.whl", hash = "sha256:ca9ddd99214aef8a113a3fcd7d7fe3951ef0e078478484f03c398a3bdee04ccb", size = 131138 }, +] + +[[package]] +name = "aws-sam-cli" +version = "1.135.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "aws-lambda-builders" }, + { name = "aws-sam-translator" }, + { name = "boto3" }, + { name = "boto3-stubs", extra = ["apigateway", "cloudformation", "ecr", "iam", "kinesis", "lambda", "s3", "schemas", "secretsmanager", "signer", "sqs", "stepfunctions", "sts", "xray"] }, + { name = "cfn-lint" }, + { name = "chevron" }, + { name = "click" }, + { name = "cookiecutter" }, + { name = "dateparser" }, + { name = "docker" }, + { name = "flask" }, + { name = "jmespath" }, + { name = "jsonschema" }, + { name = "pyopenssl" }, + { name = "pyyaml" }, + { name = "regex" }, + { name = "requests" }, + { name = "rich" }, + { name = "ruamel-yaml" }, + { name = "tomlkit" }, + { name = "typing-extensions" }, + { name = "tzlocal" }, + { name = "watchdog" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/cc/ff/92159d25b8c563de8605cb67b18c6d4ec68880d2dfd7eac689f0f4b80f57/aws_sam_cli-1.135.0.tar.gz", hash = "sha256:c630b351feeb4854ad5ecea6768920c61e7d331b3d040a677fa8744380f48808", size = 5792676 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/8d/0f/f299f9ac27d946d7bf5fb11b3d01e7d1f5affd2ec9220449636949ccc39a/aws_sam_cli-1.135.0-py3-none-any.whl", hash = "sha256:473d30202b89a9624201e46b3ecb9ad5bcd05332c3d308a888464f002c29432b", size = 6077290 }, +] + +[[package]] +name = "aws-sam-translator" +version = "1.95.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "boto3" }, + { name = "jsonschema" }, + { name = "pydantic" }, + { name = "typing-extensions" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/61/8c/4ea1c5fafdec02f2b3a91d60889219a42c18f5c3dd93ec13ef985e4249f6/aws_sam_translator-1.95.0.tar.gz", hash = "sha256:fd2b891fc4cbdde1e06130eaf2710de5cc74442a656b7859b3840691144494cf", size = 327484 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d2/5a/2edbe63d0b1c1e3c685a9b8464626f59c48bfbcc4e20142acae5ddea504c/aws_sam_translator-1.95.0-py3-none-any.whl", hash = "sha256:c9e0f22cbe83c768f7d20a3afb7e654bd6bfc087b387528bd48e98366b82ae40", size = 385846 }, +] + +[[package]] +name = "binaryornot" +version = "0.4.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "chardet" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/a7/fe/7ebfec74d49f97fc55cd38240c7a7d08134002b1e14be8c3897c0dd5e49b/binaryornot-0.4.4.tar.gz", hash = "sha256:359501dfc9d40632edc9fac890e19542db1a287bbcfa58175b66658392018061", size = 371054 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/24/7e/f7b6f453e6481d1e233540262ccbfcf89adcd43606f44a028d7f5fae5eb2/binaryornot-0.4.4-py2.py3-none-any.whl", hash = "sha256:b8b71173c917bddcd2c16070412e369c3ed7f0528926f70cac18a6c97fd563e4", size = 9006 }, +] + +[[package]] +name = "blinker" +version = "1.9.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/21/28/9b3f50ce0e048515135495f198351908d99540d69bfdc8c1d15b73dc55ce/blinker-1.9.0.tar.gz", hash = "sha256:b4ce2265a7abece45e7cc896e98dbebe6cead56bcf805a3d23136d145f5445bf", size = 22460 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl", hash = "sha256:ba0efaa9080b619ff2f3459d1d500c57bddea4a6b424b60a91141db6fd2f08bc", size = 8458 }, +] + +[[package]] +name = "boto3" +version = "1.37.11" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "botocore" }, + { name = "jmespath" }, + { name = "s3transfer" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/21/12/948ab48f2e2d4eda72f907352e67379334ded1a2a6d1ebbaac11e77dfca9/boto3-1.37.11.tar.gz", hash = "sha256:8eec08363ef5db05c2fbf58e89f0c0de6276cda2fdce01e76b3b5f423cd5c0f4", size = 111323 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/29/55/0afe0471e391f4aaa99e5216b5c9ce6493756c0b7a7d8f8ffe85ba83b7a0/boto3-1.37.11-py3-none-any.whl", hash = "sha256:da6c22fc8a7e9bca5d7fc465a877ac3d45b6b086d776bd1a6c55bdde60523741", size = 139553 }, +] + +[[package]] +name = "boto3-stubs" +version = "1.35.71" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "botocore-stubs" }, + { name = "types-s3transfer" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/9f/85/86243ad2792f8506b567c645d97ece548258203c55bcc165fd5801f4372f/boto3_stubs-1.35.71.tar.gz", hash = "sha256:50e20fa74248c96b3e3498b2d81388585583e38b9f0609d2fa58257e49c986a5", size = 93776 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a6/d1/aedf5f4a92e1e74ee29a4d43084780f2d77aeef3d734e550aa2ab304e1fb/boto3_stubs-1.35.71-py3-none-any.whl", hash = "sha256:4abf357250bdb16d1a56489a59bfc385d132a43677956bd984f6578638d599c0", size = 62964 }, +] + +[package.optional-dependencies] +apigateway = [ + { name = "mypy-boto3-apigateway" }, +] +cloudformation = [ + { name = "mypy-boto3-cloudformation" }, +] +ecr = [ + { name = "mypy-boto3-ecr" }, +] +iam = [ + { name = "mypy-boto3-iam" }, +] +kinesis = [ + { name = "mypy-boto3-kinesis" }, +] +lambda = [ + { name = "mypy-boto3-lambda" }, +] +s3 = [ + { name = "mypy-boto3-s3" }, +] +schemas = [ + { name = "mypy-boto3-schemas" }, +] +secretsmanager = [ + { name = "mypy-boto3-secretsmanager" }, +] +signer = [ + { name = "mypy-boto3-signer" }, +] +sqs = [ + { name = "mypy-boto3-sqs" }, +] +stepfunctions = [ + { name = "mypy-boto3-stepfunctions" }, +] +sts = [ + { name = "mypy-boto3-sts" }, +] +xray = [ + { name = "mypy-boto3-xray" }, +] + +[[package]] +name = "botocore" +version = "1.37.11" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "jmespath" }, + { name = "python-dateutil" }, + { name = "urllib3" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/24/ce/b11d4405b8be900bfea15d9460376ff6f07dd0e1b1f8a47e2671bf6e5ca8/botocore-1.37.11.tar.gz", hash = "sha256:72eb3a9a58b064be26ba154e5e56373633b58f951941c340ace0d379590d98b5", size = 13640593 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/63/0d/b07e9b6cd8823e520f1782742730f2e68b68ad7444825ed8dd8fcdb98fcb/botocore-1.37.11-py3-none-any.whl", hash = "sha256:02505309b1235f9f15a6da79103ca224b3f3dc5f6a62f8630fbb2c6ed05e2da8", size = 13407367 }, +] + +[[package]] +name = "botocore-stubs" +version = "1.37.11" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "types-awscrt" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/3d/6f/710664aac77cf91a663dcb291c2bbdcfe796909115aa5bb03382521359b1/botocore_stubs-1.37.11.tar.gz", hash = "sha256:9b89ba9a98eb9f088a5f82c52488013858092777c17b56265574bbf2d21da422", size = 42119 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/45/89/c8a6497055f9ecd0af5c16434c277635a4b365793d54f2d8f2b28aeeb58e/botocore_stubs-1.37.11-py3-none-any.whl", hash = "sha256:bec458a0d054892cdf82466b4d075f30a36fa03ce34f9becbcace5f36ec674bf", size = 65384 }, +] + +[[package]] +name = "certifi" +version = "2025.1.31" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/1c/ab/c9f1e32b7b1bf505bf26f0ef697775960db7932abeb7b516de930ba2705f/certifi-2025.1.31.tar.gz", hash = "sha256:3d5da6925056f6f18f119200434a4780a94263f10d1c21d032a6f6b2baa20651", size = 167577 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/38/fc/bce832fd4fd99766c04d1ee0eead6b0ec6486fb100ae5e74c1d91292b982/certifi-2025.1.31-py3-none-any.whl", hash = "sha256:ca78db4565a652026a4db2bcdf68f2fb589ea80d0be70e03929ed730746b84fe", size = 166393 }, +] + +[[package]] +name = "cffi" +version = "1.17.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "pycparser" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/fc/97/c783634659c2920c3fc70419e3af40972dbaf758daa229a7d6ea6135c90d/cffi-1.17.1.tar.gz", hash = "sha256:1c39c6016c32bc48dd54561950ebd6836e1670f2ae46128f67cf49e789c52824", size = 516621 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/5a/84/e94227139ee5fb4d600a7a4927f322e1d4aea6fdc50bd3fca8493caba23f/cffi-1.17.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:805b4371bf7197c329fcb3ead37e710d1bca9da5d583f5073b799d5c5bd1eee4", size = 183178 }, + { url = "https://files.pythonhosted.org/packages/da/ee/fb72c2b48656111c4ef27f0f91da355e130a923473bf5ee75c5643d00cca/cffi-1.17.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:733e99bc2df47476e3848417c5a4540522f234dfd4ef3ab7fafdf555b082ec0c", size = 178840 }, + { url = "https://files.pythonhosted.org/packages/cc/b6/db007700f67d151abadf508cbfd6a1884f57eab90b1bb985c4c8c02b0f28/cffi-1.17.1-cp312-cp312-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1257bdabf294dceb59f5e70c64a3e2f462c30c7ad68092d01bbbfb1c16b1ba36", size = 454803 }, + { url = "https://files.pythonhosted.org/packages/1a/df/f8d151540d8c200eb1c6fba8cd0dfd40904f1b0682ea705c36e6c2e97ab3/cffi-1.17.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:da95af8214998d77a98cc14e3a3bd00aa191526343078b530ceb0bd710fb48a5", size = 478850 }, + { url = "https://files.pythonhosted.org/packages/28/c0/b31116332a547fd2677ae5b78a2ef662dfc8023d67f41b2a83f7c2aa78b1/cffi-1.17.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d63afe322132c194cf832bfec0dc69a99fb9bb6bbd550f161a49e9e855cc78ff", size = 485729 }, + { url = "https://files.pythonhosted.org/packages/91/2b/9a1ddfa5c7f13cab007a2c9cc295b70fbbda7cb10a286aa6810338e60ea1/cffi-1.17.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f79fc4fc25f1c8698ff97788206bb3c2598949bfe0fef03d299eb1b5356ada99", size = 471256 }, + { url = "https://files.pythonhosted.org/packages/b2/d5/da47df7004cb17e4955df6a43d14b3b4ae77737dff8bf7f8f333196717bf/cffi-1.17.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b62ce867176a75d03a665bad002af8e6d54644fad99a3c70905c543130e39d93", size = 479424 }, + { url = "https://files.pythonhosted.org/packages/0b/ac/2a28bcf513e93a219c8a4e8e125534f4f6db03e3179ba1c45e949b76212c/cffi-1.17.1-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:386c8bf53c502fff58903061338ce4f4950cbdcb23e2902d86c0f722b786bbe3", size = 484568 }, + { url = "https://files.pythonhosted.org/packages/d4/38/ca8a4f639065f14ae0f1d9751e70447a261f1a30fa7547a828ae08142465/cffi-1.17.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:4ceb10419a9adf4460ea14cfd6bc43d08701f0835e979bf821052f1805850fe8", size = 488736 }, + { url = "https://files.pythonhosted.org/packages/86/c5/28b2d6f799ec0bdecf44dced2ec5ed43e0eb63097b0f58c293583b406582/cffi-1.17.1-cp312-cp312-win32.whl", hash = "sha256:a08d7e755f8ed21095a310a693525137cfe756ce62d066e53f502a83dc550f65", size = 172448 }, + { url = "https://files.pythonhosted.org/packages/50/b9/db34c4755a7bd1cb2d1603ac3863f22bcecbd1ba29e5ee841a4bc510b294/cffi-1.17.1-cp312-cp312-win_amd64.whl", hash = "sha256:51392eae71afec0d0c8fb1a53b204dbb3bcabcb3c9b807eedf3e1e6ccf2de903", size = 181976 }, + { url = "https://files.pythonhosted.org/packages/8d/f8/dd6c246b148639254dad4d6803eb6a54e8c85c6e11ec9df2cffa87571dbe/cffi-1.17.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:f3a2b4222ce6b60e2e8b337bb9596923045681d71e5a082783484d845390938e", size = 182989 }, + { url = "https://files.pythonhosted.org/packages/8b/f1/672d303ddf17c24fc83afd712316fda78dc6fce1cd53011b839483e1ecc8/cffi-1.17.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:0984a4925a435b1da406122d4d7968dd861c1385afe3b45ba82b750f229811e2", size = 178802 }, + { url = "https://files.pythonhosted.org/packages/0e/2d/eab2e858a91fdff70533cab61dcff4a1f55ec60425832ddfdc9cd36bc8af/cffi-1.17.1-cp313-cp313-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d01b12eeeb4427d3110de311e1774046ad344f5b1a7403101878976ecd7a10f3", size = 454792 }, + { url = "https://files.pythonhosted.org/packages/75/b2/fbaec7c4455c604e29388d55599b99ebcc250a60050610fadde58932b7ee/cffi-1.17.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:706510fe141c86a69c8ddc029c7910003a17353970cff3b904ff0686a5927683", size = 478893 }, + { url = "https://files.pythonhosted.org/packages/4f/b7/6e4a2162178bf1935c336d4da8a9352cccab4d3a5d7914065490f08c0690/cffi-1.17.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:de55b766c7aa2e2a3092c51e0483d700341182f08e67c63630d5b6f200bb28e5", size = 485810 }, + { url = "https://files.pythonhosted.org/packages/c7/8a/1d0e4a9c26e54746dc08c2c6c037889124d4f59dffd853a659fa545f1b40/cffi-1.17.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c59d6e989d07460165cc5ad3c61f9fd8f1b4796eacbd81cee78957842b834af4", size = 471200 }, + { url = "https://files.pythonhosted.org/packages/26/9f/1aab65a6c0db35f43c4d1b4f580e8df53914310afc10ae0397d29d697af4/cffi-1.17.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd398dbc6773384a17fe0d3e7eeb8d1a21c2200473ee6806bb5e6a8e62bb73dd", size = 479447 }, + { url = "https://files.pythonhosted.org/packages/5f/e4/fb8b3dd8dc0e98edf1135ff067ae070bb32ef9d509d6cb0f538cd6f7483f/cffi-1.17.1-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:3edc8d958eb099c634dace3c7e16560ae474aa3803a5df240542b305d14e14ed", size = 484358 }, + { url = "https://files.pythonhosted.org/packages/f1/47/d7145bf2dc04684935d57d67dff9d6d795b2ba2796806bb109864be3a151/cffi-1.17.1-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:72e72408cad3d5419375fc87d289076ee319835bdfa2caad331e377589aebba9", size = 488469 }, + { url = "https://files.pythonhosted.org/packages/bf/ee/f94057fa6426481d663b88637a9a10e859e492c73d0384514a17d78ee205/cffi-1.17.1-cp313-cp313-win32.whl", hash = "sha256:e03eab0a8677fa80d646b5ddece1cbeaf556c313dcfac435ba11f107ba117b5d", size = 172475 }, + { url = "https://files.pythonhosted.org/packages/7c/fc/6a8cb64e5f0324877d503c854da15d76c1e50eb722e320b15345c4d0c6de/cffi-1.17.1-cp313-cp313-win_amd64.whl", hash = "sha256:f6a16c31041f09ead72d69f583767292f750d24913dadacf5756b966aacb3f1a", size = 182009 }, +] + +[[package]] +name = "cfn-lint" +version = "1.25.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "aws-sam-translator" }, + { name = "jsonpatch" }, + { name = "networkx" }, + { name = "pyyaml" }, + { name = "regex" }, + { name = "sympy" }, + { name = "typing-extensions" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/4d/c0/a36a1bdc6ba1fd4a7e5f48cd23a1802ccaf745ffb5c79e3fdf800eb5ae90/cfn_lint-1.25.1.tar.gz", hash = "sha256:717012566c6034ffa7e60fcf1b350804d093ee37589a1e91a1fd867f33a930b7", size = 2837233 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/8b/1c/b03940f2213f308f19318aaa8847adfe789b834e497f8839b2c9a876618b/cfn_lint-1.25.1-py3-none-any.whl", hash = "sha256:bbf6c2d95689da466dc427217ab7ed8f3a2a4a134df70876cc63e41aaad9385a", size = 4907033 }, +] + +[[package]] +name = "chardet" +version = "5.2.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f3/0d/f7b6ab21ec75897ed80c17d79b15951a719226b9fababf1e40ea74d69079/chardet-5.2.0.tar.gz", hash = "sha256:1b3b6ff479a8c414bc3fa2c0852995695c4a026dcd6d0633b2dd092ca39c1cf7", size = 2069618 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/38/6f/f5fbc992a329ee4e0f288c1fe0e2ad9485ed064cac731ed2fe47dcc38cbf/chardet-5.2.0-py3-none-any.whl", hash = "sha256:e1cf59446890a00105fe7b7912492ea04b6e6f06d4b742b2c788469e34c82970", size = 199385 }, +] + +[[package]] +name = "charset-normalizer" +version = "3.4.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/16/b0/572805e227f01586461c80e0fd25d65a2115599cc9dad142fee4b747c357/charset_normalizer-3.4.1.tar.gz", hash = "sha256:44251f18cd68a75b56585dd00dae26183e102cd5e0f9f1466e6df5da2ed64ea3", size = 123188 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0a/9a/dd1e1cdceb841925b7798369a09279bd1cf183cef0f9ddf15a3a6502ee45/charset_normalizer-3.4.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:73d94b58ec7fecbc7366247d3b0b10a21681004153238750bb67bd9012414545", size = 196105 }, + { url = "https://files.pythonhosted.org/packages/d3/8c/90bfabf8c4809ecb648f39794cf2a84ff2e7d2a6cf159fe68d9a26160467/charset_normalizer-3.4.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dad3e487649f498dd991eeb901125411559b22e8d7ab25d3aeb1af367df5efd7", size = 140404 }, + { url = "https://files.pythonhosted.org/packages/ad/8f/e410d57c721945ea3b4f1a04b74f70ce8fa800d393d72899f0a40526401f/charset_normalizer-3.4.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c30197aa96e8eed02200a83fba2657b4c3acd0f0aa4bdc9f6c1af8e8962e0757", size = 150423 }, + { url = "https://files.pythonhosted.org/packages/f0/b8/e6825e25deb691ff98cf5c9072ee0605dc2acfca98af70c2d1b1bc75190d/charset_normalizer-3.4.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2369eea1ee4a7610a860d88f268eb39b95cb588acd7235e02fd5a5601773d4fa", size = 143184 }, + { url = "https://files.pythonhosted.org/packages/3e/a2/513f6cbe752421f16d969e32f3583762bfd583848b763913ddab8d9bfd4f/charset_normalizer-3.4.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bc2722592d8998c870fa4e290c2eec2c1569b87fe58618e67d38b4665dfa680d", size = 145268 }, + { url = "https://files.pythonhosted.org/packages/74/94/8a5277664f27c3c438546f3eb53b33f5b19568eb7424736bdc440a88a31f/charset_normalizer-3.4.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ffc9202a29ab3920fa812879e95a9e78b2465fd10be7fcbd042899695d75e616", size = 147601 }, + { url = "https://files.pythonhosted.org/packages/7c/5f/6d352c51ee763623a98e31194823518e09bfa48be2a7e8383cf691bbb3d0/charset_normalizer-3.4.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:804a4d582ba6e5b747c625bf1255e6b1507465494a40a2130978bda7b932c90b", size = 141098 }, + { url = "https://files.pythonhosted.org/packages/78/d4/f5704cb629ba5ab16d1d3d741396aec6dc3ca2b67757c45b0599bb010478/charset_normalizer-3.4.1-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:0f55e69f030f7163dffe9fd0752b32f070566451afe180f99dbeeb81f511ad8d", size = 149520 }, + { url = "https://files.pythonhosted.org/packages/c5/96/64120b1d02b81785f222b976c0fb79a35875457fa9bb40827678e54d1bc8/charset_normalizer-3.4.1-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:c4c3e6da02df6fa1410a7680bd3f63d4f710232d3139089536310d027950696a", size = 152852 }, + { url = "https://files.pythonhosted.org/packages/84/c9/98e3732278a99f47d487fd3468bc60b882920cef29d1fa6ca460a1fdf4e6/charset_normalizer-3.4.1-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:5df196eb874dae23dcfb968c83d4f8fdccb333330fe1fc278ac5ceeb101003a9", size = 150488 }, + { url = "https://files.pythonhosted.org/packages/13/0e/9c8d4cb99c98c1007cc11eda969ebfe837bbbd0acdb4736d228ccaabcd22/charset_normalizer-3.4.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:e358e64305fe12299a08e08978f51fc21fac060dcfcddd95453eabe5b93ed0e1", size = 146192 }, + { url = "https://files.pythonhosted.org/packages/b2/21/2b6b5b860781a0b49427309cb8670785aa543fb2178de875b87b9cc97746/charset_normalizer-3.4.1-cp312-cp312-win32.whl", hash = "sha256:9b23ca7ef998bc739bf6ffc077c2116917eabcc901f88da1b9856b210ef63f35", size = 95550 }, + { url = "https://files.pythonhosted.org/packages/21/5b/1b390b03b1d16c7e382b561c5329f83cc06623916aab983e8ab9239c7d5c/charset_normalizer-3.4.1-cp312-cp312-win_amd64.whl", hash = "sha256:6ff8a4a60c227ad87030d76e99cd1698345d4491638dfa6673027c48b3cd395f", size = 102785 }, + { url = "https://files.pythonhosted.org/packages/38/94/ce8e6f63d18049672c76d07d119304e1e2d7c6098f0841b51c666e9f44a0/charset_normalizer-3.4.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:aabfa34badd18f1da5ec1bc2715cadc8dca465868a4e73a0173466b688f29dda", size = 195698 }, + { url = "https://files.pythonhosted.org/packages/24/2e/dfdd9770664aae179a96561cc6952ff08f9a8cd09a908f259a9dfa063568/charset_normalizer-3.4.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:22e14b5d70560b8dd51ec22863f370d1e595ac3d024cb8ad7d308b4cd95f8313", size = 140162 }, + { url = "https://files.pythonhosted.org/packages/24/4e/f646b9093cff8fc86f2d60af2de4dc17c759de9d554f130b140ea4738ca6/charset_normalizer-3.4.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8436c508b408b82d87dc5f62496973a1805cd46727c34440b0d29d8a2f50a6c9", size = 150263 }, + { url = "https://files.pythonhosted.org/packages/5e/67/2937f8d548c3ef6e2f9aab0f6e21001056f692d43282b165e7c56023e6dd/charset_normalizer-3.4.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2d074908e1aecee37a7635990b2c6d504cd4766c7bc9fc86d63f9c09af3fa11b", size = 142966 }, + { url = "https://files.pythonhosted.org/packages/52/ed/b7f4f07de100bdb95c1756d3a4d17b90c1a3c53715c1a476f8738058e0fa/charset_normalizer-3.4.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:955f8851919303c92343d2f66165294848d57e9bba6cf6e3625485a70a038d11", size = 144992 }, + { url = "https://files.pythonhosted.org/packages/96/2c/d49710a6dbcd3776265f4c923bb73ebe83933dfbaa841c5da850fe0fd20b/charset_normalizer-3.4.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:44ecbf16649486d4aebafeaa7ec4c9fed8b88101f4dd612dcaf65d5e815f837f", size = 147162 }, + { url = "https://files.pythonhosted.org/packages/b4/41/35ff1f9a6bd380303dea55e44c4933b4cc3c4850988927d4082ada230273/charset_normalizer-3.4.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:0924e81d3d5e70f8126529951dac65c1010cdf117bb75eb02dd12339b57749dd", size = 140972 }, + { url = "https://files.pythonhosted.org/packages/fb/43/c6a0b685fe6910d08ba971f62cd9c3e862a85770395ba5d9cad4fede33ab/charset_normalizer-3.4.1-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:2967f74ad52c3b98de4c3b32e1a44e32975e008a9cd2a8cc8966d6a5218c5cb2", size = 149095 }, + { url = "https://files.pythonhosted.org/packages/4c/ff/a9a504662452e2d2878512115638966e75633519ec11f25fca3d2049a94a/charset_normalizer-3.4.1-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:c75cb2a3e389853835e84a2d8fb2b81a10645b503eca9bcb98df6b5a43eb8886", size = 152668 }, + { url = "https://files.pythonhosted.org/packages/6c/71/189996b6d9a4b932564701628af5cee6716733e9165af1d5e1b285c530ed/charset_normalizer-3.4.1-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:09b26ae6b1abf0d27570633b2b078a2a20419c99d66fb2823173d73f188ce601", size = 150073 }, + { url = "https://files.pythonhosted.org/packages/e4/93/946a86ce20790e11312c87c75ba68d5f6ad2208cfb52b2d6a2c32840d922/charset_normalizer-3.4.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:fa88b843d6e211393a37219e6a1c1df99d35e8fd90446f1118f4216e307e48cd", size = 145732 }, + { url = "https://files.pythonhosted.org/packages/cd/e5/131d2fb1b0dddafc37be4f3a2fa79aa4c037368be9423061dccadfd90091/charset_normalizer-3.4.1-cp313-cp313-win32.whl", hash = "sha256:eb8178fe3dba6450a3e024e95ac49ed3400e506fd4e9e5c32d30adda88cbd407", size = 95391 }, + { url = "https://files.pythonhosted.org/packages/27/f2/4f9a69cc7712b9b5ad8fdb87039fd89abba997ad5cbe690d1835d40405b0/charset_normalizer-3.4.1-cp313-cp313-win_amd64.whl", hash = "sha256:b1ac5992a838106edb89654e0aebfc24f5848ae2547d22c2c3f66454daa11971", size = 102702 }, + { url = "https://files.pythonhosted.org/packages/0e/f6/65ecc6878a89bb1c23a086ea335ad4bf21a588990c3f535a227b9eea9108/charset_normalizer-3.4.1-py3-none-any.whl", hash = "sha256:d98b1668f06378c6dbefec3b92299716b931cd4e6061f3c875a71ced1780ab85", size = 49767 }, +] + +[[package]] +name = "chevron" +version = "0.14.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/15/1f/ca74b65b19798895d63a6e92874162f44233467c9e7c1ed8afd19016ebe9/chevron-0.14.0.tar.gz", hash = "sha256:87613aafdf6d77b6a90ff073165a61ae5086e21ad49057aa0e53681601800ebf", size = 11440 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/52/93/342cc62a70ab727e093ed98e02a725d85b746345f05d2b5e5034649f4ec8/chevron-0.14.0-py3-none-any.whl", hash = "sha256:fbf996a709f8da2e745ef763f482ce2d311aa817d287593a5b990d6d6e4f0443", size = 11595 }, +] + +[[package]] +name = "click" +version = "8.1.8" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "colorama", marker = "sys_platform == 'win32'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/b9/2e/0090cbf739cee7d23781ad4b89a9894a41538e4fcf4c31dcdd705b78eb8b/click-8.1.8.tar.gz", hash = "sha256:ed53c9d8990d83c2a27deae68e4ee337473f6330c040a31d4225c9574d16096a", size = 226593 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/7e/d4/7ebdbd03970677812aac39c869717059dbb71a4cfc033ca6e5221787892c/click-8.1.8-py3-none-any.whl", hash = "sha256:63c132bbbed01578a06712a2d1f497bb62d9c1c0d329b7903a866228027263b2", size = 98188 }, +] + +[[package]] +name = "colorama" +version = "0.4.6" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/d8/53/6f443c9a4a8358a93a6792e2acffb9d9d5cb0a5cfd8802644b7b1c9a02e4/colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44", size = 27697 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d1/d6/3965ed04c63042e047cb6a3e6ed1a63a35087b6a609aa3a15ed8ac56c221/colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6", size = 25335 }, +] + +[[package]] +name = "cookiecutter" +version = "2.6.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "arrow" }, + { name = "binaryornot" }, + { name = "click" }, + { name = "jinja2" }, + { name = "python-slugify" }, + { name = "pyyaml" }, + { name = "requests" }, + { name = "rich" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/52/17/9f2cd228eb949a91915acd38d3eecdc9d8893dde353b603f0db7e9f6be55/cookiecutter-2.6.0.tar.gz", hash = "sha256:db21f8169ea4f4fdc2408d48ca44859349de2647fbe494a9d6c3edfc0542c21c", size = 158767 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b6/d9/0137658a353168ffa9d0fc14b812d3834772040858ddd1cb6eeaf09f7a44/cookiecutter-2.6.0-py3-none-any.whl", hash = "sha256:a54a8e37995e4ed963b3e82831072d1ad4b005af736bb17b99c2cbd9d41b6e2d", size = 39177 }, +] + +[[package]] +name = "cryptography" +version = "44.0.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "cffi", marker = "platform_python_implementation != 'PyPy'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/cd/25/4ce80c78963834b8a9fd1cc1266be5ed8d1840785c0f2e1b73b8d128d505/cryptography-44.0.2.tar.gz", hash = "sha256:c63454aa261a0cf0c5b4718349629793e9e634993538db841165b3df74f37ec0", size = 710807 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/92/ef/83e632cfa801b221570c5f58c0369db6fa6cef7d9ff859feab1aae1a8a0f/cryptography-44.0.2-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:efcfe97d1b3c79e486554efddeb8f6f53a4cdd4cf6086642784fa31fc384e1d7", size = 6676361 }, + { url = "https://files.pythonhosted.org/packages/30/ec/7ea7c1e4c8fc8329506b46c6c4a52e2f20318425d48e0fe597977c71dbce/cryptography-44.0.2-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:29ecec49f3ba3f3849362854b7253a9f59799e3763b0c9d0826259a88efa02f1", size = 3952350 }, + { url = "https://files.pythonhosted.org/packages/27/61/72e3afdb3c5ac510330feba4fc1faa0fe62e070592d6ad00c40bb69165e5/cryptography-44.0.2-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bc821e161ae88bfe8088d11bb39caf2916562e0a2dc7b6d56714a48b784ef0bb", size = 4166572 }, + { url = "https://files.pythonhosted.org/packages/26/e4/ba680f0b35ed4a07d87f9e98f3ebccb05091f3bf6b5a478b943253b3bbd5/cryptography-44.0.2-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:3c00b6b757b32ce0f62c574b78b939afab9eecaf597c4d624caca4f9e71e7843", size = 3958124 }, + { url = "https://files.pythonhosted.org/packages/9c/e8/44ae3e68c8b6d1cbc59040288056df2ad7f7f03bbcaca6b503c737ab8e73/cryptography-44.0.2-cp37-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:7bdcd82189759aba3816d1f729ce42ffded1ac304c151d0a8e89b9996ab863d5", size = 3678122 }, + { url = "https://files.pythonhosted.org/packages/27/7b/664ea5e0d1eab511a10e480baf1c5d3e681c7d91718f60e149cec09edf01/cryptography-44.0.2-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:4973da6ca3db4405c54cd0b26d328be54c7747e89e284fcff166132eb7bccc9c", size = 4191831 }, + { url = "https://files.pythonhosted.org/packages/2a/07/79554a9c40eb11345e1861f46f845fa71c9e25bf66d132e123d9feb8e7f9/cryptography-44.0.2-cp37-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:4e389622b6927d8133f314949a9812972711a111d577a5d1f4bee5e58736b80a", size = 3960583 }, + { url = "https://files.pythonhosted.org/packages/bb/6d/858e356a49a4f0b591bd6789d821427de18432212e137290b6d8a817e9bf/cryptography-44.0.2-cp37-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:f514ef4cd14bb6fb484b4a60203e912cfcb64f2ab139e88c2274511514bf7308", size = 4191753 }, + { url = "https://files.pythonhosted.org/packages/b2/80/62df41ba4916067fa6b125aa8c14d7e9181773f0d5d0bd4dcef580d8b7c6/cryptography-44.0.2-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:1bc312dfb7a6e5d66082c87c34c8a62176e684b6fe3d90fcfe1568de675e6688", size = 4079550 }, + { url = "https://files.pythonhosted.org/packages/f3/cd/2558cc08f7b1bb40683f99ff4327f8dcfc7de3affc669e9065e14824511b/cryptography-44.0.2-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:3b721b8b4d948b218c88cb8c45a01793483821e709afe5f622861fc6182b20a7", size = 4298367 }, + { url = "https://files.pythonhosted.org/packages/71/59/94ccc74788945bc3bd4cf355d19867e8057ff5fdbcac781b1ff95b700fb1/cryptography-44.0.2-cp37-abi3-win32.whl", hash = "sha256:51e4de3af4ec3899d6d178a8c005226491c27c4ba84101bfb59c901e10ca9f79", size = 2772843 }, + { url = "https://files.pythonhosted.org/packages/ca/2c/0d0bbaf61ba05acb32f0841853cfa33ebb7a9ab3d9ed8bb004bd39f2da6a/cryptography-44.0.2-cp37-abi3-win_amd64.whl", hash = "sha256:c505d61b6176aaf982c5717ce04e87da5abc9a36a5b39ac03905c4aafe8de7aa", size = 3209057 }, + { url = "https://files.pythonhosted.org/packages/9e/be/7a26142e6d0f7683d8a382dd963745e65db895a79a280a30525ec92be890/cryptography-44.0.2-cp39-abi3-macosx_10_9_universal2.whl", hash = "sha256:8e0ddd63e6bf1161800592c71ac794d3fb8001f2caebe0966e77c5234fa9efc3", size = 6677789 }, + { url = "https://files.pythonhosted.org/packages/06/88/638865be7198a84a7713950b1db7343391c6066a20e614f8fa286eb178ed/cryptography-44.0.2-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81276f0ea79a208d961c433a947029e1a15948966658cf6710bbabb60fcc2639", size = 3951919 }, + { url = "https://files.pythonhosted.org/packages/d7/fc/99fe639bcdf58561dfad1faa8a7369d1dc13f20acd78371bb97a01613585/cryptography-44.0.2-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a1e657c0f4ea2a23304ee3f964db058c9e9e635cc7019c4aa21c330755ef6fd", size = 4167812 }, + { url = "https://files.pythonhosted.org/packages/53/7b/aafe60210ec93d5d7f552592a28192e51d3c6b6be449e7fd0a91399b5d07/cryptography-44.0.2-cp39-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:6210c05941994290f3f7f175a4a57dbbb2afd9273657614c506d5976db061181", size = 3958571 }, + { url = "https://files.pythonhosted.org/packages/16/32/051f7ce79ad5a6ef5e26a92b37f172ee2d6e1cce09931646eef8de1e9827/cryptography-44.0.2-cp39-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:d1c3572526997b36f245a96a2b1713bf79ce99b271bbcf084beb6b9b075f29ea", size = 3679832 }, + { url = "https://files.pythonhosted.org/packages/78/2b/999b2a1e1ba2206f2d3bca267d68f350beb2b048a41ea827e08ce7260098/cryptography-44.0.2-cp39-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:b042d2a275c8cee83a4b7ae30c45a15e6a4baa65a179a0ec2d78ebb90e4f6699", size = 4193719 }, + { url = "https://files.pythonhosted.org/packages/72/97/430e56e39a1356e8e8f10f723211a0e256e11895ef1a135f30d7d40f2540/cryptography-44.0.2-cp39-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:d03806036b4f89e3b13b6218fefea8d5312e450935b1a2d55f0524e2ed7c59d9", size = 3960852 }, + { url = "https://files.pythonhosted.org/packages/89/33/c1cf182c152e1d262cac56850939530c05ca6c8d149aa0dcee490b417e99/cryptography-44.0.2-cp39-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:c7362add18b416b69d58c910caa217f980c5ef39b23a38a0880dfd87bdf8cd23", size = 4193906 }, + { url = "https://files.pythonhosted.org/packages/e1/99/87cf26d4f125380dc674233971069bc28d19b07f7755b29861570e513650/cryptography-44.0.2-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:8cadc6e3b5a1f144a039ea08a0bdb03a2a92e19c46be3285123d32029f40a922", size = 4081572 }, + { url = "https://files.pythonhosted.org/packages/b3/9f/6a3e0391957cc0c5f84aef9fbdd763035f2b52e998a53f99345e3ac69312/cryptography-44.0.2-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:6f101b1f780f7fc613d040ca4bdf835c6ef3b00e9bd7125a4255ec574c7916e4", size = 4298631 }, + { url = "https://files.pythonhosted.org/packages/e2/a5/5bc097adb4b6d22a24dea53c51f37e480aaec3465285c253098642696423/cryptography-44.0.2-cp39-abi3-win32.whl", hash = "sha256:3dc62975e31617badc19a906481deacdeb80b4bb454394b4098e3f2525a488c5", size = 2773792 }, + { url = "https://files.pythonhosted.org/packages/33/cf/1f7649b8b9a3543e042d3f348e398a061923ac05b507f3f4d95f11938aa9/cryptography-44.0.2-cp39-abi3-win_amd64.whl", hash = "sha256:5f6f90b72d8ccadb9c6e311c775c8305381db88374c65fa1a68250aa8a9cb3a6", size = 3210957 }, +] + +[[package]] +name = "dateparser" +version = "1.2.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "python-dateutil" }, + { name = "pytz" }, + { name = "regex" }, + { name = "tzlocal" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/bd/3f/d3207a05f5b6a78c66d86631e60bfba5af163738a599a5b9aa2c2737a09e/dateparser-1.2.1.tar.gz", hash = "sha256:7e4919aeb48481dbfc01ac9683c8e20bfe95bb715a38c1e9f6af889f4f30ccc3", size = 309924 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/cf/0a/981c438c4cd84147c781e4e96c1d72df03775deb1bc76c5a6ee8afa89c62/dateparser-1.2.1-py3-none-any.whl", hash = "sha256:bdcac262a467e6260030040748ad7c10d6bacd4f3b9cdb4cfd2251939174508c", size = 295658 }, +] + +[[package]] +name = "docker" +version = "7.1.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "pywin32", marker = "sys_platform == 'win32'" }, + { name = "requests" }, + { name = "urllib3" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/91/9b/4a2ea29aeba62471211598dac5d96825bb49348fa07e906ea930394a83ce/docker-7.1.0.tar.gz", hash = "sha256:ad8c70e6e3f8926cb8a92619b832b4ea5299e2831c14284663184e200546fa6c", size = 117834 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/e3/26/57c6fb270950d476074c087527a558ccb6f4436657314bfb6cdf484114c4/docker-7.1.0-py3-none-any.whl", hash = "sha256:c96b93b7f0a746f9e77d325bcfb87422a3d8bd4f03136ae8a85b37f1898d5fc0", size = 147774 }, +] + +[[package]] +name = "flask" +version = "3.1.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "blinker" }, + { name = "click" }, + { name = "itsdangerous" }, + { name = "jinja2" }, + { name = "werkzeug" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/89/50/dff6380f1c7f84135484e176e0cac8690af72fa90e932ad2a0a60e28c69b/flask-3.1.0.tar.gz", hash = "sha256:5f873c5184c897c8d9d1b05df1e3d01b14910ce69607a117bd3277098a5836ac", size = 680824 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/af/47/93213ee66ef8fae3b93b3e29206f6b251e65c97bd91d8e1c5596ef15af0a/flask-3.1.0-py3-none-any.whl", hash = "sha256:d667207822eb83f1c4b50949b1623c8fc8d51f2341d65f72e1a1815397551136", size = 102979 }, +] + +[[package]] +name = "idna" +version = "3.10" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f1/70/7703c29685631f5a7590aa73f1f1d3fa9a380e654b86af429e0934a32f7d/idna-3.10.tar.gz", hash = "sha256:12f65c9b470abda6dc35cf8e63cc574b1c52b11df2c86030af0ac09b01b13ea9", size = 190490 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/76/c6/c88e154df9c4e1a2a66ccf0005a88dfb2650c1dffb6f5ce603dfbd452ce3/idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3", size = 70442 }, +] + +[[package]] +name = "itsdangerous" +version = "2.2.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/9c/cb/8ac0172223afbccb63986cc25049b154ecfb5e85932587206f42317be31d/itsdangerous-2.2.0.tar.gz", hash = "sha256:e0050c0b7da1eea53ffaf149c0cfbb5c6e2e2b69c4bef22c81fa6eb73e5f6173", size = 54410 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl", hash = "sha256:c6242fc49e35958c8b15141343aa660db5fc54d4f13a1db01a3f5891b98700ef", size = 16234 }, +] + +[[package]] +name = "jinja2" +version = "3.1.6" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "markupsafe" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/df/bf/f7da0350254c0ed7c72f3e33cef02e048281fec7ecec5f032d4aac52226b/jinja2-3.1.6.tar.gz", hash = "sha256:0137fb05990d35f1275a587e9aee6d56da821fc83491a0fb838183be43f66d6d", size = 245115 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl", hash = "sha256:85ece4451f492d0c13c5dd7c13a64681a86afae63a5f347908daf103ce6d2f67", size = 134899 }, +] + +[[package]] +name = "jmespath" +version = "1.0.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/00/2a/e867e8531cf3e36b41201936b7fa7ba7b5702dbef42922193f05c8976cd6/jmespath-1.0.1.tar.gz", hash = "sha256:90261b206d6defd58fdd5e85f478bf633a2901798906be2ad389150c5c60edbe", size = 25843 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl", hash = "sha256:02e2e4cc71b5bcab88332eebf907519190dd9e6e82107fa7f83b1003a6252980", size = 20256 }, +] + +[[package]] +name = "jsonpatch" +version = "1.33" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "jsonpointer" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/42/78/18813351fe5d63acad16aec57f94ec2b70a09e53ca98145589e185423873/jsonpatch-1.33.tar.gz", hash = "sha256:9fcd4009c41e6d12348b4a0ff2563ba56a2923a7dfee731d004e212e1ee5030c", size = 21699 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl", hash = "sha256:0ae28c0cd062bbd8b8ecc26d7d164fbbea9652a1a3693f3b956c1eae5145dade", size = 12898 }, +] + +[[package]] +name = "jsonpointer" +version = "3.0.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/6a/0a/eebeb1fa92507ea94016a2a790b93c2ae41a7e18778f85471dc54475ed25/jsonpointer-3.0.0.tar.gz", hash = "sha256:2b2d729f2091522d61c3b31f82e11870f60b68f43fbc705cb76bf4b832af59ef", size = 9114 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl", hash = "sha256:13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942", size = 7595 }, +] + +[[package]] +name = "jsonschema" +version = "4.23.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "attrs" }, + { name = "jsonschema-specifications" }, + { name = "referencing" }, + { name = "rpds-py" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/38/2e/03362ee4034a4c917f697890ccd4aec0800ccf9ded7f511971c75451deec/jsonschema-4.23.0.tar.gz", hash = "sha256:d71497fef26351a33265337fa77ffeb82423f3ea21283cd9467bb03999266bc4", size = 325778 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/69/4a/4f9dbeb84e8850557c02365a0eee0649abe5eb1d84af92a25731c6c0f922/jsonschema-4.23.0-py3-none-any.whl", hash = "sha256:fbadb6f8b144a8f8cf9f0b89ba94501d143e50411a1278633f56a7acf7fd5566", size = 88462 }, +] + +[[package]] +name = "jsonschema-specifications" +version = "2024.10.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "referencing" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/10/db/58f950c996c793472e336ff3655b13fbcf1e3b359dcf52dcf3ed3b52c352/jsonschema_specifications-2024.10.1.tar.gz", hash = "sha256:0f38b83639958ce1152d02a7f062902c41c8fd20d558b0c34344292d417ae272", size = 15561 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d1/0f/8910b19ac0670a0f80ce1008e5e751c4a57e14d2c4c13a482aa6079fa9d6/jsonschema_specifications-2024.10.1-py3-none-any.whl", hash = "sha256:a09a0680616357d9a0ecf05c12ad234479f549239d0f5b55f3deea67475da9bf", size = 18459 }, +] + +[[package]] +name = "markdown-it-py" +version = "3.0.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "mdurl" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/38/71/3b932df36c1a044d397a1f92d1cf91ee0a503d91e470cbd670aa66b07ed0/markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb", size = 74596 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/42/d7/1ec15b46af6af88f19b8e5ffea08fa375d433c998b8a7639e76935c14f1f/markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1", size = 87528 }, +] + +[[package]] +name = "markupsafe" +version = "3.0.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/b2/97/5d42485e71dfc078108a86d6de8fa46db44a1a9295e89c5d6d4a06e23a62/markupsafe-3.0.2.tar.gz", hash = "sha256:ee55d3edf80167e48ea11a923c7386f4669df67d7994554387f84e7d8b0a2bf0", size = 20537 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/22/09/d1f21434c97fc42f09d290cbb6350d44eb12f09cc62c9476effdb33a18aa/MarkupSafe-3.0.2-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:9778bd8ab0a994ebf6f84c2b949e65736d5575320a17ae8984a77fab08db94cf", size = 14274 }, + { url = "https://files.pythonhosted.org/packages/6b/b0/18f76bba336fa5aecf79d45dcd6c806c280ec44538b3c13671d49099fdd0/MarkupSafe-3.0.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:846ade7b71e3536c4e56b386c2a47adf5741d2d8b94ec9dc3e92e5e1ee1e2225", size = 12348 }, + { url = "https://files.pythonhosted.org/packages/e0/25/dd5c0f6ac1311e9b40f4af06c78efde0f3b5cbf02502f8ef9501294c425b/MarkupSafe-3.0.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1c99d261bd2d5f6b59325c92c73df481e05e57f19837bdca8413b9eac4bd8028", size = 24149 }, + { url = "https://files.pythonhosted.org/packages/f3/f0/89e7aadfb3749d0f52234a0c8c7867877876e0a20b60e2188e9850794c17/MarkupSafe-3.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e17c96c14e19278594aa4841ec148115f9c7615a47382ecb6b82bd8fea3ab0c8", size = 23118 }, + { url = "https://files.pythonhosted.org/packages/d5/da/f2eeb64c723f5e3777bc081da884b414671982008c47dcc1873d81f625b6/MarkupSafe-3.0.2-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:88416bd1e65dcea10bc7569faacb2c20ce071dd1f87539ca2ab364bf6231393c", size = 22993 }, + { url = "https://files.pythonhosted.org/packages/da/0e/1f32af846df486dce7c227fe0f2398dc7e2e51d4a370508281f3c1c5cddc/MarkupSafe-3.0.2-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:2181e67807fc2fa785d0592dc2d6206c019b9502410671cc905d132a92866557", size = 24178 }, + { url = "https://files.pythonhosted.org/packages/c4/f6/bb3ca0532de8086cbff5f06d137064c8410d10779c4c127e0e47d17c0b71/MarkupSafe-3.0.2-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:52305740fe773d09cffb16f8ed0427942901f00adedac82ec8b67752f58a1b22", size = 23319 }, + { url = "https://files.pythonhosted.org/packages/a2/82/8be4c96ffee03c5b4a034e60a31294daf481e12c7c43ab8e34a1453ee48b/MarkupSafe-3.0.2-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ad10d3ded218f1039f11a75f8091880239651b52e9bb592ca27de44eed242a48", size = 23352 }, + { url = "https://files.pythonhosted.org/packages/51/ae/97827349d3fcffee7e184bdf7f41cd6b88d9919c80f0263ba7acd1bbcb18/MarkupSafe-3.0.2-cp312-cp312-win32.whl", hash = "sha256:0f4ca02bea9a23221c0182836703cbf8930c5e9454bacce27e767509fa286a30", size = 15097 }, + { url = "https://files.pythonhosted.org/packages/c1/80/a61f99dc3a936413c3ee4e1eecac96c0da5ed07ad56fd975f1a9da5bc630/MarkupSafe-3.0.2-cp312-cp312-win_amd64.whl", hash = "sha256:8e06879fc22a25ca47312fbe7c8264eb0b662f6db27cb2d3bbbc74b1df4b9b87", size = 15601 }, + { url = "https://files.pythonhosted.org/packages/83/0e/67eb10a7ecc77a0c2bbe2b0235765b98d164d81600746914bebada795e97/MarkupSafe-3.0.2-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:ba9527cdd4c926ed0760bc301f6728ef34d841f405abf9d4f959c478421e4efd", size = 14274 }, + { url = "https://files.pythonhosted.org/packages/2b/6d/9409f3684d3335375d04e5f05744dfe7e9f120062c9857df4ab490a1031a/MarkupSafe-3.0.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:f8b3d067f2e40fe93e1ccdd6b2e1d16c43140e76f02fb1319a05cf2b79d99430", size = 12352 }, + { url = "https://files.pythonhosted.org/packages/d2/f5/6eadfcd3885ea85fe2a7c128315cc1bb7241e1987443d78c8fe712d03091/MarkupSafe-3.0.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:569511d3b58c8791ab4c2e1285575265991e6d8f8700c7be0e88f86cb0672094", size = 24122 }, + { url = "https://files.pythonhosted.org/packages/0c/91/96cf928db8236f1bfab6ce15ad070dfdd02ed88261c2afafd4b43575e9e9/MarkupSafe-3.0.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15ab75ef81add55874e7ab7055e9c397312385bd9ced94920f2802310c930396", size = 23085 }, + { url = "https://files.pythonhosted.org/packages/c2/cf/c9d56af24d56ea04daae7ac0940232d31d5a8354f2b457c6d856b2057d69/MarkupSafe-3.0.2-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f3818cb119498c0678015754eba762e0d61e5b52d34c8b13d770f0719f7b1d79", size = 22978 }, + { url = "https://files.pythonhosted.org/packages/2a/9f/8619835cd6a711d6272d62abb78c033bda638fdc54c4e7f4272cf1c0962b/MarkupSafe-3.0.2-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:cdb82a876c47801bb54a690c5ae105a46b392ac6099881cdfb9f6e95e4014c6a", size = 24208 }, + { url = "https://files.pythonhosted.org/packages/f9/bf/176950a1792b2cd2102b8ffeb5133e1ed984547b75db47c25a67d3359f77/MarkupSafe-3.0.2-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:cabc348d87e913db6ab4aa100f01b08f481097838bdddf7c7a84b7575b7309ca", size = 23357 }, + { url = "https://files.pythonhosted.org/packages/ce/4f/9a02c1d335caabe5c4efb90e1b6e8ee944aa245c1aaaab8e8a618987d816/MarkupSafe-3.0.2-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:444dcda765c8a838eaae23112db52f1efaf750daddb2d9ca300bcae1039adc5c", size = 23344 }, + { url = "https://files.pythonhosted.org/packages/ee/55/c271b57db36f748f0e04a759ace9f8f759ccf22b4960c270c78a394f58be/MarkupSafe-3.0.2-cp313-cp313-win32.whl", hash = "sha256:bcf3e58998965654fdaff38e58584d8937aa3096ab5354d493c77d1fdd66d7a1", size = 15101 }, + { url = "https://files.pythonhosted.org/packages/29/88/07df22d2dd4df40aba9f3e402e6dc1b8ee86297dddbad4872bd5e7b0094f/MarkupSafe-3.0.2-cp313-cp313-win_amd64.whl", hash = "sha256:e6a2a455bd412959b57a172ce6328d2dd1f01cb2135efda2e4576e8a23fa3b0f", size = 15603 }, + { url = "https://files.pythonhosted.org/packages/62/6a/8b89d24db2d32d433dffcd6a8779159da109842434f1dd2f6e71f32f738c/MarkupSafe-3.0.2-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:b5a6b3ada725cea8a5e634536b1b01c30bcdcd7f9c6fff4151548d5bf6b3a36c", size = 14510 }, + { url = "https://files.pythonhosted.org/packages/7a/06/a10f955f70a2e5a9bf78d11a161029d278eeacbd35ef806c3fd17b13060d/MarkupSafe-3.0.2-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:a904af0a6162c73e3edcb969eeeb53a63ceeb5d8cf642fade7d39e7963a22ddb", size = 12486 }, + { url = "https://files.pythonhosted.org/packages/34/cf/65d4a571869a1a9078198ca28f39fba5fbb910f952f9dbc5220afff9f5e6/MarkupSafe-3.0.2-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4aa4e5faecf353ed117801a068ebab7b7e09ffb6e1d5e412dc852e0da018126c", size = 25480 }, + { url = "https://files.pythonhosted.org/packages/0c/e3/90e9651924c430b885468b56b3d597cabf6d72be4b24a0acd1fa0e12af67/MarkupSafe-3.0.2-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c0ef13eaeee5b615fb07c9a7dadb38eac06a0608b41570d8ade51c56539e509d", size = 23914 }, + { url = "https://files.pythonhosted.org/packages/66/8c/6c7cf61f95d63bb866db39085150df1f2a5bd3335298f14a66b48e92659c/MarkupSafe-3.0.2-cp313-cp313t-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d16a81a06776313e817c951135cf7340a3e91e8c1ff2fac444cfd75fffa04afe", size = 23796 }, + { url = "https://files.pythonhosted.org/packages/bb/35/cbe9238ec3f47ac9a7c8b3df7a808e7cb50fe149dc7039f5f454b3fba218/MarkupSafe-3.0.2-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:6381026f158fdb7c72a168278597a5e3a5222e83ea18f543112b2662a9b699c5", size = 25473 }, + { url = "https://files.pythonhosted.org/packages/e6/32/7621a4382488aa283cc05e8984a9c219abad3bca087be9ec77e89939ded9/MarkupSafe-3.0.2-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:3d79d162e7be8f996986c064d1c7c817f6df3a77fe3d6859f6f9e7be4b8c213a", size = 24114 }, + { url = "https://files.pythonhosted.org/packages/0d/80/0985960e4b89922cb5a0bac0ed39c5b96cbc1a536a99f30e8c220a996ed9/MarkupSafe-3.0.2-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:131a3c7689c85f5ad20f9f6fb1b866f402c445b220c19fe4308c0b147ccd2ad9", size = 24098 }, + { url = "https://files.pythonhosted.org/packages/82/78/fedb03c7d5380df2427038ec8d973587e90561b2d90cd472ce9254cf348b/MarkupSafe-3.0.2-cp313-cp313t-win32.whl", hash = "sha256:ba8062ed2cf21c07a9e295d5b8a2a5ce678b913b45fdf68c32d95d6c1291e0b6", size = 15208 }, + { url = "https://files.pythonhosted.org/packages/4f/65/6079a46068dfceaeabb5dcad6d674f5f5c61a6fa5673746f42a9f4c233b3/MarkupSafe-3.0.2-cp313-cp313t-win_amd64.whl", hash = "sha256:e444a31f8db13eb18ada366ab3cf45fd4b31e4db1236a4448f68778c1d1a5a2f", size = 15739 }, +] + +[[package]] +name = "mdurl" +version = "0.1.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/d6/54/cfe61301667036ec958cb99bd3efefba235e65cdeb9c84d24a8293ba1d90/mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba", size = 8729 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8", size = 9979 }, +] + +[[package]] +name = "mpmath" +version = "1.3.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/e0/47/dd32fa426cc72114383ac549964eecb20ecfd886d1e5ccf5340b55b02f57/mpmath-1.3.0.tar.gz", hash = "sha256:7a28eb2a9774d00c7bc92411c19a89209d5da7c4c9a9e227be8330a23a25b91f", size = 508106 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/43/e3/7d92a15f894aa0c9c4b49b8ee9ac9850d6e63b03c9c32c0367a13ae62209/mpmath-1.3.0-py3-none-any.whl", hash = "sha256:a0b2b9fe80bbcd81a6647ff13108738cfb482d481d826cc0e02f5b35e5c88d2c", size = 536198 }, +] + +[[package]] +name = "mypy-boto3-apigateway" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/e0/3d/c5dc7a750d9fdba2bf704d3d963be9ad4ed617fe5bb98e5c88374a3d8d69/mypy_boto3_apigateway-1.35.93.tar.gz", hash = "sha256:df90957c5f2c219663f825b905cb53b9f53fd7982e01bb21da65f5757c3d5d41", size = 44837 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f9/7d/89f26a626ab30283143222430bd39ec46cf8a2ae002e5b5c590e01ff3ad0/mypy_boto3_apigateway-1.35.93-py3-none-any.whl", hash = "sha256:a5649e9899209470c35249651f7f2faa7d6919aab6b4fcac7bd4a54c11e872bc", size = 50874 }, +] + +[[package]] +name = "mypy-boto3-cloudformation" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f3/26/e59425e30fb1783aa718f1a8ac93cdc415e279e175c953ee0a72310f7490/mypy_boto3_cloudformation-1.35.93.tar.gz", hash = "sha256:57dc112ff3e2ddc1e9e621e428490b904c0da8c1532d30e9fa2a19aefde9f719", size = 54529 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/17/52/6e73adba190fc65c5cf89ed9394cc8a1acb073989f4eda87f80f451c9b15/mypy_boto3_cloudformation-1.35.93-py3-none-any.whl", hash = "sha256:4111913cb2c9fd9099ecd616212923312fde0c126ee41f5821759ae9df4272b9", size = 66124 }, +] + +[[package]] +name = "mypy-boto3-ecr" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/92/ae/1598bf3dc7069f0e48a60a482dffa71885e1558aa076243375820de2792f/mypy_boto3_ecr-1.35.93.tar.gz", hash = "sha256:57295a72a9473b8542578ab15eb0a4909cad6f2cee1da41ce6a8a40ab7051438", size = 33904 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/83/3b/4130e22423812da282bd9ebbf08a0f14ed2e314409847bc336b841c8177b/mypy_boto3_ecr-1.35.93-py3-none-any.whl", hash = "sha256:49d98ac7376e919c0061da44aeae9577b63343eee2c1d537fd636d8886db9ad2", size = 39733 }, +] + +[[package]] +name = "mypy-boto3-iam" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/c7/24/7cb0b26c3af8207496880155441cfd7f5d8c5404d4669e39385eb307672d/mypy_boto3_iam-1.35.93.tar.gz", hash = "sha256:2595c8dac406e4e771d3b7d7835faacb936d20449b9cdd17a53f076219cc7712", size = 85815 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/dc/5a/2694c8c692fad6908c3a52f629eb87b04c242dc8bb0091e56ff3780cdb45/mypy_boto3_iam-1.35.93-py3-none-any.whl", hash = "sha256:e2955040062bf9cb587a1874e1b2f2cca33cbf167187fd3a56b6c5412cc13dc9", size = 91125 }, +] + +[[package]] +name = "mypy-boto3-kinesis" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/7d/c3/eb9f1aeaf42ea55c473b0281fe5813aafe3283733ad84fbd27c370416753/mypy_boto3_kinesis-1.35.93.tar.gz", hash = "sha256:f0718f5b54b955761790b4b33bdcab8d0c779bd50cc671c6862a8e0554515bda", size = 22476 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/56/bd/e44b999f516116dcb034262a1ed04d8ed3b830e84970b1224823ce866031/mypy_boto3_kinesis-1.35.93-py3-none-any.whl", hash = "sha256:fb11df380319e3cf5c26f43536107593836e36c6b9f3b415a7016aeaed2af1de", size = 32164 }, +] + +[[package]] +name = "mypy-boto3-lambda" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f0/ef/b90e51be87b5c226005c765a7109a26b5ce39cf349f2603336bd5c365863/mypy_boto3_lambda-1.35.93.tar.gz", hash = "sha256:c11b047743c7635ea8385abffaf97788a108b71479612e9b5e7d0bb19029d7a4", size = 41120 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/6c/f0/3c03cc63c157046106f59768e915c21377a372be6bc9f079601dd646cf4d/mypy_boto3_lambda-1.35.93-py3-none-any.whl", hash = "sha256:6bcd623c827724cde0b21b30c328515811b178763b75f0701a641cc7aa3aa414", size = 47708 }, +] + +[[package]] +name = "mypy-boto3-s3" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/15/53/99667aad21b236612ecb50eee09fdc4de6fbe39c3a75a6bad387d108ed1f/mypy_boto3_s3-1.35.93.tar.gz", hash = "sha256:b4529e57a8d5f21d4c61fe650fa6764fee2ba7ab524a455a34ba2698ef6d27a8", size = 72871 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/e0/52/9d45db5690eb2b3160c43259d70dd6890d9bc24633848bcb8ef835d44d6c/mypy_boto3_s3-1.35.93-py3-none-any.whl", hash = "sha256:4cd3f1718fa0d8a54212c495cdff493bdcc6a8ae419d95428c60fb6bc7db7980", size = 79501 }, +] + +[[package]] +name = "mypy-boto3-schemas" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/c9/f7/63c5b0db122b99265a14f179f41ab01566610c78abe14e63a4df3ebca7fa/mypy_boto3_schemas-1.35.93.tar.gz", hash = "sha256:7f2255ddd6d531101ec67fbd1afca8be02568f4e5787d1631199aa25b58a480f", size = 20680 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b2/37/cf848ce4ec07bbd7d64c91efe8d31f5aa86bf5d6d2a9f7123ca3ce3fed44/mypy_boto3_schemas-1.35.93-py3-none-any.whl", hash = "sha256:9e82b7d6e059a531359cc0304b5d4c979406d06e9d19482c7a22ccb61b40c7ff", size = 28746 }, +] + +[[package]] +name = "mypy-boto3-secretsmanager" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/d8/c6/1c69c3ac9fadeb6cc01da5a90edd5f36cbf09a4fa66e8cef638917eba4d1/mypy_boto3_secretsmanager-1.35.93.tar.gz", hash = "sha256:b6c4bc88a5fe4143124272728d41342e01c778b406db9d647a20dad0de7d6f47", size = 19624 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b6/ff/758f8869d10b10bf6bec7908bd9d532fdd26b6f04c2af4de3751d2c92b93/mypy_boto3_secretsmanager-1.35.93-py3-none-any.whl", hash = "sha256:521075d42b6d05f0d7302d1837520e9111a84d6613152d32dc8cbb3cd6fceeec", size = 26581 }, +] + +[[package]] +name = "mypy-boto3-signer" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/d1/00/954104765b3414b0221cf18efebcee656f7b8be603866682a0dcf9e00ecf/mypy_boto3_signer-1.35.93.tar.gz", hash = "sha256:f12c7c7025cc25804146431f639f3eb9db664a4695bf28d2a87f58111fc7f888", size = 20496 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/51/a0/142a49f1bd98b9a393896e0912cc8dd7a1ac91c2fff224f2c4efb166e180/mypy_boto3_signer-1.35.93-py3-none-any.whl", hash = "sha256:e1ac026096be6a52b6de45771226efbd3909a1861a638441572d926650d7fd8c", size = 28770 }, +] + +[[package]] +name = "mypy-boto3-sqs" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/29/5b/040ba82c53d5edf578ad0aafcac501b91a259b40f296ef6662db975b6595/mypy_boto3_sqs-1.35.93.tar.gz", hash = "sha256:8ea7f63e0878544705c31996ae4c064095fbb4f780f8323a84f7a75281d643fe", size = 23344 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/82/eb/d8c10da3f905921f70f008f3bca092711e316ced49287e42f45309860aca/mypy_boto3_sqs-1.35.93-py3-none-any.whl", hash = "sha256:341974f77e66851b9a4190d0014481e6baabae82d32f9ee559faa823b693609b", size = 33491 }, +] + +[[package]] +name = "mypy-boto3-stepfunctions" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/ec/f9/44a59a6c84edfd94477e5427befcbecdb4f92ae34d897536671dc4994e23/mypy_boto3_stepfunctions-1.35.93.tar.gz", hash = "sha256:20230615c42e7aabbd43b62657ca3534e96767245705d12d42672ac87cd1b59c", size = 30894 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/da/39/0964782eff12ec9c22a5dd78bc19f755df313fb6aa1215293444899dc40e/mypy_boto3_stepfunctions-1.35.93-py3-none-any.whl", hash = "sha256:7994450153298b87382119680d7fae4d8b5a6e6250cef364148ad8d0b84bd237", size = 35602 }, +] + +[[package]] +name = "mypy-boto3-sts" +version = "1.35.97" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/9f/fc/652992367bad0bae7d1c8d8bd5fa455570de77337f8d0c2021263dc4e695/mypy_boto3_sts-1.35.97.tar.gz", hash = "sha256:6df698f6a400a82ebcc2f10adb43557f66278467200e0f75588e7de3e4a1622d", size = 16487 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/8d/7c/092999366962bbe0bab5af8e18e0c8f70943ca34a42c214e3862df2fa80b/mypy_boto3_sts-1.35.97-py3-none-any.whl", hash = "sha256:50c32613aa9e8d33e5df922392e32daed6fcd0e4d4cc8d43f5948c69be1c9e1e", size = 19991 }, +] + +[[package]] +name = "mypy-boto3-xray" +version = "1.35.93" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/b6/98/1ffe456cf073fe6ee1826f053943793d4082fe02412a109c72c0f414a66c/mypy_boto3_xray-1.35.93.tar.gz", hash = "sha256:7e0af9474f06da1923aa37c8639b051042cc3a56d1a36b0141124d9de7be6709", size = 31639 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/8f/b4/826f269d883bd76df41b44fba4a49b2cd9b2a2a34a5561bc251bdb6778f2/mypy_boto3_xray-1.35.93-py3-none-any.whl", hash = "sha256:e80c2be40c5cb4851dc08c145101b4e52a6f471dab0fc5f488975f6e14f7cb93", size = 36455 }, +] + +[[package]] +name = "networkx" +version = "3.4.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/fd/1d/06475e1cd5264c0b870ea2cc6fdb3e37177c1e565c43f56ff17a10e3937f/networkx-3.4.2.tar.gz", hash = "sha256:307c3669428c5362aab27c8a1260aa8f47c4e91d3891f48be0141738d8d053e1", size = 2151368 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b9/54/dd730b32ea14ea797530a4479b2ed46a6fb250f682a9cfb997e968bf0261/networkx-3.4.2-py3-none-any.whl", hash = "sha256:df5d4365b724cf81b8c6a7312509d0c22386097011ad1abe274afd5e9d3bbc5f", size = 1723263 }, +] + +[[package]] +name = "pycparser" +version = "2.22" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/1d/b2/31537cf4b1ca988837256c910a668b553fceb8f069bedc4b1c826024b52c/pycparser-2.22.tar.gz", hash = "sha256:491c8be9c040f5390f5bf44a5b07752bd07f56edf992381b05c701439eec10f6", size = 172736 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/13/a3/a812df4e2dd5696d1f351d58b8fe16a405b234ad2886a0dab9183fb78109/pycparser-2.22-py3-none-any.whl", hash = "sha256:c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc", size = 117552 }, +] + +[[package]] +name = "pydantic" +version = "2.10.6" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "annotated-types" }, + { name = "pydantic-core" }, + { name = "typing-extensions" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/b7/ae/d5220c5c52b158b1de7ca89fc5edb72f304a70a4c540c84c8844bf4008de/pydantic-2.10.6.tar.gz", hash = "sha256:ca5daa827cce33de7a42be142548b0096bf05a7e7b365aebfa5f8eeec7128236", size = 761681 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f4/3c/8cc1cc84deffa6e25d2d0c688ebb80635dfdbf1dbea3e30c541c8cf4d860/pydantic-2.10.6-py3-none-any.whl", hash = "sha256:427d664bf0b8a2b34ff5dd0f5a18df00591adcee7198fbd71981054cef37b584", size = 431696 }, +] + +[[package]] +name = "pydantic-core" +version = "2.27.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "typing-extensions" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/fc/01/f3e5ac5e7c25833db5eb555f7b7ab24cd6f8c322d3a3ad2d67a952dc0abc/pydantic_core-2.27.2.tar.gz", hash = "sha256:eb026e5a4c1fee05726072337ff51d1efb6f59090b7da90d30ea58625b1ffb39", size = 413443 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d6/74/51c8a5482ca447871c93e142d9d4a92ead74de6c8dc5e66733e22c9bba89/pydantic_core-2.27.2-cp312-cp312-macosx_10_12_x86_64.whl", hash = "sha256:9e0c8cfefa0ef83b4da9588448b6d8d2a2bf1a53c3f1ae5fca39eb3061e2f0b0", size = 1893127 }, + { url = "https://files.pythonhosted.org/packages/d3/f3/c97e80721735868313c58b89d2de85fa80fe8dfeeed84dc51598b92a135e/pydantic_core-2.27.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:83097677b8e3bd7eaa6775720ec8e0405f1575015a463285a92bfdfe254529ef", size = 1811340 }, + { url = "https://files.pythonhosted.org/packages/9e/91/840ec1375e686dbae1bd80a9e46c26a1e0083e1186abc610efa3d9a36180/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:172fce187655fece0c90d90a678424b013f8fbb0ca8b036ac266749c09438cb7", size = 1822900 }, + { url = "https://files.pythonhosted.org/packages/f6/31/4240bc96025035500c18adc149aa6ffdf1a0062a4b525c932065ceb4d868/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:519f29f5213271eeeeb3093f662ba2fd512b91c5f188f3bb7b27bc5973816934", size = 1869177 }, + { url = "https://files.pythonhosted.org/packages/fa/20/02fbaadb7808be578317015c462655c317a77a7c8f0ef274bc016a784c54/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:05e3a55d124407fffba0dd6b0c0cd056d10e983ceb4e5dbd10dda135c31071d6", size = 2038046 }, + { url = "https://files.pythonhosted.org/packages/06/86/7f306b904e6c9eccf0668248b3f272090e49c275bc488a7b88b0823444a4/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9c3ed807c7b91de05e63930188f19e921d1fe90de6b4f5cd43ee7fcc3525cb8c", size = 2685386 }, + { url = "https://files.pythonhosted.org/packages/8d/f0/49129b27c43396581a635d8710dae54a791b17dfc50c70164866bbf865e3/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6fb4aadc0b9a0c063206846d603b92030eb6f03069151a625667f982887153e2", size = 1997060 }, + { url = "https://files.pythonhosted.org/packages/0d/0f/943b4af7cd416c477fd40b187036c4f89b416a33d3cc0ab7b82708a667aa/pydantic_core-2.27.2-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:28ccb213807e037460326424ceb8b5245acb88f32f3d2777427476e1b32c48c4", size = 2004870 }, + { url = "https://files.pythonhosted.org/packages/35/40/aea70b5b1a63911c53a4c8117c0a828d6790483f858041f47bab0b779f44/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:de3cd1899e2c279b140adde9357c4495ed9d47131b4a4eaff9052f23398076b3", size = 1999822 }, + { url = "https://files.pythonhosted.org/packages/f2/b3/807b94fd337d58effc5498fd1a7a4d9d59af4133e83e32ae39a96fddec9d/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_armv7l.whl", hash = "sha256:220f892729375e2d736b97d0e51466252ad84c51857d4d15f5e9692f9ef12be4", size = 2130364 }, + { url = "https://files.pythonhosted.org/packages/fc/df/791c827cd4ee6efd59248dca9369fb35e80a9484462c33c6649a8d02b565/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:a0fcd29cd6b4e74fe8ddd2c90330fd8edf2e30cb52acda47f06dd615ae72da57", size = 2158303 }, + { url = "https://files.pythonhosted.org/packages/9b/67/4e197c300976af185b7cef4c02203e175fb127e414125916bf1128b639a9/pydantic_core-2.27.2-cp312-cp312-win32.whl", hash = "sha256:1e2cb691ed9834cd6a8be61228471d0a503731abfb42f82458ff27be7b2186fc", size = 1834064 }, + { url = "https://files.pythonhosted.org/packages/1f/ea/cd7209a889163b8dcca139fe32b9687dd05249161a3edda62860430457a5/pydantic_core-2.27.2-cp312-cp312-win_amd64.whl", hash = "sha256:cc3f1a99a4f4f9dd1de4fe0312c114e740b5ddead65bb4102884b384c15d8bc9", size = 1989046 }, + { url = "https://files.pythonhosted.org/packages/bc/49/c54baab2f4658c26ac633d798dab66b4c3a9bbf47cff5284e9c182f4137a/pydantic_core-2.27.2-cp312-cp312-win_arm64.whl", hash = "sha256:3911ac9284cd8a1792d3cb26a2da18f3ca26c6908cc434a18f730dc0db7bfa3b", size = 1885092 }, + { url = "https://files.pythonhosted.org/packages/41/b1/9bc383f48f8002f99104e3acff6cba1231b29ef76cfa45d1506a5cad1f84/pydantic_core-2.27.2-cp313-cp313-macosx_10_12_x86_64.whl", hash = "sha256:7d14bd329640e63852364c306f4d23eb744e0f8193148d4044dd3dacdaacbd8b", size = 1892709 }, + { url = "https://files.pythonhosted.org/packages/10/6c/e62b8657b834f3eb2961b49ec8e301eb99946245e70bf42c8817350cbefc/pydantic_core-2.27.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:82f91663004eb8ed30ff478d77c4d1179b3563df6cdb15c0817cd1cdaf34d154", size = 1811273 }, + { url = "https://files.pythonhosted.org/packages/ba/15/52cfe49c8c986e081b863b102d6b859d9defc63446b642ccbbb3742bf371/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:71b24c7d61131bb83df10cc7e687433609963a944ccf45190cfc21e0887b08c9", size = 1823027 }, + { url = "https://files.pythonhosted.org/packages/b1/1c/b6f402cfc18ec0024120602bdbcebc7bdd5b856528c013bd4d13865ca473/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:fa8e459d4954f608fa26116118bb67f56b93b209c39b008277ace29937453dc9", size = 1868888 }, + { url = "https://files.pythonhosted.org/packages/bd/7b/8cb75b66ac37bc2975a3b7de99f3c6f355fcc4d89820b61dffa8f1e81677/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ce8918cbebc8da707ba805b7fd0b382816858728ae7fe19a942080c24e5b7cd1", size = 2037738 }, + { url = "https://files.pythonhosted.org/packages/c8/f1/786d8fe78970a06f61df22cba58e365ce304bf9b9f46cc71c8c424e0c334/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:eda3f5c2a021bbc5d976107bb302e0131351c2ba54343f8a496dc8783d3d3a6a", size = 2685138 }, + { url = "https://files.pythonhosted.org/packages/a6/74/d12b2cd841d8724dc8ffb13fc5cef86566a53ed358103150209ecd5d1999/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd8086fa684c4775c27f03f062cbb9eaa6e17f064307e86b21b9e0abc9c0f02e", size = 1997025 }, + { url = "https://files.pythonhosted.org/packages/a0/6e/940bcd631bc4d9a06c9539b51f070b66e8f370ed0933f392db6ff350d873/pydantic_core-2.27.2-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:8d9b3388db186ba0c099a6d20f0604a44eabdeef1777ddd94786cdae158729e4", size = 2004633 }, + { url = "https://files.pythonhosted.org/packages/50/cc/a46b34f1708d82498c227d5d80ce615b2dd502ddcfd8376fc14a36655af1/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:7a66efda2387de898c8f38c0cf7f14fca0b51a8ef0b24bfea5849f1b3c95af27", size = 1999404 }, + { url = "https://files.pythonhosted.org/packages/ca/2d/c365cfa930ed23bc58c41463bae347d1005537dc8db79e998af8ba28d35e/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_armv7l.whl", hash = "sha256:18a101c168e4e092ab40dbc2503bdc0f62010e95d292b27827871dc85450d7ee", size = 2130130 }, + { url = "https://files.pythonhosted.org/packages/f4/d7/eb64d015c350b7cdb371145b54d96c919d4db516817f31cd1c650cae3b21/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:ba5dd002f88b78a4215ed2f8ddbdf85e8513382820ba15ad5ad8955ce0ca19a1", size = 2157946 }, + { url = "https://files.pythonhosted.org/packages/a4/99/bddde3ddde76c03b65dfd5a66ab436c4e58ffc42927d4ff1198ffbf96f5f/pydantic_core-2.27.2-cp313-cp313-win32.whl", hash = "sha256:1ebaf1d0481914d004a573394f4be3a7616334be70261007e47c2a6fe7e50130", size = 1834387 }, + { url = "https://files.pythonhosted.org/packages/71/47/82b5e846e01b26ac6f1893d3c5f9f3a2eb6ba79be26eef0b759b4fe72946/pydantic_core-2.27.2-cp313-cp313-win_amd64.whl", hash = "sha256:953101387ecf2f5652883208769a79e48db18c6df442568a0b5ccd8c2723abee", size = 1990453 }, + { url = "https://files.pythonhosted.org/packages/51/b2/b2b50d5ecf21acf870190ae5d093602d95f66c9c31f9d5de6062eb329ad1/pydantic_core-2.27.2-cp313-cp313-win_arm64.whl", hash = "sha256:ac4dbfd1691affb8f48c2c13241a2e3b60ff23247cbcf981759c768b6633cf8b", size = 1885186 }, +] + +[[package]] +name = "pygments" +version = "2.19.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/7c/2d/c3338d48ea6cc0feb8446d8e6937e1408088a72a39937982cc6111d17f84/pygments-2.19.1.tar.gz", hash = "sha256:61c16d2a8576dc0649d9f39e089b5f02bcd27fba10d8fb4dcc28173f7a45151f", size = 4968581 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/8a/0b/9fcc47d19c48b59121088dd6da2488a49d5f72dacf8262e2790a1d2c7d15/pygments-2.19.1-py3-none-any.whl", hash = "sha256:9ea1544ad55cecf4b8242fab6dd35a93bbce657034b0611ee383099054ab6d8c", size = 1225293 }, +] + +[[package]] +name = "pyopenssl" +version = "24.3.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "cryptography" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/c1/d4/1067b82c4fc674d6f6e9e8d26b3dff978da46d351ca3bac171544693e085/pyopenssl-24.3.0.tar.gz", hash = "sha256:49f7a019577d834746bc55c5fce6ecbcec0f2b4ec5ce1cf43a9a173b8138bb36", size = 178944 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/42/22/40f9162e943f86f0fc927ebc648078be87def360d9d8db346619fb97df2b/pyOpenSSL-24.3.0-py3-none-any.whl", hash = "sha256:e474f5a473cd7f92221cc04976e48f4d11502804657a08a989fb3be5514c904a", size = 56111 }, +] + +[[package]] +name = "python-dateutil" +version = "2.9.0.post0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "six" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/66/c0/0c8b6ad9f17a802ee498c46e004a0eb49bc148f2fd230864601a86dcf6db/python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3", size = 342432 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427", size = 229892 }, +] + +[[package]] +name = "python-slugify" +version = "8.0.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "text-unidecode" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/87/c7/5e1547c44e31da50a460df93af11a535ace568ef89d7a811069ead340c4a/python-slugify-8.0.4.tar.gz", hash = "sha256:59202371d1d05b54a9e7720c5e038f928f45daaffe41dd10822f3907b937c856", size = 10921 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a4/62/02da182e544a51a5c3ccf4b03ab79df279f9c60c5e82d5e8bec7ca26ac11/python_slugify-8.0.4-py2.py3-none-any.whl", hash = "sha256:276540b79961052b66b7d116620b36518847f52d5fd9e3a70164fc8c50faa6b8", size = 10051 }, +] + +[[package]] +name = "pytz" +version = "2025.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/5f/57/df1c9157c8d5a05117e455d66fd7cf6dbc46974f832b1058ed4856785d8a/pytz-2025.1.tar.gz", hash = "sha256:c2db42be2a2518b28e65f9207c4d05e6ff547d1efa4086469ef855e4ab70178e", size = 319617 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/eb/38/ac33370d784287baa1c3d538978b5e2ea064d4c1b93ffbd12826c190dd10/pytz-2025.1-py2.py3-none-any.whl", hash = "sha256:89dd22dca55b46eac6eda23b2d72721bf1bdfef212645d81513ef5d03038de57", size = 507930 }, +] + +[[package]] +name = "pywin32" +version = "309" +source = { registry = "https://pypi.org/simple" } +wheels = [ + { url = "https://files.pythonhosted.org/packages/20/2c/b0240b14ff3dba7a8a7122dc9bbf7fbd21ed0e8b57c109633675b5d1761f/pywin32-309-cp312-cp312-win32.whl", hash = "sha256:de9acacced5fa82f557298b1fed5fef7bd49beee04190f68e1e4783fbdc19926", size = 8790648 }, + { url = "https://files.pythonhosted.org/packages/dd/11/c36884c732e2b3397deee808b5dac1abbb170ec37f94c6606fcb04d1e9d7/pywin32-309-cp312-cp312-win_amd64.whl", hash = "sha256:6ff9eebb77ffc3d59812c68db33c0a7817e1337e3537859499bd27586330fc9e", size = 9497399 }, + { url = "https://files.pythonhosted.org/packages/18/9f/79703972958f8ba3fd38bc9bf1165810bd75124982419b0cc433a2894d46/pywin32-309-cp312-cp312-win_arm64.whl", hash = "sha256:619f3e0a327b5418d833f44dc87859523635cf339f86071cc65a13c07be3110f", size = 8454122 }, + { url = "https://files.pythonhosted.org/packages/6c/c3/51aca6887cc5e410aa4cdc55662cf8438212440c67335c3f141b02eb8d52/pywin32-309-cp313-cp313-win32.whl", hash = "sha256:008bffd4afd6de8ca46c6486085414cc898263a21a63c7f860d54c9d02b45c8d", size = 8789700 }, + { url = "https://files.pythonhosted.org/packages/dd/66/330f265140fa814b4ed1bf16aea701f9d005f8f4ab57a54feb17f53afe7e/pywin32-309-cp313-cp313-win_amd64.whl", hash = "sha256:bd0724f58492db4cbfbeb1fcd606495205aa119370c0ddc4f70e5771a3ab768d", size = 9496714 }, + { url = "https://files.pythonhosted.org/packages/2c/84/9a51e6949a03f25cd329ece54dbf0846d57fadd2e79046c3b8d140aaa132/pywin32-309-cp313-cp313-win_arm64.whl", hash = "sha256:8fd9669cfd41863b688a1bc9b1d4d2d76fd4ba2128be50a70b0ea66b8d37953b", size = 8453052 }, +] + +[[package]] +name = "pyyaml" +version = "6.0.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/54/ed/79a089b6be93607fa5cdaedf301d7dfb23af5f25c398d5ead2525b063e17/pyyaml-6.0.2.tar.gz", hash = "sha256:d584d9ec91ad65861cc08d42e834324ef890a082e591037abe114850ff7bbc3e", size = 130631 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/86/0c/c581167fc46d6d6d7ddcfb8c843a4de25bdd27e4466938109ca68492292c/PyYAML-6.0.2-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:c70c95198c015b85feafc136515252a261a84561b7b1d51e3384e0655ddf25ab", size = 183873 }, + { url = "https://files.pythonhosted.org/packages/a8/0c/38374f5bb272c051e2a69281d71cba6fdb983413e6758b84482905e29a5d/PyYAML-6.0.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:ce826d6ef20b1bc864f0a68340c8b3287705cae2f8b4b1d932177dcc76721725", size = 173302 }, + { url = "https://files.pythonhosted.org/packages/c3/93/9916574aa8c00aa06bbac729972eb1071d002b8e158bd0e83a3b9a20a1f7/PyYAML-6.0.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1f71ea527786de97d1a0cc0eacd1defc0985dcf6b3f17bb77dcfc8c34bec4dc5", size = 739154 }, + { url = "https://files.pythonhosted.org/packages/95/0f/b8938f1cbd09739c6da569d172531567dbcc9789e0029aa070856f123984/PyYAML-6.0.2-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9b22676e8097e9e22e36d6b7bda33190d0d400f345f23d4065d48f4ca7ae0425", size = 766223 }, + { url = "https://files.pythonhosted.org/packages/b9/2b/614b4752f2e127db5cc206abc23a8c19678e92b23c3db30fc86ab731d3bd/PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:80bab7bfc629882493af4aa31a4cfa43a4c57c83813253626916b8c7ada83476", size = 767542 }, + { url = "https://files.pythonhosted.org/packages/d4/00/dd137d5bcc7efea1836d6264f049359861cf548469d18da90cd8216cf05f/PyYAML-6.0.2-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:0833f8694549e586547b576dcfaba4a6b55b9e96098b36cdc7ebefe667dfed48", size = 731164 }, + { url = "https://files.pythonhosted.org/packages/c9/1f/4f998c900485e5c0ef43838363ba4a9723ac0ad73a9dc42068b12aaba4e4/PyYAML-6.0.2-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:8b9c7197f7cb2738065c481a0461e50ad02f18c78cd75775628afb4d7137fb3b", size = 756611 }, + { url = "https://files.pythonhosted.org/packages/df/d1/f5a275fdb252768b7a11ec63585bc38d0e87c9e05668a139fea92b80634c/PyYAML-6.0.2-cp312-cp312-win32.whl", hash = "sha256:ef6107725bd54b262d6dedcc2af448a266975032bc85ef0172c5f059da6325b4", size = 140591 }, + { url = "https://files.pythonhosted.org/packages/0c/e8/4f648c598b17c3d06e8753d7d13d57542b30d56e6c2dedf9c331ae56312e/PyYAML-6.0.2-cp312-cp312-win_amd64.whl", hash = "sha256:7e7401d0de89a9a855c839bc697c079a4af81cf878373abd7dc625847d25cbd8", size = 156338 }, + { url = "https://files.pythonhosted.org/packages/ef/e3/3af305b830494fa85d95f6d95ef7fa73f2ee1cc8ef5b495c7c3269fb835f/PyYAML-6.0.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:efdca5630322a10774e8e98e1af481aad470dd62c3170801852d752aa7a783ba", size = 181309 }, + { url = "https://files.pythonhosted.org/packages/45/9f/3b1c20a0b7a3200524eb0076cc027a970d320bd3a6592873c85c92a08731/PyYAML-6.0.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:50187695423ffe49e2deacb8cd10510bc361faac997de9efef88badc3bb9e2d1", size = 171679 }, + { url = "https://files.pythonhosted.org/packages/7c/9a/337322f27005c33bcb656c655fa78325b730324c78620e8328ae28b64d0c/PyYAML-6.0.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0ffe8360bab4910ef1b9e87fb812d8bc0a308b0d0eef8c8f44e0254ab3b07133", size = 733428 }, + { url = "https://files.pythonhosted.org/packages/a3/69/864fbe19e6c18ea3cc196cbe5d392175b4cf3d5d0ac1403ec3f2d237ebb5/PyYAML-6.0.2-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:17e311b6c678207928d649faa7cb0d7b4c26a0ba73d41e99c4fff6b6c3276484", size = 763361 }, + { url = "https://files.pythonhosted.org/packages/04/24/b7721e4845c2f162d26f50521b825fb061bc0a5afcf9a386840f23ea19fa/PyYAML-6.0.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:70b189594dbe54f75ab3a1acec5f1e3faa7e8cf2f1e08d9b561cb41b845f69d5", size = 759523 }, + { url = "https://files.pythonhosted.org/packages/2b/b2/e3234f59ba06559c6ff63c4e10baea10e5e7df868092bf9ab40e5b9c56b6/PyYAML-6.0.2-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:41e4e3953a79407c794916fa277a82531dd93aad34e29c2a514c2c0c5fe971cc", size = 726660 }, + { url = "https://files.pythonhosted.org/packages/fe/0f/25911a9f080464c59fab9027482f822b86bf0608957a5fcc6eaac85aa515/PyYAML-6.0.2-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:68ccc6023a3400877818152ad9a1033e3db8625d899c72eacb5a668902e4d652", size = 751597 }, + { url = "https://files.pythonhosted.org/packages/14/0d/e2c3b43bbce3cf6bd97c840b46088a3031085179e596d4929729d8d68270/PyYAML-6.0.2-cp313-cp313-win32.whl", hash = "sha256:bc2fa7c6b47d6bc618dd7fb02ef6fdedb1090ec036abab80d4681424b84c1183", size = 140527 }, + { url = "https://files.pythonhosted.org/packages/fa/de/02b54f42487e3d3c6efb3f89428677074ca7bf43aae402517bc7cca949f3/PyYAML-6.0.2-cp313-cp313-win_amd64.whl", hash = "sha256:8388ee1976c416731879ac16da0aff3f63b286ffdd57cdeb95f3f2e085687563", size = 156446 }, +] + +[[package]] +name = "referencing" +version = "0.36.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "attrs" }, + { name = "rpds-py" }, + { name = "typing-extensions", marker = "python_full_version < '3.13'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/2f/db/98b5c277be99dd18bfd91dd04e1b759cad18d1a338188c936e92f921c7e2/referencing-0.36.2.tar.gz", hash = "sha256:df2e89862cd09deabbdba16944cc3f10feb6b3e6f18e902f7cc25609a34775aa", size = 74744 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/c1/b1/3baf80dc6d2b7bc27a95a67752d0208e410351e3feb4eb78de5f77454d8d/referencing-0.36.2-py3-none-any.whl", hash = "sha256:e8699adbbf8b5c7de96d8ffa0eb5c158b3beafce084968e2ea8bb08c6794dcd0", size = 26775 }, +] + +[[package]] +name = "regex" +version = "2024.11.6" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/8e/5f/bd69653fbfb76cf8604468d3b4ec4c403197144c7bfe0e6a5fc9e02a07cb/regex-2024.11.6.tar.gz", hash = "sha256:7ab159b063c52a0333c884e4679f8d7a85112ee3078fe3d9004b2dd875585519", size = 399494 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/ba/30/9a87ce8336b172cc232a0db89a3af97929d06c11ceaa19d97d84fa90a8f8/regex-2024.11.6-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:52fb28f528778f184f870b7cf8f225f5eef0a8f6e3778529bdd40c7b3920796a", size = 483781 }, + { url = "https://files.pythonhosted.org/packages/01/e8/00008ad4ff4be8b1844786ba6636035f7ef926db5686e4c0f98093612add/regex-2024.11.6-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:fdd6028445d2460f33136c55eeb1f601ab06d74cb3347132e1c24250187500d9", size = 288455 }, + { url = "https://files.pythonhosted.org/packages/60/85/cebcc0aff603ea0a201667b203f13ba75d9fc8668fab917ac5b2de3967bc/regex-2024.11.6-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:805e6b60c54bf766b251e94526ebad60b7de0c70f70a4e6210ee2891acb70bf2", size = 284759 }, + { url = "https://files.pythonhosted.org/packages/94/2b/701a4b0585cb05472a4da28ee28fdfe155f3638f5e1ec92306d924e5faf0/regex-2024.11.6-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b85c2530be953a890eaffde05485238f07029600e8f098cdf1848d414a8b45e4", size = 794976 }, + { url = "https://files.pythonhosted.org/packages/4b/bf/fa87e563bf5fee75db8915f7352e1887b1249126a1be4813837f5dbec965/regex-2024.11.6-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb26437975da7dc36b7efad18aa9dd4ea569d2357ae6b783bf1118dabd9ea577", size = 833077 }, + { url = "https://files.pythonhosted.org/packages/a1/56/7295e6bad94b047f4d0834e4779491b81216583c00c288252ef625c01d23/regex-2024.11.6-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:abfa5080c374a76a251ba60683242bc17eeb2c9818d0d30117b4486be10c59d3", size = 823160 }, + { url = "https://files.pythonhosted.org/packages/fb/13/e3b075031a738c9598c51cfbc4c7879e26729c53aa9cca59211c44235314/regex-2024.11.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:70b7fa6606c2881c1db9479b0eaa11ed5dfa11c8d60a474ff0e095099f39d98e", size = 796896 }, + { url = "https://files.pythonhosted.org/packages/24/56/0b3f1b66d592be6efec23a795b37732682520b47c53da5a32c33ed7d84e3/regex-2024.11.6-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0c32f75920cf99fe6b6c539c399a4a128452eaf1af27f39bce8909c9a3fd8cbe", size = 783997 }, + { url = "https://files.pythonhosted.org/packages/f9/a1/eb378dada8b91c0e4c5f08ffb56f25fcae47bf52ad18f9b2f33b83e6d498/regex-2024.11.6-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:982e6d21414e78e1f51cf595d7f321dcd14de1f2881c5dc6a6e23bbbbd68435e", size = 781725 }, + { url = "https://files.pythonhosted.org/packages/83/f2/033e7dec0cfd6dda93390089864732a3409246ffe8b042e9554afa9bff4e/regex-2024.11.6-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:a7c2155f790e2fb448faed6dd241386719802296ec588a8b9051c1f5c481bc29", size = 789481 }, + { url = "https://files.pythonhosted.org/packages/83/23/15d4552ea28990a74e7696780c438aadd73a20318c47e527b47a4a5a596d/regex-2024.11.6-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:149f5008d286636e48cd0b1dd65018548944e495b0265b45e1bffecce1ef7f39", size = 852896 }, + { url = "https://files.pythonhosted.org/packages/e3/39/ed4416bc90deedbfdada2568b2cb0bc1fdb98efe11f5378d9892b2a88f8f/regex-2024.11.6-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:e5364a4502efca094731680e80009632ad6624084aff9a23ce8c8c6820de3e51", size = 860138 }, + { url = "https://files.pythonhosted.org/packages/93/2d/dd56bb76bd8e95bbce684326302f287455b56242a4f9c61f1bc76e28360e/regex-2024.11.6-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:0a86e7eeca091c09e021db8eb72d54751e527fa47b8d5787caf96d9831bd02ad", size = 787692 }, + { url = "https://files.pythonhosted.org/packages/0b/55/31877a249ab7a5156758246b9c59539abbeba22461b7d8adc9e8475ff73e/regex-2024.11.6-cp312-cp312-win32.whl", hash = "sha256:32f9a4c643baad4efa81d549c2aadefaeba12249b2adc5af541759237eee1c54", size = 262135 }, + { url = "https://files.pythonhosted.org/packages/38/ec/ad2d7de49a600cdb8dd78434a1aeffe28b9d6fc42eb36afab4a27ad23384/regex-2024.11.6-cp312-cp312-win_amd64.whl", hash = "sha256:a93c194e2df18f7d264092dc8539b8ffb86b45b899ab976aa15d48214138e81b", size = 273567 }, + { url = "https://files.pythonhosted.org/packages/90/73/bcb0e36614601016552fa9344544a3a2ae1809dc1401b100eab02e772e1f/regex-2024.11.6-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:a6ba92c0bcdf96cbf43a12c717eae4bc98325ca3730f6b130ffa2e3c3c723d84", size = 483525 }, + { url = "https://files.pythonhosted.org/packages/0f/3f/f1a082a46b31e25291d830b369b6b0c5576a6f7fb89d3053a354c24b8a83/regex-2024.11.6-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:525eab0b789891ac3be914d36893bdf972d483fe66551f79d3e27146191a37d4", size = 288324 }, + { url = "https://files.pythonhosted.org/packages/09/c9/4e68181a4a652fb3ef5099e077faf4fd2a694ea6e0f806a7737aff9e758a/regex-2024.11.6-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:086a27a0b4ca227941700e0b31425e7a28ef1ae8e5e05a33826e17e47fbfdba0", size = 284617 }, + { url = "https://files.pythonhosted.org/packages/fc/fd/37868b75eaf63843165f1d2122ca6cb94bfc0271e4428cf58c0616786dce/regex-2024.11.6-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bde01f35767c4a7899b7eb6e823b125a64de314a8ee9791367c9a34d56af18d0", size = 795023 }, + { url = "https://files.pythonhosted.org/packages/c4/7c/d4cd9c528502a3dedb5c13c146e7a7a539a3853dc20209c8e75d9ba9d1b2/regex-2024.11.6-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b583904576650166b3d920d2bcce13971f6f9e9a396c673187f49811b2769dc7", size = 833072 }, + { url = "https://files.pythonhosted.org/packages/4f/db/46f563a08f969159c5a0f0e722260568425363bea43bb7ae370becb66a67/regex-2024.11.6-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1c4de13f06a0d54fa0d5ab1b7138bfa0d883220965a29616e3ea61b35d5f5fc7", size = 823130 }, + { url = "https://files.pythonhosted.org/packages/db/60/1eeca2074f5b87df394fccaa432ae3fc06c9c9bfa97c5051aed70e6e00c2/regex-2024.11.6-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3cde6e9f2580eb1665965ce9bf17ff4952f34f5b126beb509fee8f4e994f143c", size = 796857 }, + { url = "https://files.pythonhosted.org/packages/10/db/ac718a08fcee981554d2f7bb8402f1faa7e868c1345c16ab1ebec54b0d7b/regex-2024.11.6-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0d7f453dca13f40a02b79636a339c5b62b670141e63efd511d3f8f73fba162b3", size = 784006 }, + { url = "https://files.pythonhosted.org/packages/c2/41/7da3fe70216cea93144bf12da2b87367590bcf07db97604edeea55dac9ad/regex-2024.11.6-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:59dfe1ed21aea057a65c6b586afd2a945de04fc7db3de0a6e3ed5397ad491b07", size = 781650 }, + { url = "https://files.pythonhosted.org/packages/a7/d5/880921ee4eec393a4752e6ab9f0fe28009435417c3102fc413f3fe81c4e5/regex-2024.11.6-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:b97c1e0bd37c5cd7902e65f410779d39eeda155800b65fc4d04cc432efa9bc6e", size = 789545 }, + { url = "https://files.pythonhosted.org/packages/dc/96/53770115e507081122beca8899ab7f5ae28ae790bfcc82b5e38976df6a77/regex-2024.11.6-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:f9d1e379028e0fc2ae3654bac3cbbef81bf3fd571272a42d56c24007979bafb6", size = 853045 }, + { url = "https://files.pythonhosted.org/packages/31/d3/1372add5251cc2d44b451bd94f43b2ec78e15a6e82bff6a290ef9fd8f00a/regex-2024.11.6-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:13291b39131e2d002a7940fb176e120bec5145f3aeb7621be6534e46251912c4", size = 860182 }, + { url = "https://files.pythonhosted.org/packages/ed/e3/c446a64984ea9f69982ba1a69d4658d5014bc7a0ea468a07e1a1265db6e2/regex-2024.11.6-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:4f51f88c126370dcec4908576c5a627220da6c09d0bff31cfa89f2523843316d", size = 787733 }, + { url = "https://files.pythonhosted.org/packages/2b/f1/e40c8373e3480e4f29f2692bd21b3e05f296d3afebc7e5dcf21b9756ca1c/regex-2024.11.6-cp313-cp313-win32.whl", hash = "sha256:63b13cfd72e9601125027202cad74995ab26921d8cd935c25f09c630436348ff", size = 262122 }, + { url = "https://files.pythonhosted.org/packages/45/94/bc295babb3062a731f52621cdc992d123111282e291abaf23faa413443ea/regex-2024.11.6-cp313-cp313-win_amd64.whl", hash = "sha256:2b3361af3198667e99927da8b84c1b010752fa4b1115ee30beaa332cabc3ef1a", size = 273545 }, +] + +[[package]] +name = "requests" +version = "2.32.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "certifi" }, + { name = "charset-normalizer" }, + { name = "idna" }, + { name = "urllib3" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/63/70/2bf7780ad2d390a8d301ad0b550f1581eadbd9a20f896afe06353c2a2913/requests-2.32.3.tar.gz", hash = "sha256:55365417734eb18255590a9ff9eb97e9e1da868d4ccd6402399eaf68af20a760", size = 131218 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f9/9b/335f9764261e915ed497fcdeb11df5dfd6f7bf257d4a6a2a686d80da4d54/requests-2.32.3-py3-none-any.whl", hash = "sha256:70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6", size = 64928 }, +] + +[[package]] +name = "rich" +version = "13.9.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "markdown-it-py" }, + { name = "pygments" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/ab/3a/0316b28d0761c6734d6bc14e770d85506c986c85ffb239e688eeaab2c2bc/rich-13.9.4.tar.gz", hash = "sha256:439594978a49a09530cff7ebc4b5c7103ef57baf48d5ea3184f21d9a2befa098", size = 223149 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/19/71/39c7c0d87f8d4e6c020a393182060eaefeeae6c01dab6a84ec346f2567df/rich-13.9.4-py3-none-any.whl", hash = "sha256:6049d5e6ec054bf2779ab3358186963bac2ea89175919d699e378b99738c2a90", size = 242424 }, +] + +[[package]] +name = "rpds-py" +version = "0.23.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/0a/79/2ce611b18c4fd83d9e3aecb5cba93e1917c050f556db39842889fa69b79f/rpds_py-0.23.1.tar.gz", hash = "sha256:7f3240dcfa14d198dba24b8b9cb3b108c06b68d45b7babd9eefc1038fdf7e707", size = 26806 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f3/8c/d17efccb9f5b9137ddea706664aebae694384ae1d5997c0202093e37185a/rpds_py-0.23.1-cp312-cp312-macosx_10_12_x86_64.whl", hash = "sha256:3902df19540e9af4cc0c3ae75974c65d2c156b9257e91f5101a51f99136d834c", size = 364369 }, + { url = "https://files.pythonhosted.org/packages/6e/c0/ab030f696b5c573107115a88d8d73d80f03309e60952b64c584c70c659af/rpds_py-0.23.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:66f8d2a17e5838dd6fb9be6baaba8e75ae2f5fa6b6b755d597184bfcd3cb0eba", size = 349965 }, + { url = "https://files.pythonhosted.org/packages/b3/55/b40170f5a079c4fb0b6a82b299689e66e744edca3c3375a8b160fb797660/rpds_py-0.23.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:112b8774b0b4ee22368fec42749b94366bd9b536f8f74c3d4175d4395f5cbd31", size = 389064 }, + { url = "https://files.pythonhosted.org/packages/ab/1c/b03a912c59ec7c1e16b26e587b9dfa8ddff3b07851e781e8c46e908a365a/rpds_py-0.23.1-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:e0df046f2266e8586cf09d00588302a32923eb6386ced0ca5c9deade6af9a149", size = 397741 }, + { url = "https://files.pythonhosted.org/packages/52/6f/151b90792b62fb6f87099bcc9044c626881fdd54e31bf98541f830b15cea/rpds_py-0.23.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0f3288930b947cbebe767f84cf618d2cbe0b13be476e749da0e6a009f986248c", size = 448784 }, + { url = "https://files.pythonhosted.org/packages/71/2a/6de67c0c97ec7857e0e9e5cd7c52405af931b303eb1e5b9eff6c50fd9a2e/rpds_py-0.23.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ce473a2351c018b06dd8d30d5da8ab5a0831056cc53b2006e2a8028172c37ce5", size = 440203 }, + { url = "https://files.pythonhosted.org/packages/db/5e/e759cd1c276d98a4b1f464b17a9bf66c65d29f8f85754e27e1467feaa7c3/rpds_py-0.23.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d550d7e9e7d8676b183b37d65b5cd8de13676a738973d330b59dc8312df9c5dc", size = 391611 }, + { url = "https://files.pythonhosted.org/packages/1c/1e/2900358efcc0d9408c7289769cba4c0974d9db314aa884028ed7f7364f61/rpds_py-0.23.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e14f86b871ea74c3fddc9a40e947d6a5d09def5adc2076ee61fb910a9014fb35", size = 423306 }, + { url = "https://files.pythonhosted.org/packages/23/07/6c177e6d059f5d39689352d6c69a926ee4805ffdb6f06203570234d3d8f7/rpds_py-0.23.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:1bf5be5ba34e19be579ae873da515a2836a2166d8d7ee43be6ff909eda42b72b", size = 562323 }, + { url = "https://files.pythonhosted.org/packages/70/e4/f9097fd1c02b516fff9850792161eb9fc20a2fd54762f3c69eae0bdb67cb/rpds_py-0.23.1-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:d7031d493c4465dbc8d40bd6cafefef4bd472b17db0ab94c53e7909ee781b9ef", size = 588351 }, + { url = "https://files.pythonhosted.org/packages/87/39/5db3c6f326bfbe4576ae2af6435bd7555867d20ae690c786ff33659f293b/rpds_py-0.23.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:55ff4151cfd4bc635e51cfb1c59ac9f7196b256b12e3a57deb9e5742e65941ad", size = 557252 }, + { url = "https://files.pythonhosted.org/packages/fd/14/2d5ad292f144fa79bafb78d2eb5b8a3a91c358b6065443cb9c49b5d1fedf/rpds_py-0.23.1-cp312-cp312-win32.whl", hash = "sha256:a9d3b728f5a5873d84cba997b9d617c6090ca5721caaa691f3b1a78c60adc057", size = 222181 }, + { url = "https://files.pythonhosted.org/packages/a3/4f/0fce63e0f5cdd658e71e21abd17ac1bc9312741ebb8b3f74eeed2ebdf771/rpds_py-0.23.1-cp312-cp312-win_amd64.whl", hash = "sha256:b03a8d50b137ee758e4c73638b10747b7c39988eb8e6cd11abb7084266455165", size = 237426 }, + { url = "https://files.pythonhosted.org/packages/13/9d/b8b2c0edffb0bed15be17b6d5ab06216f2f47f9ee49259c7e96a3ad4ca42/rpds_py-0.23.1-cp313-cp313-macosx_10_12_x86_64.whl", hash = "sha256:4caafd1a22e5eaa3732acb7672a497123354bef79a9d7ceed43387d25025e935", size = 363672 }, + { url = "https://files.pythonhosted.org/packages/bd/c2/5056fa29e6894144d7ba4c938b9b0445f75836b87d2dd00ed4999dc45a8c/rpds_py-0.23.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:178f8a60fc24511c0eb756af741c476b87b610dba83270fce1e5a430204566a4", size = 349602 }, + { url = "https://files.pythonhosted.org/packages/b0/bc/33779a1bb0ee32d8d706b173825aab75c628521d23ce72a7c1e6a6852f86/rpds_py-0.23.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c632419c3870507ca20a37c8f8f5352317aca097639e524ad129f58c125c61c6", size = 388746 }, + { url = "https://files.pythonhosted.org/packages/62/0b/71db3e36b7780a619698ec82a9c87ab44ad7ca7f5480913e8a59ff76f050/rpds_py-0.23.1-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:698a79d295626ee292d1730bc2ef6e70a3ab135b1d79ada8fde3ed0047b65a10", size = 397076 }, + { url = "https://files.pythonhosted.org/packages/bb/2e/494398f613edf77ba10a916b1ddea2acce42ab0e3b62e2c70ffc0757ce00/rpds_py-0.23.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:271fa2184cf28bdded86bb6217c8e08d3a169fe0bbe9be5e8d96e8476b707122", size = 448399 }, + { url = "https://files.pythonhosted.org/packages/dd/53/4bd7f5779b1f463243ee5fdc83da04dd58a08f86e639dbffa7a35f969a84/rpds_py-0.23.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b91cceb5add79ee563bd1f70b30896bd63bc5f78a11c1f00a1e931729ca4f1f4", size = 439764 }, + { url = "https://files.pythonhosted.org/packages/f6/55/b3c18c04a460d951bf8e91f2abf46ce5b6426fb69784166a6a25827cb90a/rpds_py-0.23.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f3a6cb95074777f1ecda2ca4fa7717caa9ee6e534f42b7575a8f0d4cb0c24013", size = 390662 }, + { url = "https://files.pythonhosted.org/packages/2a/65/cc463044a3cbd616029b2aa87a651cdee8288d2fdd7780b2244845e934c1/rpds_py-0.23.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50fb62f8d8364978478b12d5f03bf028c6bc2af04082479299139dc26edf4c64", size = 422680 }, + { url = "https://files.pythonhosted.org/packages/fa/8e/1fa52990c7836d72e8d70cd7753f2362c72fbb0a49c1462e8c60e7176d0b/rpds_py-0.23.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:c8f7e90b948dc9dcfff8003f1ea3af08b29c062f681c05fd798e36daa3f7e3e8", size = 561792 }, + { url = "https://files.pythonhosted.org/packages/57/b8/fe3b612979b1a29d0c77f8585903d8b3a292604b26d4b300e228b8ac6360/rpds_py-0.23.1-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:5b98b6c953e5c2bda51ab4d5b4f172617d462eebc7f4bfdc7c7e6b423f6da957", size = 588127 }, + { url = "https://files.pythonhosted.org/packages/44/2d/fde474de516bbc4b9b230f43c98e7f8acc5da7fc50ceed8e7af27553d346/rpds_py-0.23.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:2893d778d4671ee627bac4037a075168b2673c57186fb1a57e993465dbd79a93", size = 556981 }, + { url = "https://files.pythonhosted.org/packages/18/57/767deeb27b81370bbab8f74ef6e68d26c4ea99018f3c71a570e506fede85/rpds_py-0.23.1-cp313-cp313-win32.whl", hash = "sha256:2cfa07c346a7ad07019c33fb9a63cf3acb1f5363c33bc73014e20d9fe8b01cdd", size = 221936 }, + { url = "https://files.pythonhosted.org/packages/7d/6c/3474cfdd3cafe243f97ab8474ea8949236eb2a1a341ca55e75ce00cd03da/rpds_py-0.23.1-cp313-cp313-win_amd64.whl", hash = "sha256:3aaf141d39f45322e44fc2c742e4b8b4098ead5317e5f884770c8df0c332da70", size = 237145 }, + { url = "https://files.pythonhosted.org/packages/ec/77/e985064c624230f61efa0423759bb066da56ebe40c654f8b5ba225bd5d63/rpds_py-0.23.1-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:759462b2d0aa5a04be5b3e37fb8183615f47014ae6b116e17036b131985cb731", size = 359623 }, + { url = "https://files.pythonhosted.org/packages/62/d9/a33dcbf62b29e40559e012d525bae7d516757cf042cc9234bd34ca4b6aeb/rpds_py-0.23.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:3e9212f52074fc9d72cf242a84063787ab8e21e0950d4d6709886fb62bcb91d5", size = 345900 }, + { url = "https://files.pythonhosted.org/packages/92/eb/f81a4be6397861adb2cb868bb6a28a33292c2dcac567d1dc575226055e55/rpds_py-0.23.1-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9e9f3a3ac919406bc0414bbbd76c6af99253c507150191ea79fab42fdb35982a", size = 386426 }, + { url = "https://files.pythonhosted.org/packages/09/47/1f810c9b5e83be005341201b5389f1d240dfa440346ea7189f9b3fd6961d/rpds_py-0.23.1-cp313-cp313t-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c04ca91dda8a61584165825907f5c967ca09e9c65fe8966ee753a3f2b019fe1e", size = 392314 }, + { url = "https://files.pythonhosted.org/packages/83/bd/bc95831432fd6c46ed8001f01af26de0763a059d6d7e6d69e3c5bf02917a/rpds_py-0.23.1-cp313-cp313t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4ab923167cfd945abb9b51a407407cf19f5bee35001221f2911dc85ffd35ff4f", size = 447706 }, + { url = "https://files.pythonhosted.org/packages/19/3e/567c04c226b1802dc6dc82cad3d53e1fa0a773258571c74ac5d8fbde97ed/rpds_py-0.23.1-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ed6f011bedca8585787e5082cce081bac3d30f54520097b2411351b3574e1219", size = 437060 }, + { url = "https://files.pythonhosted.org/packages/fe/77/a77d2c6afe27ae7d0d55fc32f6841502648070dc8d549fcc1e6d47ff8975/rpds_py-0.23.1-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6959bb9928c5c999aba4a3f5a6799d571ddc2c59ff49917ecf55be2bbb4e3722", size = 389347 }, + { url = "https://files.pythonhosted.org/packages/3f/47/6b256ff20a74cfebeac790ab05586e0ac91f88e331125d4740a6c86fc26f/rpds_py-0.23.1-cp313-cp313t-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:1ed7de3c86721b4e83ac440751329ec6a1102229aa18163f84c75b06b525ad7e", size = 415554 }, + { url = "https://files.pythonhosted.org/packages/fc/29/d4572469a245bc9fc81e35166dca19fc5298d5c43e1a6dd64bf145045193/rpds_py-0.23.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:5fb89edee2fa237584e532fbf78f0ddd1e49a47c7c8cfa153ab4849dc72a35e6", size = 557418 }, + { url = "https://files.pythonhosted.org/packages/9c/0a/68cf7228895b1a3f6f39f51b15830e62456795e61193d2c8b87fd48c60db/rpds_py-0.23.1-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:7e5413d2e2d86025e73f05510ad23dad5950ab8417b7fc6beaad99be8077138b", size = 583033 }, + { url = "https://files.pythonhosted.org/packages/14/18/017ab41dcd6649ad5db7d00155b4c212b31ab05bd857d5ba73a1617984eb/rpds_py-0.23.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:d31ed4987d72aabdf521eddfb6a72988703c091cfc0064330b9e5f8d6a042ff5", size = 554880 }, + { url = "https://files.pythonhosted.org/packages/2e/dd/17de89431268da8819d8d51ce67beac28d9b22fccf437bc5d6d2bcd1acdb/rpds_py-0.23.1-cp313-cp313t-win32.whl", hash = "sha256:f3429fb8e15b20961efca8c8b21432623d85db2228cc73fe22756c6637aa39e7", size = 219743 }, + { url = "https://files.pythonhosted.org/packages/68/15/6d22d07e063ce5e9bfbd96db9ec2fbb4693591b4503e3a76996639474d02/rpds_py-0.23.1-cp313-cp313t-win_amd64.whl", hash = "sha256:d6f6512a90bd5cd9030a6237f5346f046c6f0e40af98657568fa45695d4de59d", size = 235415 }, +] + +[[package]] +name = "ruamel-yaml" +version = "0.18.10" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "ruamel-yaml-clib", marker = "python_full_version < '3.13' and platform_python_implementation == 'CPython'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/ea/46/f44d8be06b85bc7c4d8c95d658be2b68f27711f279bf9dd0612a5e4794f5/ruamel.yaml-0.18.10.tar.gz", hash = "sha256:20c86ab29ac2153f80a428e1254a8adf686d3383df04490514ca3b79a362db58", size = 143447 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/c2/36/dfc1ebc0081e6d39924a2cc53654497f967a084a436bb64402dfce4254d9/ruamel.yaml-0.18.10-py3-none-any.whl", hash = "sha256:30f22513ab2301b3d2b577adc121c6471f28734d3d9728581245f1e76468b4f1", size = 117729 }, +] + +[[package]] +name = "ruamel-yaml-clib" +version = "0.2.12" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/20/84/80203abff8ea4993a87d823a5f632e4d92831ef75d404c9fc78d0176d2b5/ruamel.yaml.clib-0.2.12.tar.gz", hash = "sha256:6c8fbb13ec503f99a91901ab46e0b07ae7941cd527393187039aec586fdfd36f", size = 225315 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/48/41/e7a405afbdc26af961678474a55373e1b323605a4f5e2ddd4a80ea80f628/ruamel.yaml.clib-0.2.12-cp312-cp312-macosx_14_0_arm64.whl", hash = "sha256:20b0f8dc160ba83b6dcc0e256846e1a02d044e13f7ea74a3d1d56ede4e48c632", size = 133433 }, + { url = "https://files.pythonhosted.org/packages/ec/b0/b850385604334c2ce90e3ee1013bd911aedf058a934905863a6ea95e9eb4/ruamel.yaml.clib-0.2.12-cp312-cp312-manylinux2014_aarch64.whl", hash = "sha256:943f32bc9dedb3abff9879edc134901df92cfce2c3d5c9348f172f62eb2d771d", size = 647362 }, + { url = "https://files.pythonhosted.org/packages/44/d0/3f68a86e006448fb6c005aee66565b9eb89014a70c491d70c08de597f8e4/ruamel.yaml.clib-0.2.12-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:95c3829bb364fdb8e0332c9931ecf57d9be3519241323c5274bd82f709cebc0c", size = 754118 }, + { url = "https://files.pythonhosted.org/packages/52/a9/d39f3c5ada0a3bb2870d7db41901125dbe2434fa4f12ca8c5b83a42d7c53/ruamel.yaml.clib-0.2.12-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:749c16fcc4a2b09f28843cda5a193e0283e47454b63ec4b81eaa2242f50e4ccd", size = 706497 }, + { url = "https://files.pythonhosted.org/packages/b0/fa/097e38135dadd9ac25aecf2a54be17ddf6e4c23e43d538492a90ab3d71c6/ruamel.yaml.clib-0.2.12-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:bf165fef1f223beae7333275156ab2022cffe255dcc51c27f066b4370da81e31", size = 698042 }, + { url = "https://files.pythonhosted.org/packages/ec/d5/a659ca6f503b9379b930f13bc6b130c9f176469b73b9834296822a83a132/ruamel.yaml.clib-0.2.12-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:32621c177bbf782ca5a18ba4d7af0f1082a3f6e517ac2a18b3974d4edf349680", size = 745831 }, + { url = "https://files.pythonhosted.org/packages/db/5d/36619b61ffa2429eeaefaab4f3374666adf36ad8ac6330d855848d7d36fd/ruamel.yaml.clib-0.2.12-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:b82a7c94a498853aa0b272fd5bc67f29008da798d4f93a2f9f289feb8426a58d", size = 715692 }, + { url = "https://files.pythonhosted.org/packages/b1/82/85cb92f15a4231c89b95dfe08b09eb6adca929ef7df7e17ab59902b6f589/ruamel.yaml.clib-0.2.12-cp312-cp312-win32.whl", hash = "sha256:e8c4ebfcfd57177b572e2040777b8abc537cdef58a2120e830124946aa9b42c5", size = 98777 }, + { url = "https://files.pythonhosted.org/packages/d7/8f/c3654f6f1ddb75daf3922c3d8fc6005b1ab56671ad56ffb874d908bfa668/ruamel.yaml.clib-0.2.12-cp312-cp312-win_amd64.whl", hash = "sha256:0467c5965282c62203273b838ae77c0d29d7638c8a4e3a1c8bdd3602c10904e4", size = 115523 }, + { url = "https://files.pythonhosted.org/packages/29/00/4864119668d71a5fa45678f380b5923ff410701565821925c69780356ffa/ruamel.yaml.clib-0.2.12-cp313-cp313-macosx_14_0_arm64.whl", hash = "sha256:4c8c5d82f50bb53986a5e02d1b3092b03622c02c2eb78e29bec33fd9593bae1a", size = 132011 }, + { url = "https://files.pythonhosted.org/packages/7f/5e/212f473a93ae78c669ffa0cb051e3fee1139cb2d385d2ae1653d64281507/ruamel.yaml.clib-0.2.12-cp313-cp313-manylinux2014_aarch64.whl", hash = "sha256:e7e3736715fbf53e9be2a79eb4db68e4ed857017344d697e8b9749444ae57475", size = 642488 }, + { url = "https://files.pythonhosted.org/packages/1f/8f/ecfbe2123ade605c49ef769788f79c38ddb1c8fa81e01f4dbf5cf1a44b16/ruamel.yaml.clib-0.2.12-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b7e75b4965e1d4690e93021adfcecccbca7d61c7bddd8e22406ef2ff20d74ef", size = 745066 }, + { url = "https://files.pythonhosted.org/packages/e2/a9/28f60726d29dfc01b8decdb385de4ced2ced9faeb37a847bd5cf26836815/ruamel.yaml.clib-0.2.12-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:96777d473c05ee3e5e3c3e999f5d23c6f4ec5b0c38c098b3a5229085f74236c6", size = 701785 }, + { url = "https://files.pythonhosted.org/packages/84/7e/8e7ec45920daa7f76046578e4f677a3215fe8f18ee30a9cb7627a19d9b4c/ruamel.yaml.clib-0.2.12-cp313-cp313-musllinux_1_1_i686.whl", hash = "sha256:3bc2a80e6420ca8b7d3590791e2dfc709c88ab9152c00eeb511c9875ce5778bf", size = 693017 }, + { url = "https://files.pythonhosted.org/packages/c5/b3/d650eaade4ca225f02a648321e1ab835b9d361c60d51150bac49063b83fa/ruamel.yaml.clib-0.2.12-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:e188d2699864c11c36cdfdada94d781fd5d6b0071cd9c427bceb08ad3d7c70e1", size = 741270 }, + { url = "https://files.pythonhosted.org/packages/87/b8/01c29b924dcbbed75cc45b30c30d565d763b9c4d540545a0eeecffb8f09c/ruamel.yaml.clib-0.2.12-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:4f6f3eac23941b32afccc23081e1f50612bdbe4e982012ef4f5797986828cd01", size = 709059 }, + { url = "https://files.pythonhosted.org/packages/30/8c/ed73f047a73638257aa9377ad356bea4d96125b305c34a28766f4445cc0f/ruamel.yaml.clib-0.2.12-cp313-cp313-win32.whl", hash = "sha256:6442cb36270b3afb1b4951f060eccca1ce49f3d087ca1ca4563a6eb479cb3de6", size = 98583 }, + { url = "https://files.pythonhosted.org/packages/b0/85/e8e751d8791564dd333d5d9a4eab0a7a115f7e349595417fd50ecae3395c/ruamel.yaml.clib-0.2.12-cp313-cp313-win_amd64.whl", hash = "sha256:e5b8daf27af0b90da7bb903a876477a9e6d7270be6146906b276605997c7e9a3", size = 115190 }, +] + +[[package]] +name = "s3transfer" +version = "0.11.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "botocore" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/0f/ec/aa1a215e5c126fe5decbee2e107468f51d9ce190b9763cb649f76bb45938/s3transfer-0.11.4.tar.gz", hash = "sha256:559f161658e1cf0a911f45940552c696735f5c74e64362e515f333ebed87d679", size = 148419 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/86/62/8d3fc3ec6640161a5649b2cddbbf2b9fa39c92541225b33f117c37c5a2eb/s3transfer-0.11.4-py3-none-any.whl", hash = "sha256:ac265fa68318763a03bf2dc4f39d5cbd6a9e178d81cc9483ad27da33637e320d", size = 84412 }, +] + +[[package]] +name = "setuptools" +version = "76.0.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/32/d2/7b171caf085ba0d40d8391f54e1c75a1cda9255f542becf84575cfd8a732/setuptools-76.0.0.tar.gz", hash = "sha256:43b4ee60e10b0d0ee98ad11918e114c70701bc6051662a9a675a0496c1a158f4", size = 1349387 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/37/66/d2d7e6ad554f3a7c7297c3f8ef6e22643ad3d35ef5c63bf488bc89f32f31/setuptools-76.0.0-py3-none-any.whl", hash = "sha256:199466a166ff664970d0ee145839f5582cb9bca7a0a3a2e795b6a9cb2308e9c6", size = 1236106 }, +] + +[[package]] +name = "six" +version = "1.17.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/94/e7/b2c673351809dca68a0e064b6af791aa332cf192da575fd474ed7d6f16a2/six-1.17.0.tar.gz", hash = "sha256:ff70335d468e7eb6ec65b95b99d3a2836546063f63acc5171de367e834932a81", size = 34031 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl", hash = "sha256:4721f391ed90541fddacab5acf947aa0d3dc7d27b2e1e8eda2be8970586c3274", size = 11050 }, +] + +[[package]] +name = "sympy" +version = "1.13.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "mpmath" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/11/8a/5a7fd6284fa8caac23a26c9ddf9c30485a48169344b4bd3b0f02fef1890f/sympy-1.13.3.tar.gz", hash = "sha256:b27fd2c6530e0ab39e275fc9b683895367e51d5da91baa8d3d64db2565fec4d9", size = 7533196 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/99/ff/c87e0622b1dadea79d2fb0b25ade9ed98954c9033722eb707053d310d4f3/sympy-1.13.3-py3-none-any.whl", hash = "sha256:54612cf55a62755ee71824ce692986f23c88ffa77207b30c1368eda4a7060f73", size = 6189483 }, +] + +[[package]] +name = "test-lambda-locally" +version = "0" +source = { virtual = "." } +dependencies = [ + { name = "aws-sam-cli" }, +] + +[package.metadata] +requires-dist = [{ name = "aws-sam-cli", specifier = ">=1.135.0" }] + +[[package]] +name = "text-unidecode" +version = "1.3" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/ab/e2/e9a00f0ccb71718418230718b3d900e71a5d16e701a3dae079a21e9cd8f8/text-unidecode-1.3.tar.gz", hash = "sha256:bad6603bb14d279193107714b288be206cac565dfa49aa5b105294dd5c4aab93", size = 76885 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a6/a5/c0b6468d3824fe3fde30dbb5e1f687b291608f9473681bbf7dabbf5a87d7/text_unidecode-1.3-py2.py3-none-any.whl", hash = "sha256:1311f10e8b895935241623731c2ba64f4c455287888b18189350b67134a822e8", size = 78154 }, +] + +[[package]] +name = "tomlkit" +version = "0.13.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/b1/09/a439bec5888f00a54b8b9f05fa94d7f901d6735ef4e55dcec9bc37b5d8fa/tomlkit-0.13.2.tar.gz", hash = "sha256:fff5fe59a87295b278abd31bec92c15d9bc4a06885ab12bcea52c71119392e79", size = 192885 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f9/b6/a447b5e4ec71e13871be01ba81f5dfc9d0af7e473da256ff46bc0e24026f/tomlkit-0.13.2-py3-none-any.whl", hash = "sha256:7a974427f6e119197f670fbbbeae7bef749a6c14e793db934baefc1b5f03efde", size = 37955 }, +] + +[[package]] +name = "types-awscrt" +version = "0.24.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/88/6e/32779b967eee6ef627eaf10f3414163482b3980fc45ba21765fdd05359d4/types_awscrt-0.24.1.tar.gz", hash = "sha256:fc6eae56f8dc5a3f8cc93cc2c7c332fa82909f8284fbe25e014c575757af397d", size = 15450 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a6/1a/22e327d29fe231a10ed00e35ed2a100d2462cea253c3d24d41162769711a/types_awscrt-0.24.1-py3-none-any.whl", hash = "sha256:f3f2578ff74a254a79882b95961fb493ba217cebc350b3eb239d1cd948d4d7fa", size = 19414 }, +] + +[[package]] +name = "types-python-dateutil" +version = "2.9.0.20241206" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/a9/60/47d92293d9bc521cd2301e423a358abfac0ad409b3a1606d8fbae1321961/types_python_dateutil-2.9.0.20241206.tar.gz", hash = "sha256:18f493414c26ffba692a72369fea7a154c502646301ebfe3d56a04b3767284cb", size = 13802 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0f/b3/ca41df24db5eb99b00d97f89d7674a90cb6b3134c52fb8121b6d8d30f15c/types_python_dateutil-2.9.0.20241206-py3-none-any.whl", hash = "sha256:e248a4bc70a486d3e3ec84d0dc30eec3a5f979d6e7ee4123ae043eedbb987f53", size = 14384 }, +] + +[[package]] +name = "types-s3transfer" +version = "0.11.4" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/93/a9/440d8ba72a81bcf2cc5a56ef63f23b58ce93e7b9b62409697553bdcdd181/types_s3transfer-0.11.4.tar.gz", hash = "sha256:05fde593c84270f19fd053f0b1e08f5a057d7c5f036b9884e68fb8cd3041ac30", size = 14074 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d0/69/0b5ae42c3c33d31a32f7dcb9f35a3e327365360a6e4a2a7b491904bd38aa/types_s3transfer-0.11.4-py3-none-any.whl", hash = "sha256:2a76d92c07d4a3cb469e5343b2e7560e0b8078b2e03696a65407b8c44c861b61", size = 19516 }, +] + +[[package]] +name = "typing-extensions" +version = "4.12.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/df/db/f35a00659bc03fec321ba8bce9420de607a1d37f8342eee1863174c69557/typing_extensions-4.12.2.tar.gz", hash = "sha256:1a7ead55c7e559dd4dee8856e3a88b41225abfe1ce8df57b7c13915fe121ffb8", size = 85321 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/26/9f/ad63fc0248c5379346306f8668cda6e2e2e9c95e01216d2b8ffd9ff037d0/typing_extensions-4.12.2-py3-none-any.whl", hash = "sha256:04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d", size = 37438 }, +] + +[[package]] +name = "tzdata" +version = "2025.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/43/0f/fa4723f22942480be4ca9527bbde8d43f6c3f2fe8412f00e7f5f6746bc8b/tzdata-2025.1.tar.gz", hash = "sha256:24894909e88cdb28bd1636c6887801df64cb485bd593f2fd83ef29075a81d694", size = 194950 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0f/dd/84f10e23edd882c6f968c21c2434fe67bd4a528967067515feca9e611e5e/tzdata-2025.1-py2.py3-none-any.whl", hash = "sha256:7e127113816800496f027041c570f50bcd464a020098a3b6b199517772303639", size = 346762 }, +] + +[[package]] +name = "tzlocal" +version = "5.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "tzdata", marker = "sys_platform == 'win32'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/04/d3/c19d65ae67636fe63953b20c2e4a8ced4497ea232c43ff8d01db16de8dc0/tzlocal-5.2.tar.gz", hash = "sha256:8d399205578f1a9342816409cc1e46a93ebd5755e39ea2d85334bea911bf0e6e", size = 30201 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/97/3f/c4c51c55ff8487f2e6d0e618dba917e3c3ee2caae6cf0fbb59c9b1876f2e/tzlocal-5.2-py3-none-any.whl", hash = "sha256:49816ef2fe65ea8ac19d19aa7a1ae0551c834303d5014c6d5a62e4cbda8047b8", size = 17859 }, +] + +[[package]] +name = "urllib3" +version = "2.3.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/aa/63/e53da845320b757bf29ef6a9062f5c669fe997973f966045cb019c3f4b66/urllib3-2.3.0.tar.gz", hash = "sha256:f8c5449b3cf0861679ce7e0503c7b44b5ec981bec0d1d3795a07f1ba96f0204d", size = 307268 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/c8/19/4ec628951a74043532ca2cf5d97b7b14863931476d117c471e8e2b1eb39f/urllib3-2.3.0-py3-none-any.whl", hash = "sha256:1cee9ad369867bfdbbb48b7dd50374c0967a0bb7710050facf0dd6911440e3df", size = 128369 }, +] + +[[package]] +name = "watchdog" +version = "4.0.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/4f/38/764baaa25eb5e35c9a043d4c4588f9836edfe52a708950f4b6d5f714fd42/watchdog-4.0.2.tar.gz", hash = "sha256:b4dfbb6c49221be4535623ea4474a4d6ee0a9cef4a80b20c28db4d858b64e270", size = 126587 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/92/f5/ea22b095340545faea37ad9a42353b265ca751f543da3fb43f5d00cdcd21/watchdog-4.0.2-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:1cdcfd8142f604630deef34722d695fb455d04ab7cfe9963055df1fc69e6727a", size = 100342 }, + { url = "https://files.pythonhosted.org/packages/cb/d2/8ce97dff5e465db1222951434e3115189ae54a9863aef99c6987890cc9ef/watchdog-4.0.2-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:d7ab624ff2f663f98cd03c8b7eedc09375a911794dfea6bf2a359fcc266bff29", size = 92306 }, + { url = "https://files.pythonhosted.org/packages/49/c4/1aeba2c31b25f79b03b15918155bc8c0b08101054fc727900f1a577d0d54/watchdog-4.0.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:132937547a716027bd5714383dfc40dc66c26769f1ce8a72a859d6a48f371f3a", size = 92915 }, + { url = "https://files.pythonhosted.org/packages/79/63/eb8994a182672c042d85a33507475c50c2ee930577524dd97aea05251527/watchdog-4.0.2-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:cd67c7df93eb58f360c43802acc945fa8da70c675b6fa37a241e17ca698ca49b", size = 100343 }, + { url = "https://files.pythonhosted.org/packages/ce/82/027c0c65c2245769580605bcd20a1dc7dfd6c6683c8c4e2ef43920e38d27/watchdog-4.0.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:bcfd02377be80ef3b6bc4ce481ef3959640458d6feaae0bd43dd90a43da90a7d", size = 92313 }, + { url = "https://files.pythonhosted.org/packages/2a/89/ad4715cbbd3440cb0d336b78970aba243a33a24b1a79d66f8d16b4590d6a/watchdog-4.0.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:980b71510f59c884d684b3663d46e7a14b457c9611c481e5cef08f4dd022eed7", size = 92919 }, + { url = "https://files.pythonhosted.org/packages/8a/b1/25acf6767af6f7e44e0086309825bd8c098e301eed5868dc5350642124b9/watchdog-4.0.2-py3-none-manylinux2014_aarch64.whl", hash = "sha256:936acba76d636f70db8f3c66e76aa6cb5136a936fc2a5088b9ce1c7a3508fc83", size = 82947 }, + { url = "https://files.pythonhosted.org/packages/e8/90/aebac95d6f954bd4901f5d46dcd83d68e682bfd21798fd125a95ae1c9dbf/watchdog-4.0.2-py3-none-manylinux2014_armv7l.whl", hash = "sha256:e252f8ca942a870f38cf785aef420285431311652d871409a64e2a0a52a2174c", size = 82942 }, + { url = "https://files.pythonhosted.org/packages/15/3a/a4bd8f3b9381824995787488b9282aff1ed4667e1110f31a87b871ea851c/watchdog-4.0.2-py3-none-manylinux2014_i686.whl", hash = "sha256:0e83619a2d5d436a7e58a1aea957a3c1ccbf9782c43c0b4fed80580e5e4acd1a", size = 82947 }, + { url = "https://files.pythonhosted.org/packages/09/cc/238998fc08e292a4a18a852ed8274159019ee7a66be14441325bcd811dfd/watchdog-4.0.2-py3-none-manylinux2014_ppc64.whl", hash = "sha256:88456d65f207b39f1981bf772e473799fcdc10801062c36fd5ad9f9d1d463a73", size = 82946 }, + { url = "https://files.pythonhosted.org/packages/80/f1/d4b915160c9d677174aa5fae4537ae1f5acb23b3745ab0873071ef671f0a/watchdog-4.0.2-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:32be97f3b75693a93c683787a87a0dc8db98bb84701539954eef991fb35f5fbc", size = 82947 }, + { url = "https://files.pythonhosted.org/packages/db/02/56ebe2cf33b352fe3309588eb03f020d4d1c061563d9858a9216ba004259/watchdog-4.0.2-py3-none-manylinux2014_s390x.whl", hash = "sha256:c82253cfc9be68e3e49282831afad2c1f6593af80c0daf1287f6a92657986757", size = 82944 }, + { url = "https://files.pythonhosted.org/packages/01/d2/c8931ff840a7e5bd5dcb93f2bb2a1fd18faf8312e9f7f53ff1cf76ecc8ed/watchdog-4.0.2-py3-none-manylinux2014_x86_64.whl", hash = "sha256:c0b14488bd336c5b1845cee83d3e631a1f8b4e9c5091ec539406e4a324f882d8", size = 82947 }, + { url = "https://files.pythonhosted.org/packages/d0/d8/cdb0c21a4a988669d7c210c75c6a2c9a0e16a3b08d9f7e633df0d9a16ad8/watchdog-4.0.2-py3-none-win32.whl", hash = "sha256:0d8a7e523ef03757a5aa29f591437d64d0d894635f8a50f370fe37f913ce4e19", size = 82935 }, + { url = "https://files.pythonhosted.org/packages/99/2e/b69dfaae7a83ea64ce36538cc103a3065e12c447963797793d5c0a1d5130/watchdog-4.0.2-py3-none-win_amd64.whl", hash = "sha256:c344453ef3bf875a535b0488e3ad28e341adbd5a9ffb0f7d62cefacc8824ef2b", size = 82934 }, + { url = "https://files.pythonhosted.org/packages/b0/0b/43b96a9ecdd65ff5545b1b13b687ca486da5c6249475b1a45f24d63a1858/watchdog-4.0.2-py3-none-win_ia64.whl", hash = "sha256:baececaa8edff42cd16558a639a9b0ddf425f93d892e8392a56bf904f5eff22c", size = 82933 }, +] + +[[package]] +name = "werkzeug" +version = "3.1.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "markupsafe" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/9f/69/83029f1f6300c5fb2471d621ab06f6ec6b3324685a2ce0f9777fd4a8b71e/werkzeug-3.1.3.tar.gz", hash = "sha256:60723ce945c19328679790e3282cc758aa4a6040e4bb330f53d30fa546d44746", size = 806925 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl", hash = "sha256:54b78bf3716d19a65be4fceccc0d1d7b89e608834989dfae50ea87564639213e", size = 224498 }, +] + +[[package]] +name = "wheel" +version = "0.45.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/8a/98/2d9906746cdc6a6ef809ae6338005b3f21bb568bea3165cfc6a243fdc25c/wheel-0.45.1.tar.gz", hash = "sha256:661e1abd9198507b1409a20c02106d9670b2576e916d58f520316666abca6729", size = 107545 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0b/2c/87f3254fd8ffd29e4c02732eee68a83a1d3c346ae39bc6822dcbcb697f2b/wheel-0.45.1-py3-none-any.whl", hash = "sha256:708e7481cc80179af0e556bbf0cc00b8444c7321e2700b8d8580231d13017248", size = 72494 }, +] From 4c9731bbe68b6523cccec73fb764e04e61e441cb Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 12 Mar 2025 16:04:18 +0100 Subject: [PATCH 173/591] Coerce None values into strings in logentry params. (#4121) Nice rendering of log messages containing parameters that are `None` values does not work. There we coerce `None` values into strings to have nicer messages in Sentry UI. Fixes #3660 --- sentry_sdk/integrations/logging.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index b792510d6c..28809de4ab 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -248,7 +248,11 @@ def _emit(self, record): else: event["logentry"] = { "message": to_string(record.msg), - "params": record.args, + "params": ( + tuple(str(arg) if arg is None else arg for arg in record.args) + if record.args + else () + ), } event["extra"] = self._extra_from_record(record) From 78db2ec6b787b89c948ca1f049b688bb6300cff5 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Mar 2025 16:12:57 +0100 Subject: [PATCH 174/591] fix(bottle): Prevent internal error on 404 (#4131) `request.route` can throw a `RuntimeError: This request is not connected to a route.`. Closes https://github.com/getsentry/sentry-python/issues/3583 --- sentry_sdk/integrations/bottle.py | 18 ++++++++++++------ 1 file changed, 12 insertions(+), 6 deletions(-) diff --git a/sentry_sdk/integrations/bottle.py b/sentry_sdk/integrations/bottle.py index 148b86852e..8a9fc41208 100644 --- a/sentry_sdk/integrations/bottle.py +++ b/sentry_sdk/integrations/bottle.py @@ -177,14 +177,20 @@ def _set_transaction_name_and_source(event, transaction_style, request): name = "" if transaction_style == "url": - name = request.route.rule or "" + try: + name = request.route.rule or "" + except RuntimeError: + pass elif transaction_style == "endpoint": - name = ( - request.route.name - or transaction_from_function(request.route.callback) - or "" - ) + try: + name = ( + request.route.name + or transaction_from_function(request.route.callback) + or "" + ) + except RuntimeError: + pass event["transaction"] = name event["transaction_info"] = {"source": SOURCE_FOR_STYLE[transaction_style]} From 4ffefe42dc7135c4bd72efe652d2f066679bc7d8 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Mar 2025 16:20:32 +0100 Subject: [PATCH 175/591] tests: Add concurrency testcase for arq (#4125) --- tests/integrations/arq/test_arq.py | 47 ++++++++++++++++++++++++++++++ 1 file changed, 47 insertions(+) diff --git a/tests/integrations/arq/test_arq.py b/tests/integrations/arq/test_arq.py index e74395e26c..d8b7e715f2 100644 --- a/tests/integrations/arq/test_arq.py +++ b/tests/integrations/arq/test_arq.py @@ -1,4 +1,6 @@ import asyncio +from datetime import timedelta + import pytest from sentry_sdk import get_client, start_transaction @@ -376,3 +378,48 @@ async def job(ctx): assert event["contexts"]["trace"]["origin"] == "auto.queue.arq" assert event["spans"][0]["origin"] == "auto.db.redis" assert event["spans"][1]["origin"] == "auto.db.redis" + + +@pytest.mark.asyncio +async def test_job_concurrency(capture_events, init_arq): + """ + 10 - division starts + 70 - sleepy starts + 110 - division raises error + 120 - sleepy finishes + + """ + + async def sleepy(_): + await asyncio.sleep(0.05) + + async def division(_): + await asyncio.sleep(0.1) + return 1 / 0 + + sleepy.__qualname__ = sleepy.__name__ + division.__qualname__ = division.__name__ + + pool, worker = init_arq([sleepy, division]) + + events = capture_events() + + await pool.enqueue_job( + "division", _job_id="123", _defer_by=timedelta(milliseconds=10) + ) + await pool.enqueue_job( + "sleepy", _job_id="456", _defer_by=timedelta(milliseconds=70) + ) + + loop = asyncio.get_event_loop() + task = loop.create_task(worker.async_run()) + await asyncio.sleep(1) + + task.cancel() + + await worker.close() + + exception_event = events[1] + assert exception_event["exception"]["values"][0]["type"] == "ZeroDivisionError" + assert exception_event["transaction"] == "division" + assert exception_event["extra"]["arq-job"]["task"] == "division" From 4f51ff37a26b1e774b8050119da75074d1a1d5ed Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Mar 2025 16:21:27 +0100 Subject: [PATCH 176/591] fix(quart): Support `quart_flask_patch` (#4132) See https://github.com/getsentry/sentry-python/issues/2709#issuecomment-2006932012 If `quart_flask_patch` is imported, it monkeypatches stuff so that the Quart app appears to be a Flask app. This confuses our Flask integration, which tries to enable itself and fails. This commit: - Makes the Flask integration detect that what it sees as Flask might actually be Quart. - Reorganizes the Quart test suite a little to allow to test this case (a bit tricky since `import quart_flask_patch` needs to happen before anything else due to its monkeypatching nature). Closes https://github.com/getsentry/sentry-python/issues/2709 --- requirements-testing.txt | 2 +- scripts/populate_tox/tox.jinja | 1 + sentry_sdk/integrations/flask.py | 12 +++++ tests/integrations/quart/test_quart.py | 67 +++++++++++++++++++++----- tox.ini | 1 + 5 files changed, 71 insertions(+), 12 deletions(-) diff --git a/requirements-testing.txt b/requirements-testing.txt index 503ab5de68..cbc515eec2 100644 --- a/requirements-testing.txt +++ b/requirements-testing.txt @@ -14,4 +14,4 @@ socksio httpcore[http2] setuptools Brotli -docker \ No newline at end of file +docker diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 9da986a35a..5f1a26ac5e 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -384,6 +384,7 @@ deps = # Quart quart: quart-auth quart: pytest-asyncio + quart-{v0.19,latest}: quart-flask-patch quart-v0.16: blinker<1.6 quart-v0.16: jinja2<3.1.0 quart-v0.16: Werkzeug<2.1.0 diff --git a/sentry_sdk/integrations/flask.py b/sentry_sdk/integrations/flask.py index 45b4f0b2b1..f45ec6db20 100644 --- a/sentry_sdk/integrations/flask.py +++ b/sentry_sdk/integrations/flask.py @@ -72,6 +72,18 @@ def __init__( @staticmethod def setup_once(): # type: () -> None + try: + from quart import Quart # type: ignore + + if Flask == Quart: + # This is Quart masquerading as Flask, don't enable the Flask + # integration. See https://github.com/getsentry/sentry-python/issues/2709 + raise DidNotEnable( + "This is not a Flask app but rather Quart pretending to be Flask" + ) + except ImportError: + pass + version = package_version("flask") _check_minimum_version(FlaskIntegration, version) diff --git a/tests/integrations/quart/test_quart.py b/tests/integrations/quart/test_quart.py index f15b968ac5..100642d245 100644 --- a/tests/integrations/quart/test_quart.py +++ b/tests/integrations/quart/test_quart.py @@ -1,3 +1,4 @@ +import importlib import json import threading from unittest import mock @@ -13,22 +14,22 @@ from sentry_sdk.integrations.logging import LoggingIntegration import sentry_sdk.integrations.quart as quart_sentry -from quart import Quart, Response, abort, stream_with_context -from quart.views import View -from quart_auth import AuthUser, login_user - -try: - from quart_auth import QuartAuth +def quart_app_factory(): + # These imports are inlined because the `test_quart_flask_patch` testcase + # tests behavior that is triggered by importing a package before any Quart + # imports happen, so we can't have these on the module level + from quart import Quart - auth_manager = QuartAuth() -except ImportError: - from quart_auth import AuthManager + try: + from quart_auth import QuartAuth - auth_manager = AuthManager() + auth_manager = QuartAuth() + except ImportError: + from quart_auth import AuthManager + auth_manager = AuthManager() -def quart_app_factory(): app = Quart(__name__) app.debug = False app.config["TESTING"] = False @@ -71,6 +72,42 @@ def integration_enabled_params(request): raise ValueError(request.param) +@pytest.mark.asyncio +@pytest.mark.forked +@pytest.mark.skipif( + not importlib.util.find_spec("quart_flask_patch"), + reason="requires quart_flask_patch", +) +async def test_quart_flask_patch(sentry_init, capture_events, reset_integrations): + # This testcase is forked because `import quart_flask_patch` needs to run + # before anything else Quart-related is imported (since it monkeypatches + # some things) and we don't want this to affect other testcases. + # + # It's also important this testcase be run before any other testcase + # that uses `quart_app_factory`. + import quart_flask_patch # noqa: F401 + + app = quart_app_factory() + sentry_init( + integrations=[quart_sentry.QuartIntegration()], + ) + + @app.route("/") + async def index(): + 1 / 0 + + events = capture_events() + + client = app.test_client() + try: + await client.get("/") + except ZeroDivisionError: + pass + + (event,) = events + assert event["exception"]["values"][0]["mechanism"]["type"] == "quart" + + @pytest.mark.asyncio async def test_has_context(sentry_init, capture_events): sentry_init(integrations=[quart_sentry.QuartIntegration()]) @@ -213,6 +250,8 @@ async def test_quart_auth_configured( monkeypatch, integration_enabled_params, ): + from quart_auth import AuthUser, login_user + sentry_init(send_default_pii=send_default_pii, **integration_enabled_params) app = quart_app_factory() @@ -368,6 +407,8 @@ async def error_handler(err): @pytest.mark.asyncio async def test_bad_request_not_captured(sentry_init, capture_events): + from quart import abort + sentry_init(integrations=[quart_sentry.QuartIntegration()]) app = quart_app_factory() events = capture_events() @@ -385,6 +426,8 @@ async def index(): @pytest.mark.asyncio async def test_does_not_leak_scope(sentry_init, capture_events): + from quart import Response, stream_with_context + sentry_init(integrations=[quart_sentry.QuartIntegration()]) app = quart_app_factory() events = capture_events() @@ -514,6 +557,8 @@ async def error(): @pytest.mark.asyncio async def test_class_based_views(sentry_init, capture_events): + from quart.views import View + sentry_init(integrations=[quart_sentry.QuartIntegration()]) app = quart_app_factory() events = capture_events() diff --git a/tox.ini b/tox.ini index 932ef256ab..2294fcc00b 100644 --- a/tox.ini +++ b/tox.ini @@ -501,6 +501,7 @@ deps = # Quart quart: quart-auth quart: pytest-asyncio + quart-{v0.19,latest}: quart-flask-patch quart-v0.16: blinker<1.6 quart-v0.16: jinja2<3.1.0 quart-v0.16: Werkzeug<2.1.0 From 37930840dcefba96e7708b19e461013a919e83a5 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 13 Mar 2025 16:35:27 +0100 Subject: [PATCH 177/591] fix(debug): Take into account parent handlers for debug logger (#4133) We only check `logger.handlers` for existing handlers. This ignores any potential parent handlers. By using `hasHandlers()` ([docs](https://docs.python.org/3/library/logging.html#logging.Logger.hasHandlers)) instead we take those into account as well. Closes https://github.com/getsentry/sentry-python/issues/3944 --- sentry_sdk/debug.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/debug.py b/sentry_sdk/debug.py index e4c686a3e8..f740d92dec 100644 --- a/sentry_sdk/debug.py +++ b/sentry_sdk/debug.py @@ -19,7 +19,7 @@ def filter(self, record): def init_debug_support(): # type: () -> None - if not logger.handlers: + if not logger.hasHandlers(): configure_logger() From 380e32f29121bd203cd752f9c920fe54e4e8509d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 14 Mar 2025 13:43:17 +0100 Subject: [PATCH 178/591] Updating Readme (#4134) Dusting off our Readme a bit. It has been quite some time since it was last updated. --- README.md | 88 ++++++++++++++++++++++++++++++------------------------- 1 file changed, 48 insertions(+), 40 deletions(-) diff --git a/README.md b/README.md index 29501064f3..10bc8eb2ed 100644 --- a/README.md +++ b/README.md @@ -1,19 +1,32 @@ Sentry for Python +
+_Bad software is everywhere, and we're tired of it. Sentry is on a mission to help developers write better software faster, so we can get back to enjoying technology. If you want to join us +[**Check out our open positions**](https://sentry.io/careers/)_. + +[![Discord](https://img.shields.io/discord/621778831602221064?logo=discord&labelColor=%20%235462eb&logoColor=%20%23f5f5f5&color=%20%235462eb)](https://discord.gg/wdNEHETs87) +[![Twitter Follow](https://img.shields.io/twitter/follow/getsentry?label=@getsentry&style=social)](https://twitter.com/intent/follow?screen_name=getsentry) +[![PyPi page link -- version](https://img.shields.io/pypi/v/sentry-sdk.svg)](https://pypi.python.org/pypi/sentry-sdk) +python +[![Build Status](https://github.com/getsentry/sentry-python/actions/workflows/ci.yml/badge.svg)](https://github.com/getsentry/sentry-python/actions/workflows/ci.yml) + +
+ +
-_Bad software is everywhere, and we're tired of it. Sentry is on a mission to help developers write better software faster, so we can get back to enjoying technology. If you want to join us, [**check out our open positions**](https://sentry.io/careers/)_. # Official Sentry SDK for Python -[![Build Status](https://github.com/getsentry/sentry-python/actions/workflows/ci.yml/badge.svg)](https://github.com/getsentry/sentry-python/actions/workflows/ci.yml) -[![PyPi page link -- version](https://img.shields.io/pypi/v/sentry-sdk.svg)](https://pypi.python.org/pypi/sentry-sdk) -[![Discord](https://img.shields.io/discord/621778831602221064)](https://discord.gg/cWnMQeA) +Welcome to the official Python SDK for **[Sentry](http://sentry.io/)**. + + +## 📦 Getting Started -Welcome to the official Python SDK for **[Sentry](http://sentry.io/)**! +### Prerequisites -## Getting Started +You need a Sentry [account](https://sentry.io/signup/) and [project](https://docs.sentry.io/product/projects/). ### Installation @@ -25,7 +38,7 @@ pip install --upgrade sentry-sdk ### Basic Configuration -Here’s a quick configuration example to get Sentry up and running: +Here's a quick configuration example to get Sentry up and running: ```python import sentry_sdk @@ -34,7 +47,7 @@ sentry_sdk.init( "https://12927b5f211046b575ee51fd8b1ac34f@o1.ingest.sentry.io/1", # Your DSN here # Set traces_sample_rate to 1.0 to capture 100% - # of transactions for performance monitoring. + # of traces for performance monitoring. traces_sample_rate=1.0, ) ``` @@ -46,36 +59,26 @@ With this configuration, Sentry will monitor for exceptions and performance issu To generate some events that will show up in Sentry, you can log messages or capture errors: ```python -from sentry_sdk import capture_message -capture_message("Hello Sentry!") # You'll see this in your Sentry dashboard. +import sentry_sdk +sentry_sdk.init(...) # same as above + +sentry_sdk.capture_message("Hello Sentry!") # You'll see this in your Sentry dashboard. raise ValueError("Oops, something went wrong!") # This will create an error event in Sentry. ``` -#### Explore the Docs - -For more details on advanced usage, integrations, and customization, check out the full documentation: - -- [Official SDK Docs](https://docs.sentry.io/platforms/python/) -- [API Reference](https://getsentry.github.io/sentry-python/) -## Integrations +## 📚 Documentation -Sentry integrates with many popular Python libraries and frameworks, including: +For more details on advanced usage, integrations, and customization, check out the full documentation on [https://docs.sentry.io](https://docs.sentry.io/). -- [Django](https://docs.sentry.io/platforms/python/integrations/django/) -- [Flask](https://docs.sentry.io/platforms/python/integrations/flask/) -- [FastAPI](https://docs.sentry.io/platforms/python/integrations/fastapi/) -- [Celery](https://docs.sentry.io/platforms/python/integrations/celery/) -- [AWS Lambda](https://docs.sentry.io/platforms/python/integrations/aws-lambda/) -Want more? [Check out the full list of integrations](https://docs.sentry.io/platforms/python/integrations/). +## 🧩 Integrations -### Rolling Your Own Integration? +Sentry integrates with a ton of popular Python libraries and frameworks, including [FastAPI](https://docs.sentry.io/platforms/python/integrations/fastapi/), [Django](https://docs.sentry.io/platforms/python/integrations/django/), [Celery](https://docs.sentry.io/platforms/python/integrations/celery/), [OpenAI](https://docs.sentry.io/platforms/python/integrations/openai/) and many, many more. Check out the [full list of integrations](https://docs.sentry.io/platforms/python/integrations/) to get the full picture. -If you want to create a new integration or improve an existing one, we’d welcome your contributions! Please read our [contributing guide](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md) before starting. -## Migrating Between Versions? +## 🚧 Migrating Between Versions? ### From `1.x` to `2.x` @@ -85,30 +88,35 @@ If you're using the older `1.x` version of the SDK, now's the time to upgrade to Using the legacy `raven-python` client? It's now in maintenance mode, and we recommend migrating to the new SDK for an improved experience. Get all the details in our [migration guide](https://docs.sentry.io/platforms/python/migration/raven-to-sentry-sdk/). -## Want to Contribute? -We’d love your help in improving the Sentry SDK! Whether it’s fixing bugs, adding features, or enhancing documentation, every contribution is valuable. +## 🙌 Want to Contribute? -For details on how to contribute, please check out [CONTRIBUTING.md](CONTRIBUTING.md) and explore the [open issues](https://github.com/getsentry/sentry-python/issues). +We'd love your help in improving the Sentry SDK! Whether it's fixing bugs, adding features, writing new integrations, or enhancing documentation, every contribution is valuable. -## Need Help? +For details on how to contribute, please read our [contribution guide](CONTRIBUTING.md) and explore the [open issues](https://github.com/getsentry/sentry-python/issues). -If you encounter issues or need help setting up or configuring the SDK, don’t hesitate to reach out to the [Sentry Community on Discord](https://discord.com/invite/Ww9hbqr). There is a ton of great people there ready to help! -## Resources +## 🛟 Need Help? -Here are additional resources to help you make the most of Sentry: +If you encounter issues or need help setting up or configuring the SDK, don't hesitate to reach out to the [Sentry Community on Discord](https://discord.com/invite/Ww9hbqr). There is a ton of great people there ready to help! -- [![Documentation](https://img.shields.io/badge/documentation-sentry.io-green.svg)](https://docs.sentry.io/quickstart/) – Official documentation to get started. -- [![Discord](https://img.shields.io/discord/621778831602221064)](https://discord.gg/Ww9hbqr) – Join our Discord community. -- [![Twitter Follow](https://img.shields.io/twitter/follow/getsentry?label=getsentry&style=social)](https://twitter.com/intent/follow?screen_name=getsentry) – Follow us on X (Twitter) for updates. -- [![Stack Overflow](https://img.shields.io/badge/stack%20overflow-sentry-green.svg)](http://stackoverflow.com/questions/tagged/sentry) – Questions and answers related to Sentry. -## License +## 🔗 Resources + +Here are all resources to help you make the most of Sentry: + +- [Documentation](https://docs.sentry.io/platforms/python/) - Official documentation to get started. +- [Discord](https://img.shields.io/discord/621778831602221064) - Join our Discord community. +- [X/Twitter](https://twitter.com/intent/follow?screen_name=getsentry) - Follow us on X (Twitter) for updates. +- [Stack Overflow](https://stackoverflow.com/questions/tagged/sentry) - Questions and answers related to Sentry. + + +## 📃 License The SDK is open-source and available under the MIT license. Check out the [LICENSE](LICENSE) file for more information. ---- + +## 😘 Contributors Thanks to everyone who has helped improve the SDK! From 486d7338c5fff11c047ef657fff4217dc1f8b541 Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Mon, 17 Mar 2025 04:43:41 -0400 Subject: [PATCH 179/591] feat(logs): Add alpha version of Sentry logs (#4126) Logs are coming to sentry! This commit: - Adds `sentry_sdk._experimental_logger.{info, warn, ...}` methods - Adds `_experimental` options for `before_send_log` and `enable_sentry_logs` There are no tests (yet), and this still uses the otel_log schema. Example usage: ```python sentry_sdk.init( dsn=..., _experiments={"enable_sentry_logs": True}, ) from sentry_sdk import _experimental_logger as sentry_logger sentry_logger.info('Finished sending answer! #chunks={num_chunks}', num_chunks=10) ``` --------- Co-authored-by: Anton Pirker --- sentry_sdk/__init__.py | 1 + sentry_sdk/_experimental_logger.py | 20 +++ sentry_sdk/_types.py | 13 ++ sentry_sdk/client.py | 113 +++++++++++++- sentry_sdk/envelope.py | 8 + tests/test_logs.py | 242 +++++++++++++++++++++++++++++ 6 files changed, 396 insertions(+), 1 deletion(-) create mode 100644 sentry_sdk/_experimental_logger.py create mode 100644 tests/test_logs.py diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index 1c9cedec5f..4a0d551e5a 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -45,6 +45,7 @@ "start_transaction", "trace", "monitor", + "_experimental_logger.py", ] # Initialize the debug support after everything is loaded diff --git a/sentry_sdk/_experimental_logger.py b/sentry_sdk/_experimental_logger.py new file mode 100644 index 0000000000..1f3cd5e443 --- /dev/null +++ b/sentry_sdk/_experimental_logger.py @@ -0,0 +1,20 @@ +# NOTE: this is the logger sentry exposes to users, not some generic logger. +import functools +from typing import Any + +from sentry_sdk import get_client, get_current_scope + + +def _capture_log(severity_text, severity_number, template, **kwargs): + # type: (str, int, str, **Any) -> None + client = get_client() + scope = get_current_scope() + client.capture_log(scope, severity_text, severity_number, template, **kwargs) + + +trace = functools.partial(_capture_log, "trace", 1) +debug = functools.partial(_capture_log, "debug", 5) +info = functools.partial(_capture_log, "info", 9) +warn = functools.partial(_capture_log, "warn", 13) +error = functools.partial(_capture_log, "error", 17) +fatal = functools.partial(_capture_log, "fatal", 21) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 883b4cbc81..bc730719d2 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -207,6 +207,17 @@ class SDKInfo(TypedDict): ] Hint = Dict[str, Any] + Log = TypedDict( + "Log", + { + "severity_text": str, + "severity_number": int, + "body": str, + "attributes": dict[str, str | bool | float | int], + "time_unix_nano": int, + "trace_id": Optional[str], + }, + ) Breadcrumb = Dict[str, Any] BreadcrumbHint = Dict[str, Any] @@ -217,6 +228,7 @@ class SDKInfo(TypedDict): ErrorProcessor = Callable[[Event, ExcInfo], Optional[Event]] BreadcrumbProcessor = Callable[[Breadcrumb, BreadcrumbHint], Optional[Breadcrumb]] TransactionProcessor = Callable[[Event, Hint], Optional[Event]] + LogProcessor = Callable[[Log, Hint], Optional[Log]] TracesSampler = Callable[[SamplingContext], Union[float, int, bool]] @@ -237,6 +249,7 @@ class SDKInfo(TypedDict): "metric_bucket", "monitor", "span", + "log", ] SessionStatus = Literal["ok", "exited", "crashed", "abnormal"] diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 4f5c1566b3..5bbf919c02 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -1,7 +1,10 @@ +import json import os +import time import uuid import random import socket +import logging from collections.abc import Mapping from datetime import datetime, timezone from importlib import import_module @@ -55,7 +58,7 @@ from typing import Union from typing import TypeVar - from sentry_sdk._types import Event, Hint, SDKInfo + from sentry_sdk._types import Event, Hint, SDKInfo, Log from sentry_sdk.integrations import Integration from sentry_sdk.metrics import MetricsAggregator from sentry_sdk.scope import Scope @@ -206,6 +209,10 @@ def capture_event(self, *args, **kwargs): # type: (*Any, **Any) -> Optional[str] return None + def capture_log(self, scope, severity_text, severity_number, template, **kwargs): + # type: (Scope, str, int, str, **Any) -> None + pass + def capture_session(self, *args, **kwargs): # type: (*Any, **Any) -> None return None @@ -847,6 +854,110 @@ def capture_event( return return_value + def capture_log(self, scope, severity_text, severity_number, template, **kwargs): + # type: (Scope, str, int, str, **Any) -> None + logs_enabled = self.options["_experiments"].get("enable_sentry_logs", False) + if not logs_enabled: + return + + headers = { + "sent_at": format_timestamp(datetime.now(timezone.utc)), + } # type: dict[str, object] + + attrs = { + "sentry.message.template": template, + } # type: dict[str, str | bool | float | int] + + kwargs_attributes = kwargs.get("attributes") + if kwargs_attributes is not None: + attrs.update(kwargs_attributes) + + environment = self.options.get("environment") + if environment is not None: + attrs["sentry.environment"] = environment + + release = self.options.get("release") + if release is not None: + attrs["sentry.release"] = release + + span = scope.span + if span is not None: + attrs["sentry.trace.parent_span_id"] = span.span_id + + for k, v in kwargs.items(): + attrs[f"sentry.message.parameters.{k}"] = v + + log = { + "severity_text": severity_text, + "severity_number": severity_number, + "body": template.format(**kwargs), + "attributes": attrs, + "time_unix_nano": time.time_ns(), + "trace_id": None, + } # type: Log + + # If debug is enabled, log the log to the console + debug = self.options.get("debug", False) + if debug: + severity_text_to_logging_level = { + "trace": logging.DEBUG, + "debug": logging.DEBUG, + "info": logging.INFO, + "warn": logging.WARNING, + "error": logging.ERROR, + "fatal": logging.CRITICAL, + } + logger.log( + severity_text_to_logging_level.get(severity_text, logging.DEBUG), + f'[Sentry Logs] {log["body"]}', + ) + + propagation_context = scope.get_active_propagation_context() + if propagation_context is not None: + headers["trace_id"] = propagation_context.trace_id + log["trace_id"] = propagation_context.trace_id + + envelope = Envelope(headers=headers) + + before_emit_log = self.options["_experiments"].get("before_emit_log") + if before_emit_log is not None: + log = before_emit_log(log, {}) + if log is None: + return + + def format_attribute(key, val): + # type: (str, int | float | str | bool) -> Any + if isinstance(val, bool): + return {"key": key, "value": {"boolValue": val}} + if isinstance(val, int): + return {"key": key, "value": {"intValue": str(val)}} + if isinstance(val, float): + return {"key": key, "value": {"doubleValue": val}} + if isinstance(val, str): + return {"key": key, "value": {"stringValue": val}} + return {"key": key, "value": {"stringValue": json.dumps(val)}} + + otel_log = { + "severityText": log["severity_text"], + "severityNumber": log["severity_number"], + "body": {"stringValue": log["body"]}, + "timeUnixNano": str(log["time_unix_nano"]), + "attributes": [ + format_attribute(k, v) for (k, v) in log["attributes"].items() + ], + } + + if "trace_id" in log: + otel_log["traceId"] = log["trace_id"] + + envelope.add_log(otel_log) # TODO: batch these + + if self.spotlight: + self.spotlight.capture_envelope(envelope) + + if self.transport is not None: + self.transport.capture_envelope(envelope) + def capture_session( self, session # type: Session ): diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index 760116daa1..5f61e689c5 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -102,6 +102,12 @@ def add_sessions( # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=sessions), type="sessions")) + def add_log( + self, log # type: Any + ): + # type: (...) -> None + self.add_item(Item(payload=PayloadRef(json=log), type="otel_log")) + def add_item( self, item # type: Item ): @@ -268,6 +274,8 @@ def data_category(self): return "transaction" elif ty == "event": return "error" + elif ty == "otel_log": + return "log" elif ty == "client_report": return "internal" elif ty == "profile": diff --git a/tests/test_logs.py b/tests/test_logs.py new file mode 100644 index 0000000000..173a4028d6 --- /dev/null +++ b/tests/test_logs.py @@ -0,0 +1,242 @@ +import sys +from unittest import mock +import pytest + +import sentry_sdk +from sentry_sdk import _experimental_logger as sentry_logger + + +minimum_python_37 = pytest.mark.skipif( + sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" +) + + +@minimum_python_37 +def test_logs_disabled_by_default(sentry_init, capture_envelopes): + sentry_init() + envelopes = capture_envelopes() + + sentry_logger.trace("This is a 'trace' log.") + sentry_logger.debug("This is a 'debug' log...") + sentry_logger.info("This is a 'info' log...") + sentry_logger.warn("This is a 'warn' log...") + sentry_logger.error("This is a 'error' log...") + sentry_logger.fatal("This is a 'fatal' log...") + + assert len(envelopes) == 0 + + +@minimum_python_37 +def test_logs_basics(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + sentry_logger.trace("This is a 'trace' log...") + sentry_logger.debug("This is a 'debug' log...") + sentry_logger.info("This is a 'info' log...") + sentry_logger.warn("This is a 'warn' log...") + sentry_logger.error("This is a 'error' log...") + sentry_logger.fatal("This is a 'fatal' log...") + + assert ( + len(envelopes) == 6 + ) # We will batch those log items into a single envelope at some point + + assert envelopes[0].items[0].payload.json["severityText"] == "trace" + assert envelopes[0].items[0].payload.json["severityNumber"] == 1 + + assert envelopes[1].items[0].payload.json["severityText"] == "debug" + assert envelopes[1].items[0].payload.json["severityNumber"] == 5 + + assert envelopes[2].items[0].payload.json["severityText"] == "info" + assert envelopes[2].items[0].payload.json["severityNumber"] == 9 + + assert envelopes[3].items[0].payload.json["severityText"] == "warn" + assert envelopes[3].items[0].payload.json["severityNumber"] == 13 + + assert envelopes[4].items[0].payload.json["severityText"] == "error" + assert envelopes[4].items[0].payload.json["severityNumber"] == 17 + + assert envelopes[5].items[0].payload.json["severityText"] == "fatal" + assert envelopes[5].items[0].payload.json["severityNumber"] == 21 + + +@minimum_python_37 +def test_logs_before_emit_log(sentry_init, capture_envelopes): + def _before_log(record, hint): + assert list(record.keys()) == [ + "severity_text", + "severity_number", + "body", + "attributes", + "time_unix_nano", + "trace_id", + ] + + if record["severity_text"] in ["fatal", "error"]: + return None + + return record + + sentry_init( + _experiments={ + "enable_sentry_logs": True, + "before_emit_log": _before_log, + } + ) + envelopes = capture_envelopes() + + sentry_logger.trace("This is a 'trace' log...") + sentry_logger.debug("This is a 'debug' log...") + sentry_logger.info("This is a 'info' log...") + sentry_logger.warn("This is a 'warn' log...") + sentry_logger.error("This is a 'error' log...") + sentry_logger.fatal("This is a 'fatal' log...") + + assert len(envelopes) == 4 + + assert envelopes[0].items[0].payload.json["severityText"] == "trace" + assert envelopes[1].items[0].payload.json["severityText"] == "debug" + assert envelopes[2].items[0].payload.json["severityText"] == "info" + assert envelopes[3].items[0].payload.json["severityText"] == "warn" + + +@minimum_python_37 +def test_logs_attributes(sentry_init, capture_envelopes): + """ + Passing arbitrary attributes to log messages. + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + attrs = { + "attr_int": 1, + "attr_float": 2.0, + "attr_bool": True, + "attr_string": "string attribute", + } + + sentry_logger.warn( + "The recorded value was '{my_var}'", my_var="some value", attributes=attrs + ) + + log_item = envelopes[0].items[0].payload.json + assert log_item["body"]["stringValue"] == "The recorded value was 'some value'" + + assert log_item["attributes"][1] == { + "key": "attr_int", + "value": {"intValue": "1"}, + } # TODO: this is strange. + assert log_item["attributes"][2] == { + "key": "attr_float", + "value": {"doubleValue": 2.0}, + } + assert log_item["attributes"][3] == { + "key": "attr_bool", + "value": {"boolValue": True}, + } + assert log_item["attributes"][4] == { + "key": "attr_string", + "value": {"stringValue": "string attribute"}, + } + assert log_item["attributes"][5] == { + "key": "sentry.environment", + "value": {"stringValue": "production"}, + } + assert log_item["attributes"][6] == { + "key": "sentry.release", + "value": {"stringValue": mock.ANY}, + } + assert log_item["attributes"][7] == { + "key": "sentry.message.parameters.my_var", + "value": {"stringValue": "some value"}, + } + + +@minimum_python_37 +def test_logs_message_params(sentry_init, capture_envelopes): + """ + This is the official way of how to pass vars to log messages. + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + sentry_logger.warn("The recorded value was '{int_var}'", int_var=1) + sentry_logger.warn("The recorded value was '{float_var}'", float_var=2.0) + sentry_logger.warn("The recorded value was '{bool_var}'", bool_var=False) + sentry_logger.warn( + "The recorded value was '{string_var}'", string_var="some string value" + ) + + assert ( + envelopes[0].items[0].payload.json["body"]["stringValue"] + == "The recorded value was '1'" + ) + assert envelopes[0].items[0].payload.json["attributes"][-1] == { + "key": "sentry.message.parameters.int_var", + "value": {"intValue": "1"}, + } # TODO: this is strange. + + assert ( + envelopes[1].items[0].payload.json["body"]["stringValue"] + == "The recorded value was '2.0'" + ) + assert envelopes[1].items[0].payload.json["attributes"][-1] == { + "key": "sentry.message.parameters.float_var", + "value": {"doubleValue": 2.0}, + } + + assert ( + envelopes[2].items[0].payload.json["body"]["stringValue"] + == "The recorded value was 'False'" + ) + assert envelopes[2].items[0].payload.json["attributes"][-1] == { + "key": "sentry.message.parameters.bool_var", + "value": {"boolValue": False}, + } + + assert ( + envelopes[3].items[0].payload.json["body"]["stringValue"] + == "The recorded value was 'some string value'" + ) + assert envelopes[3].items[0].payload.json["attributes"][-1] == { + "key": "sentry.message.parameters.string_var", + "value": {"stringValue": "some string value"}, + } + + +@minimum_python_37 +def test_logs_tied_to_transactions(sentry_init, capture_envelopes): + """ + Log messages are also tied to transactions. + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + with sentry_sdk.start_transaction(name="test-transaction") as trx: + sentry_logger.warn("This is a log tied to a transaction") + + log_entry = envelopes[0].items[0].payload.json + assert log_entry["attributes"][-1] == { + "key": "sentry.trace.parent_span_id", + "value": {"stringValue": trx.span_id}, + } + + +@minimum_python_37 +def test_logs_tied_to_spans(sentry_init, capture_envelopes): + """ + Log messages are also tied to spans. + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + with sentry_sdk.start_transaction(name="test-transaction"): + with sentry_sdk.start_span(description="test-span") as span: + sentry_logger.warn("This is a log tied to a span") + + log_entry = envelopes[0].items[0].payload.json + assert log_entry["attributes"][-1] == { + "key": "sentry.trace.parent_span_id", + "value": {"stringValue": span.span_id}, + } From 5771f3e39e4bb0da0d158d31c701dda70511071d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 17 Mar 2025 09:49:37 +0100 Subject: [PATCH 180/591] Add `init()` parameters to ApiDocs. (#4100) Copied the text from docs.sentry.io and added it to the ApiDocs. (some parameters are undocumented, it seems) --- docs/api.rst | 8 + sentry_sdk/consts.py | 381 +++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 389 insertions(+) diff --git a/docs/api.rst b/docs/api.rst index 034652e05c..87c2535abd 100644 --- a/docs/api.rst +++ b/docs/api.rst @@ -5,6 +5,14 @@ Top Level API This is the user facing API of the SDK. It's exposed as ``sentry_sdk``. With this API you can implement a custom performance monitoring or error reporting solution. +Initializing the SDK +==================== + +.. autoclass:: sentry_sdk.client.ClientConstructor + :members: + :undoc-members: + :special-members: __init__ + :noindex: Capturing Data ============== diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 20179e2231..e617581b9e 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -561,6 +561,387 @@ def __init__( max_stack_frames=DEFAULT_MAX_STACK_FRAMES, # type: Optional[int] ): # type: (...) -> None + """Initialize the Sentry SDK with the given parameters. All parameters described here can be used in a call to `sentry_sdk.init()`. + + :param dsn: The DSN tells the SDK where to send the events. + + If this option is not set, the SDK will just not send any data. + + The `dsn` config option takes precedence over the environment variable. + + Learn more about `DSN utilization `_. + + :param debug: Turns debug mode on or off. + + When `True`, the SDK will attempt to print out debugging information. This can be useful if something goes + wrong with event sending. + + The default is always `False`. It's generally not recommended to turn it on in production because of the + increase in log output. + + The `debug` config option takes precedence over the environment variable. + + :param release: Sets the release. + + If not set, the SDK will try to automatically configure a release out of the box but it's a better idea to + manually set it to guarantee that the release is in sync with your deploy integrations. + + Release names are strings, but some formats are detected by Sentry and might be rendered differently. + + See `the releases documentation `_ to learn how the SDK tries to + automatically configure a release. + + The `release` config option takes precedence over the environment variable. + + Learn more about how to send release data so Sentry can tell you about regressions between releases and + identify the potential source in `the product documentation `_. + + :param environment: Sets the environment. This string is freeform and set to `production` by default. + + A release can be associated with more than one environment to separate them in the UI (think `staging` vs + `production` or similar). + + The `environment` config option takes precedence over the environment variable. + + :param dist: The distribution of the application. + + Distributions are used to disambiguate build or deployment variants of the same release of an application. + + The dist can be for example a build number. + + :param sample_rate: Configures the sample rate for error events, in the range of `0.0` to `1.0`. + + The default is `1.0`, which means that 100% of error events will be sent. If set to `0.1`, only 10% of + error events will be sent. + + Events are picked randomly. + + :param error_sampler: Dynamically configures the sample rate for error events on a per-event basis. + + This configuration option accepts a function, which takes two parameters (the `event` and the `hint`), and + which returns a boolean (indicating whether the event should be sent to Sentry) or a floating-point number + between `0.0` and `1.0`, inclusive. + + The number indicates the probability the event is sent to Sentry; the SDK will randomly decide whether to + send the event with the given probability. + + If this configuration option is specified, the `sample_rate` option is ignored. + + :param ignore_errors: A list of exception class names that shouldn't be sent to Sentry. + + Errors that are an instance of these exceptions or a subclass of them, will be filtered out before they're + sent to Sentry. + + By default, all errors are sent. + + :param max_breadcrumbs: This variable controls the total amount of breadcrumbs that should be captured. + + This defaults to `100`, but you can set this to any number. + + However, you should be aware that Sentry has a `maximum payload size `_ + and any events exceeding that payload size will be dropped. + + :param attach_stacktrace: When enabled, stack traces are automatically attached to all messages logged. + + Stack traces are always attached to exceptions; however, when this option is set, stack traces are also + sent with messages. + + This option means that stack traces appear next to all log messages. + + Grouping in Sentry is different for events with stack traces and without. As a result, you will get new + groups as you enable or disable this flag for certain events. + + :param send_default_pii: If this flag is enabled, `certain personally identifiable information (PII) + `_ is added by active integrations. + + If you enable this option, be sure to manually remove what you don't want to send using our features for + managing `Sensitive Data `_. + + :param event_scrubber: Scrubs the event payload for sensitive information such as cookies, sessions, and + passwords from a `denylist`. + + It can additionally be used to scrub from another `pii_denylist` if `send_default_pii` is disabled. + + See how to `configure the scrubber here `_. + + :param include_source_context: When enabled, source context will be included in events sent to Sentry. + + This source context includes the five lines of code above and below the line of code where an error + happened. + + :param include_local_variables: When enabled, the SDK will capture a snapshot of local variables to send with + the event to help with debugging. + + :param add_full_stack: When capturing errors, Sentry stack traces typically only include frames that start the + moment an error occurs. + + But if the `add_full_stack` option is enabled (set to `True`), all frames from the start of execution will + be included in the stack trace sent to Sentry. + + :param max_stack_frames: This option limits the number of stack frames that will be captured when + `add_full_stack` is enabled. + + :param server_name: This option can be used to supply a server name. + + When provided, the name of the server is sent along and persisted in the event. + + For many integrations, the server name actually corresponds to the device hostname, even in situations + where the machine is not actually a server. + + :param project_root: The full path to the root directory of your application. + + The `project_root` is used to mark frames in a stack trace either as being in your application or outside + of the application. + + :param in_app_include: A list of string prefixes of module names that belong to the app. + + This option takes precedence over `in_app_exclude`. + + Sentry differentiates stack frames that are directly related to your application ("in application") from + stack frames that come from other packages such as the standard library, frameworks, or other dependencies. + + The application package is automatically marked as `inApp`. + + The difference is visible in [sentry.io](https://sentry.io), where only the "in application" frames are + displayed by default. + + :param in_app_exclude: A list of string prefixes of module names that do not belong to the app, but rather to + third-party packages. + + Modules considered not part of the app will be hidden from stack traces by default. + + This option can be overridden using `in_app_include`. + + :param max_request_body_size: This parameter controls whether integrations should capture HTTP request bodies. + It can be set to one of the following values: + + - `never`: Request bodies are never sent. + - `small`: Only small request bodies will be captured. The cutoff for small depends on the SDK (typically + 4KB). + - `medium`: Medium and small requests will be captured (typically 10KB). + - `always`: The SDK will always capture the request body as long as Sentry can make sense of it. + + Please note that the Sentry server [limits HTTP request body size](https://develop.sentry.dev/sdk/ + expected-features/data-handling/#variable-size). The server always enforces its size limit, regardless of + how you configure this option. + + :param max_value_length: The number of characters after which the values containing text in the event payload + will be truncated. + + WARNING: If the value you set for this is exceptionally large, the event may exceed 1 MiB and will be + dropped by Sentry. + + :param ca_certs: A path to an alternative CA bundle file in PEM-format. + + :param send_client_reports: Set this boolean to `False` to disable sending of client reports. + + Client reports allow the client to send status reports about itself to Sentry, such as information about + events that were dropped before being sent. + + :param integrations: List of integrations to enable in addition to `auto-enabling integrations (overview) + `_. + + This setting can be used to override the default config options for a specific auto-enabling integration + or to add an integration that is not auto-enabled. + + :param disabled_integrations: List of integrations that will be disabled. + + This setting can be used to explicitly turn off specific `auto-enabling integrations (list) + `_ or + `default `_ integrations. + + :param auto_enabling_integrations: Configures whether `auto-enabling integrations (configuration) + `_ should be enabled. + + When set to `False`, no auto-enabling integrations will be enabled by default, even if the corresponding + framework/library is detected. + + :param default_integrations: Configures whether `default integrations + `_ should be enabled. + + Setting `default_integrations` to `False` disables all default integrations **as well as all auto-enabling + integrations**, unless they are specifically added in the `integrations` option, described above. + + :param before_send: This function is called with an SDK-specific message or error event object, and can return + a modified event object, or `null` to skip reporting the event. + + This can be used, for instance, for manual PII stripping before sending. + + By the time `before_send` is executed, all scope data has already been applied to the event. Further + modification of the scope won't have any effect. + + :param before_send_transaction: This function is called with an SDK-specific transaction event object, and can + return a modified transaction event object, or `null` to skip reporting the event. + + One way this might be used is for manual PII stripping before sending. + + :param before_breadcrumb: This function is called with an SDK-specific breadcrumb object before the breadcrumb + is added to the scope. + + When nothing is returned from the function, the breadcrumb is dropped. + + To pass the breadcrumb through, return the first argument, which contains the breadcrumb object. + + The callback typically gets a second argument (called a "hint") which contains the original object from + which the breadcrumb was created to further customize what the breadcrumb should look like. + + :param transport: Switches out the transport used to send events. + + How this works depends on the SDK. It can, for instance, be used to capture events for unit-testing or to + send it through some more complex setup that requires proxy authentication. + + :param transport_queue_size: The maximum number of events that will be queued before the transport is forced to + flush. + + :param http_proxy: When set, a proxy can be configured that should be used for outbound requests. + + This is also used for HTTPS requests unless a separate `https_proxy` is configured. However, not all SDKs + support a separate HTTPS proxy. + + SDKs will attempt to default to the system-wide configured proxy, if possible. For instance, on Unix + systems, the `http_proxy` environment variable will be picked up. + + :param https_proxy: Configures a separate proxy for outgoing HTTPS requests. + + This value might not be supported by all SDKs. When not supported the `http-proxy` value is also used for + HTTPS requests at all times. + + :param proxy_headers: A dict containing additional proxy headers (usually for authentication) to be forwarded + to `urllib3`'s `ProxyManager `_. + + :param shutdown_timeout: Controls how many seconds to wait before shutting down. + + Sentry SDKs send events from a background queue. This queue is given a certain amount to drain pending + events. The default is SDK specific but typically around two seconds. + + Setting this value too low may cause problems for sending events from command line applications. + + Setting the value too high will cause the application to block for a long time for users experiencing + network connectivity problems. + + :param keep_alive: Determines whether to keep the connection alive between requests. + + This can be useful in environments where you encounter frequent network issues such as connection resets. + + :param cert_file: Path to the client certificate to use. + + If set, supersedes the `CLIENT_CERT_FILE` environment variable. + + :param key_file: Path to the key file to use. + + If set, supersedes the `CLIENT_KEY_FILE` environment variable. + + :param socket_options: An optional list of socket options to use. + + These provide fine-grained, low-level control over the way the SDK connects to Sentry. + + If provided, the options will override the default `urllib3` `socket options + `_. + + :param traces_sample_rate: A number between `0` and `1`, controlling the percentage chance a given transaction + will be sent to Sentry. + + (`0` represents 0% while `1` represents 100%.) Applies equally to all transactions created in the app. + + Either this or `traces_sampler` must be defined to enable tracing. + + If `traces_sample_rate` is `0`, this means that no new traces will be created. However, if you have + another service (for example a JS frontend) that makes requests to your service that include trace + information, those traces will be continued and thus transactions will be sent to Sentry. + + If you want to disable all tracing you need to set `traces_sample_rate=None`. In this case, no new traces + will be started and no incoming traces will be continued. + + :param traces_sampler: A function responsible for determining the percentage chance a given transaction will be + sent to Sentry. + + It will automatically be passed information about the transaction and the context in which it's being + created, and must return a number between `0` (0% chance of being sent) and `1` (100% chance of being + sent). + + Can also be used for filtering transactions, by returning `0` for those that are unwanted. + + Either this or `traces_sample_rate` must be defined to enable tracing. + + :param trace_propagation_targets: An optional property that controls which downstream services receive tracing + data, in the form of a `sentry-trace` and a `baggage` header attached to any outgoing HTTP requests. + + The option may contain a list of strings or regex against which the URLs of outgoing requests are matched. + + If one of the entries in the list matches the URL of an outgoing request, trace data will be attached to + that request. + + String entries do not have to be full matches, meaning the URL of a request is matched when it _contains_ + a string provided through the option. + + If `trace_propagation_targets` is not provided, trace data is attached to every outgoing request from the + instrumented client. + + :param functions_to_trace: An optional list of functions that should be set up for tracing. + + For each function in the list, a span will be created when the function is executed. + + Functions in the list are represented as strings containing the fully qualified name of the function. + + This is a convenient option, making it possible to have one central place for configuring what functions + to trace, instead of having custom instrumentation scattered all over your code base. + + To learn more, see the `Custom Instrumentation `_ documentation. + + :param enable_backpressure_handling: When enabled, a new monitor thread will be spawned to perform health + checks on the SDK. + + If the system is unhealthy, the SDK will keep halving the `traces_sample_rate` set by you in 10 second + intervals until recovery. + + This down sampling helps ensure that the system stays stable and reduces SDK overhead under high load. + + This option is enabled by default. + + :param enable_db_query_source: When enabled, the source location will be added to database queries. + + :param db_query_source_threshold_ms: The threshold in milliseconds for adding the source location to database + queries. + + The query location will be added to the query for queries slower than the specified threshold. + + :param custom_repr: A custom `repr `_ function to run + while serializing an object. + + Use this to control how your custom objects and classes are visible in Sentry. + + Return a string for that repr value to be used or `None` to continue serializing how Sentry would have + done it anyway. + + :param profiles_sample_rate: A number between `0` and `1`, controlling the percentage chance a given sampled + transaction will be profiled. + + (`0` represents 0% while `1` represents 100%.) Applies equally to all transactions created in the app. + + This is relative to the tracing sample rate - e.g. `0.5` means 50% of sampled transactions will be + profiled. + + :param profiles_sampler: + + :param profiler_mode: + + :param profile_lifecycle: + + :param profile_session_sample_rate: + + + :param enable_tracing: + + :param propagate_traces: + + :param auto_session_tracking: + + :param spotlight: + + :param instrumenter: + + :param _experiments: + """ pass From 7a3834776135715bd0d8cd6fc0a8a6d98b9f0fdc Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 17 Mar 2025 10:06:42 +0100 Subject: [PATCH 181/591] docs(baggage): Document that caller must check `mutable` (#4010) The `Baggage` class does not enforce mutability. Document this to avoid confusion. --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. The AWS Lambda tests additionally require a maintainer to add a special label, and they will fail until this label is added. Co-authored-by: Anton Pirker --- sentry_sdk/tracing_utils.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index b1e2050708..6aa4e4882a 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -543,6 +543,10 @@ def _sample_rand(self): class Baggage: """ The W3C Baggage header information (see https://www.w3.org/TR/baggage/). + + Before mutating a `Baggage` object, calling code must check that `mutable` is `True`. + Mutating a `Baggage` object that has `mutable` set to `False` is not allowed, but + it is the caller's responsibility to enforce this restriction. """ __slots__ = ("sentry_items", "third_party_items", "mutable") From 59ed713dfd620758c7bb373302b84937378088d2 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Mar 2025 09:16:49 +0000 Subject: [PATCH 182/591] build(deps): bump codecov/codecov-action from 5.3.1 to 5.4.0 (#4112) --- .github/workflows/test-integrations-ai.yml | 4 +- .github/workflows/test-integrations-aws.yml | 126 ++++++++++++++++++ .github/workflows/test-integrations-cloud.yml | 4 +- .../workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 +- .github/workflows/test-integrations-flags.yml | 2 +- .../workflows/test-integrations-gevent.yml | 2 +- .../workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .../workflows/test-integrations-network.yml | 4 +- .github/workflows/test-integrations-tasks.yml | 4 +- .github/workflows/test-integrations-web-1.yml | 4 +- .github/workflows/test-integrations-web-2.yml | 4 +- .../templates/test_group.jinja | 2 +- 14 files changed, 146 insertions(+), 20 deletions(-) create mode 100644 .github/workflows/test-integrations-aws.yml diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 1a5df1d00f..2b2e13059b 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -80,7 +80,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -152,7 +152,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml new file mode 100644 index 0000000000..9d9994dcfb --- /dev/null +++ b/.github/workflows/test-integrations-aws.yml @@ -0,0 +1,126 @@ +# Do not edit this YAML file. This file is generated automatically by executing +# python scripts/split_tox_gh_actions/split_tox_gh_actions.py +# The template responsible for it is in +# scripts/split_tox_gh_actions/templates/base.jinja +name: Test AWS +on: + push: + branches: + - master + - release/** + - potel-base + # XXX: We are using `pull_request_target` instead of `pull_request` because we want + # this to run on forks with access to the secrets necessary to run the test suite. + # Prefer to use `pull_request` when possible. + pull_request_target: + types: [labeled, opened, reopened, synchronize] +# Cancel in progress workflows on pull_requests. +# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value +concurrency: + group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} + cancel-in-progress: true +permissions: + contents: read + # `write` is needed to remove the `Trigger: tests using secrets` label + pull-requests: write +env: + SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID: ${{ secrets.SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID }} + SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY: ${{ secrets.SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY }} + BUILD_CACHE_KEY: ${{ github.sha }} + CACHED_BUILD_PATHS: | + ${{ github.workspace }}/dist-serverless +jobs: + check-permissions: + name: permissions check + runs-on: ubuntu-20.04 + steps: + - uses: actions/checkout@v4.2.2 + with: + persist-credentials: false + - name: Check permissions on PR + if: github.event_name == 'pull_request_target' + run: | + python3 -uS .github/workflows/scripts/trigger_tests_on_label.py \ + --repo-id ${{ github.event.repository.id }} \ + --pr ${{ github.event.number }} \ + --event ${{ github.event.action }} \ + --username "$ARG_USERNAME" \ + --label-names "$ARG_LABEL_NAMES" + env: + GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} + # these can contain special characters + ARG_USERNAME: ${{ github.event.pull_request.user.login }} + ARG_LABEL_NAMES: ${{ toJSON(github.event.pull_request.labels.*.name) }} + - name: Check permissions on repo branch + if: github.event_name == 'push' + run: true + test-aws-pinned: + name: AWS (pinned) + timeout-minutes: 30 + runs-on: ${{ matrix.os }} + strategy: + fail-fast: false + matrix: + python-version: ["3.9"] + # python3.6 reached EOL and is no longer being supported on + # new versions of hosted runners on Github Actions + # ubuntu-20.04 is the last version that supported python3.6 + # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 + os: [ubuntu-20.04] + needs: check-permissions + steps: + - uses: actions/checkout@v4.2.2 + with: + ref: ${{ github.event.pull_request.head.sha || github.ref }} + - uses: actions/setup-python@v5 + with: + python-version: ${{ matrix.python-version }} + allow-prereleases: true + - name: Setup Test Env + run: | + pip install "coverage[toml]" tox + - name: Erase coverage + run: | + coverage erase + - name: Test aws_lambda pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aws_lambda" + - name: Generate coverage XML (Python 3.6) + if: ${{ !cancelled() && matrix.python-version == '3.6' }} + run: | + export COVERAGE_RCFILE=.coveragerc36 + coverage combine .coverage-sentry-* + coverage xml --ignore-errors + - name: Generate coverage XML + if: ${{ !cancelled() && matrix.python-version != '3.6' }} + run: | + coverage combine .coverage-sentry-* + coverage xml + - name: Upload coverage to Codecov + if: ${{ !cancelled() }} + uses: codecov/codecov-action@v5.4.0 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: coverage.xml + # make sure no plugins alter our coverage reports + plugin: noop + verbose: true + - name: Upload test results to Codecov + if: ${{ !cancelled() }} + uses: codecov/test-results-action@v1 + with: + token: ${{ secrets.CODECOV_TOKEN }} + files: .junitxml + verbose: true + check_required_tests: + name: All pinned AWS tests passed + needs: test-aws-pinned + # Always run this, even if a dependent job failed + if: always() + runs-on: ubuntu-20.04 + steps: + - name: Check for failures + if: contains(needs.test-aws-pinned.result, 'failure') || contains(needs.test-aws-pinned.result, 'skipped') + run: | + echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index efa71c8e0c..0468518ec6 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -84,7 +84,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -160,7 +160,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 11506d0f0f..b1bdc564f3 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -64,7 +64,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 1fb0aa0715..ed35630da6 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -104,7 +104,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -200,7 +200,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index ad344762ae..d3ec53de62 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 2729c3e701..e9c64d568b 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -64,7 +64,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index f3015ae5bf..235e660474 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -76,7 +76,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 4e582c6c71..0db363c3c1 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -84,7 +84,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index aae29ab7f9..96ecdbe5ad 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -72,7 +72,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -136,7 +136,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 6abefa29f4..a5ed395f32 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index e243ceb69a..72cc958308 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -94,7 +94,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -180,7 +180,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index b3973aa960..a06ad23b32 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -100,7 +100,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -192,7 +192,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 9fcc0b1527..5ff68e37dc 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -89,7 +89,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.3.1 + uses: codecov/codecov-action@v5.4.0 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From e06ea8dec22e4986a8485ee6dee64c99520e9282 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Mar 2025 09:32:30 +0000 Subject: [PATCH 183/591] build(deps): bump actions/create-github-app-token from 1.11.5 to 1.11.6 (#4113) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.5 to 1.11.6.
Release notes

Sourced from actions/create-github-app-token's releases.

v1.11.6

1.11.6 (2025-03-03)

Bug Fixes

  • deps: bump the production-dependencies group with 2 updates (#210) (1ff1dea)
Commits
  • 21cfef2 build(release): 1.11.6 [skip ci]
  • 1ff1dea fix(deps): bump the production-dependencies group with 2 updates (#210)
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=1.11.5&new-version=1.11.6)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/release.yml | 2 +- .github/workflows/test-integrations-aws.yml | 126 -------------------- 2 files changed, 1 insertion(+), 127 deletions(-) delete mode 100644 .github/workflows/test-integrations-aws.yml diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 4d8c060f6a..c1861ce182 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@0d564482f06ca65fa9e77e2510873638c82206f2 # v1.11.5 + uses: actions/create-github-app-token@21cfef2b496dd8ef5b904c159339626a10ad380e # v1.11.6 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} diff --git a/.github/workflows/test-integrations-aws.yml b/.github/workflows/test-integrations-aws.yml deleted file mode 100644 index 9d9994dcfb..0000000000 --- a/.github/workflows/test-integrations-aws.yml +++ /dev/null @@ -1,126 +0,0 @@ -# Do not edit this YAML file. This file is generated automatically by executing -# python scripts/split_tox_gh_actions/split_tox_gh_actions.py -# The template responsible for it is in -# scripts/split_tox_gh_actions/templates/base.jinja -name: Test AWS -on: - push: - branches: - - master - - release/** - - potel-base - # XXX: We are using `pull_request_target` instead of `pull_request` because we want - # this to run on forks with access to the secrets necessary to run the test suite. - # Prefer to use `pull_request` when possible. - pull_request_target: - types: [labeled, opened, reopened, synchronize] -# Cancel in progress workflows on pull_requests. -# https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value -concurrency: - group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} - cancel-in-progress: true -permissions: - contents: read - # `write` is needed to remove the `Trigger: tests using secrets` label - pull-requests: write -env: - SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID: ${{ secrets.SENTRY_PYTHON_TEST_AWS_ACCESS_KEY_ID }} - SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY: ${{ secrets.SENTRY_PYTHON_TEST_AWS_SECRET_ACCESS_KEY }} - BUILD_CACHE_KEY: ${{ github.sha }} - CACHED_BUILD_PATHS: | - ${{ github.workspace }}/dist-serverless -jobs: - check-permissions: - name: permissions check - runs-on: ubuntu-20.04 - steps: - - uses: actions/checkout@v4.2.2 - with: - persist-credentials: false - - name: Check permissions on PR - if: github.event_name == 'pull_request_target' - run: | - python3 -uS .github/workflows/scripts/trigger_tests_on_label.py \ - --repo-id ${{ github.event.repository.id }} \ - --pr ${{ github.event.number }} \ - --event ${{ github.event.action }} \ - --username "$ARG_USERNAME" \ - --label-names "$ARG_LABEL_NAMES" - env: - GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} - # these can contain special characters - ARG_USERNAME: ${{ github.event.pull_request.user.login }} - ARG_LABEL_NAMES: ${{ toJSON(github.event.pull_request.labels.*.name) }} - - name: Check permissions on repo branch - if: github.event_name == 'push' - run: true - test-aws-pinned: - name: AWS (pinned) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.9"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] - needs: check-permissions - steps: - - uses: actions/checkout@v4.2.2 - with: - ref: ${{ github.event.pull_request.head.sha || github.ref }} - - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test aws_lambda pinned - run: | - set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aws_lambda" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true - check_required_tests: - name: All pinned AWS tests passed - needs: test-aws-pinned - # Always run this, even if a dependent job failed - if: always() - runs-on: ubuntu-20.04 - steps: - - name: Check for failures - if: contains(needs.test-aws-pinned.result, 'failure') || contains(needs.test-aws-pinned.result, 'skipped') - run: | - echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 From 88a048ff21f70a65d1b8b8c0b9eb5729acae5e6d Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 17 Mar 2025 09:45:14 +0000 Subject: [PATCH 184/591] release: 2.23.0 --- CHANGELOG.md | 31 +++++++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 34 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 939a612bc0..55e23c1436 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,36 @@ # Changelog +## 2.23.0 + +### Various fixes & improvements + +- build(deps): bump actions/create-github-app-token from 1.11.5 to 1.11.6 (#4113) by @dependabot +- build(deps): bump codecov/codecov-action from 5.3.1 to 5.4.0 (#4112) by @dependabot +- docs(baggage): Document that caller must check `mutable` (#4010) by @szokeasaurusrex +- Add `init()` parameters to ApiDocs. (#4100) by @antonpirker +- feat(logs): Add alpha version of Sentry logs (#4126) by @colin-sentry +- Updating Readme (#4134) by @antonpirker +- fix(debug): Take into account parent handlers for debug logger (#4133) by @sentrivana +- fix(quart): Support `quart_flask_patch` (#4132) by @sentrivana +- tests: Add concurrency testcase for arq (#4125) by @sentrivana +- fix(bottle): Prevent internal error on 404 (#4131) by @sentrivana +- Coerce None values into strings in logentry params. (#4121) by @antonpirker +- A way to locally run AWS Lambda functions (#4128) by @antonpirker +- fix(pyspark): Grab `attemptId` more defensively (#4130) by @sentrivana +- Improve asyncio integration error handling. (#4129) by @antonpirker +- Run AWS Lambda tests locally (#3988) by @antonpirker +- Added timeout to HTTP requests in CloudResourceContextIntegration (#4120) by @antonpirker +- Fix FastAPI/Starlette middleware with positional arguments. (#4118) by @antonpirker +- fix(typing): Set correct type for set_context everywhere (#4123) by @sentrivana +- chore(tests): Regenerate tox.ini (#4108) by @sentrivana +- Fixed bug when `cron_jobs` is set to `None` in arq integration (#4115) by @antonpirker +- feat(tracing): Backfill missing `sample_rand` on `PropagationContext` (#4038) by @szokeasaurusrex +- fix(asgi): Fix KeyError if transaction does not exist (#4095) by @kevinji +- security(gha): fix potential for shell injection (#4099) by @mdtro +- ref(tracing): Move `TRANSACTION_SOURCE_*` constants to `Enum` (#3889) by @mgaligniana + +_Plus 12 more_ + ## 2.22.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 0928eea74f..223097b514 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.22.0" +release = "2.23.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index e617581b9e..af811a59ec 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -965,4 +965,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.22.0" +VERSION = "2.23.0" diff --git a/setup.py b/setup.py index 675f5bb1bc..6bbbb77749 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.22.0", + version="2.23.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From c5352c70270f517c3b17f235d52cf2586a719fdb Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 17 Mar 2025 11:02:18 +0100 Subject: [PATCH 185/591] Updated changelog --- CHANGELOG.md | 59 +++++++++++++++++++++++++++++----------------------- 1 file changed, 33 insertions(+), 26 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 55e23c1436..c516461c70 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,32 +4,39 @@ ### Various fixes & improvements -- build(deps): bump actions/create-github-app-token from 1.11.5 to 1.11.6 (#4113) by @dependabot -- build(deps): bump codecov/codecov-action from 5.3.1 to 5.4.0 (#4112) by @dependabot -- docs(baggage): Document that caller must check `mutable` (#4010) by @szokeasaurusrex -- Add `init()` parameters to ApiDocs. (#4100) by @antonpirker -- feat(logs): Add alpha version of Sentry logs (#4126) by @colin-sentry -- Updating Readme (#4134) by @antonpirker -- fix(debug): Take into account parent handlers for debug logger (#4133) by @sentrivana -- fix(quart): Support `quart_flask_patch` (#4132) by @sentrivana -- tests: Add concurrency testcase for arq (#4125) by @sentrivana -- fix(bottle): Prevent internal error on 404 (#4131) by @sentrivana -- Coerce None values into strings in logentry params. (#4121) by @antonpirker -- A way to locally run AWS Lambda functions (#4128) by @antonpirker -- fix(pyspark): Grab `attemptId` more defensively (#4130) by @sentrivana -- Improve asyncio integration error handling. (#4129) by @antonpirker -- Run AWS Lambda tests locally (#3988) by @antonpirker -- Added timeout to HTTP requests in CloudResourceContextIntegration (#4120) by @antonpirker -- Fix FastAPI/Starlette middleware with positional arguments. (#4118) by @antonpirker -- fix(typing): Set correct type for set_context everywhere (#4123) by @sentrivana -- chore(tests): Regenerate tox.ini (#4108) by @sentrivana -- Fixed bug when `cron_jobs` is set to `None` in arq integration (#4115) by @antonpirker -- feat(tracing): Backfill missing `sample_rand` on `PropagationContext` (#4038) by @szokeasaurusrex -- fix(asgi): Fix KeyError if transaction does not exist (#4095) by @kevinji -- security(gha): fix potential for shell injection (#4099) by @mdtro -- ref(tracing): Move `TRANSACTION_SOURCE_*` constants to `Enum` (#3889) by @mgaligniana - -_Plus 12 more_ +- Feat(profiling): Add new functions to start/stop continuous profiler (#4056) by @Zylphrex +- Feat(profiling): Export start/stop profile session (#4079) by @Zylphrex +- Feat(tracing): Backfill missing `sample_rand` on `PropagationContext` (#4038) by @szokeasaurusrex +- Feat(logs): Add alpha version of Sentry logs (#4126) by @colin-sentry +- Security(gha): fix potential for shell injection (#4099) by @mdtro +- Docs: Add `init()` parameters to ApiDocs. (#4100) by @antonpirker +- Docs: Document that caller must check `mutable` (#4010) by @szokeasaurusrex +- Fix(Anthropic): Add partial json support to streams (#3674) +- Fix(ASGI): Fix KeyError if transaction does not exist (#4095) by @kevinji +- Fix(asyncio): Improve asyncio integration error handling. (#4129) by @antonpirker +- Fix(AWS Lambda): Fix capturing errors during AWS Lambda INIT phase (#3943) +- Fix(Bottle): Prevent internal error on 404 (#4131) by @sentrivana +- Fix(CI): Fix API doc failure in CI (#4075) by @sentrivana +- Fix(ClickHouse) ClickHouse in test suite (#4087) by @antonpirker +- Fix(cloudresourcecontext): Added timeout to HTTP requests in CloudResourceContextIntegration (#4120) by @antonpirker +- Fix(crons): Fixed bug when `cron_jobs` is set to `None` in arq integration (#4115) by @antonpirker +- Fix(debug): Take into account parent handlers for debug logger (#4133) by @sentrivana +- Fix(FastAPI/Starlette): Fix middleware with positional arguments. (#4118) by @antonpirker +- Fix(featureflags): add LRU update/dedupe test coverage (#4082) +- Fix(logging): Coerce None values into strings in logentry params. (#4121) by @antonpirker +- Fix(pyspark): Grab `attemptId` more defensively (#4130) by @sentrivana +- Fix(Quart): Support `quart_flask_patch` (#4132) by @sentrivana +- Fix(tests): A way to locally run AWS Lambda functions (#4128) by @antonpirker +- Fix(tests): Add concurrency testcase for arq (#4125) by @sentrivana +- Fix(tests): Add fail_on_changes to toxgen by @sentrivana +- Fix(tests): Run AWS Lambda tests locally (#3988) by @antonpirker +- Fix(tests): Test relevant prereleases and allow to ignore releases +- Fix(tracing): Move `TRANSACTION_SOURCE_*` constants to `Enum` (#3889) by @mgaligniana +- Fix(typing): Add more typing info to Scope.update_from_kwargs's "contexts" (#4080) +- Fix(typing): Set correct type for `set_context` everywhere (#4123) by @sentrivana +- Chore(tests): Regenerate tox.ini (#4108) by @sentrivana +- Build(deps): bump actions/create-github-app-token from 1.11.5 to 1.11.6 (#4113) by @dependabot +- Build(deps): bump codecov/codecov-action from 5.3.1 to 5.4.0 (#4112) by @dependabot ## 2.22.0 From 08d231961a6d6d4374bc66110ae09ef183062fda Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 17 Mar 2025 13:28:55 +0100 Subject: [PATCH 186/591] Fix import problem in release 2.23.0 (#4140) Fixes #4139 --- sentry_sdk/__init__.py | 2 +- tests/test_import.py | 7 +++++++ 2 files changed, 8 insertions(+), 1 deletion(-) create mode 100644 tests/test_import.py diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index 4a0d551e5a..e7e069e377 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -45,7 +45,7 @@ "start_transaction", "trace", "monitor", - "_experimental_logger.py", + "_experimental_logger", ] # Initialize the debug support after everything is loaded diff --git a/tests/test_import.py b/tests/test_import.py new file mode 100644 index 0000000000..e5b07817cb --- /dev/null +++ b/tests/test_import.py @@ -0,0 +1,7 @@ +# As long as this file can be imported, we are good. +from sentry_sdk import * # noqa: F403, F401 + + +def test_import(): + # As long as this file can be imported, we are good. + assert True From 7a82725ce5a8e1b915f4809050ac1a9615dbc072 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 17 Mar 2025 12:29:51 +0000 Subject: [PATCH 187/591] release: 2.23.1 --- CHANGELOG.md | 6 ++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 9 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index c516461c70..2bf4da0e29 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,11 @@ # Changelog +## 2.23.1 + +### Various fixes & improvements + +- Fix import problem in release 2.23.0 (#4140) by @antonpirker + ## 2.23.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 223097b514..9408338941 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.23.0" +release = "2.23.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index af811a59ec..a24903e0ff 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -965,4 +965,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.23.0" +VERSION = "2.23.1" diff --git a/setup.py b/setup.py index 6bbbb77749..a134913fe4 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.23.0", + version="2.23.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From e85715a0ca19e586f567e79c52f6ed62b5099d3d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 18 Mar 2025 16:07:17 +0100 Subject: [PATCH 188/591] Support Starlette/FastAPI `app.host` (#4157) In Starlette/FastAPI you're able to create subapps. When using `transaction_style="url"` in our integration, this would throw an exception because we try to access `route.path` to determine the transaction name, but `Host` routes have no `path` attribute. Closes https://github.com/getsentry/sentry-python/issues/2631 --- sentry_sdk/integrations/starlette.py | 6 +++- tests/integrations/fastapi/test_fastapi.py | 35 ++++++++++++++++++++++ 2 files changed, 40 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index deb05059d5..dbb47dff58 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -693,7 +693,11 @@ def _transaction_name_from_router(scope): for route in router.routes: match = route.matches(scope) if match[0] == Match.FULL: - return route.path + try: + return route.path + except AttributeError: + # routes added via app.host() won't have a path attribute + return scope.get("path") return None diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index f1c0a69305..4cb9ea1716 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -682,3 +682,38 @@ async def _error(): client.get("/error") assert len(events) == int(expected_error) + + +@pytest.mark.parametrize("transaction_style", ["endpoint", "url"]) +def test_app_host(sentry_init, capture_events, transaction_style): + sentry_init( + traces_sample_rate=1.0, + integrations=[ + StarletteIntegration(transaction_style=transaction_style), + FastApiIntegration(transaction_style=transaction_style), + ], + ) + + app = FastAPI() + subapp = FastAPI() + + @subapp.get("/subapp") + async def subapp_route(): + return {"message": "Hello world!"} + + app.host("subapp", subapp) + + events = capture_events() + + client = TestClient(app) + client.get("/subapp", headers={"Host": "subapp"}) + + assert len(events) == 1 + + (event,) = events + assert "transaction" in event + + if transaction_style == "url": + assert event["transaction"] == "/subapp" + else: + assert event["transaction"].endswith("subapp_route") From bc54a1dbc63240a41ee40e6a20b8a6b2e9e52fa2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 18 Mar 2025 16:08:24 +0100 Subject: [PATCH 189/591] feat(tests): Update tox.ini (#4146) Regular `tox.ini` update --- tox.ini | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/tox.ini b/tox.ini index 2294fcc00b..40cbf74475 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-03-10T11:46:25.287445+00:00 +# Last generated: 2025-03-18T10:29:17.585636+00:00 [tox] requires = @@ -187,12 +187,13 @@ envlist = {py3.6,py3.7}-sqlalchemy-v1.3.9 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 {py3.7,py3.10,py3.11}-sqlalchemy-v2.0.9 - {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.38 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.39 # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 + {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.0 @@ -222,15 +223,14 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.227.7 {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.262.1 + {py3.9,py3.12,py3.13}-strawberry-v0.262.5 # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.44.0 {py3.7,py3.10,py3.11}-grpc-v1.58.3 - {py3.8,py3.12,py3.13}-grpc-v1.70.0 - {py3.9,py3.12,py3.13}-grpc-v1.71.0rc2 + {py3.9,py3.12,py3.13}-grpc-v1.71.0 # ~~~ Tasks ~~~ @@ -294,7 +294,7 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.9 - {py3.8,py3.11,py3.12}-trytond-v7.4.7 + {py3.8,py3.11,py3.12}-trytond-v7.4.8 {py3.7,py3.12,py3.13}-typer-v0.15.2 @@ -578,12 +578,13 @@ deps = sqlalchemy-v1.3.9: sqlalchemy==1.3.9 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 sqlalchemy-v2.0.9: sqlalchemy==2.0.9 - sqlalchemy-v2.0.38: sqlalchemy==2.0.38 + sqlalchemy-v2.0.39: sqlalchemy==2.0.39 # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 + launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.0: openfeature-sdk==0.8.0 @@ -622,7 +623,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.262.1: strawberry-graphql[fastapi,flask]==0.262.1 + strawberry-v0.262.5: strawberry-graphql[fastapi,flask]==0.262.5 strawberry: httpx @@ -630,8 +631,7 @@ deps = grpc-v1.32.0: grpcio==1.32.0 grpc-v1.44.0: grpcio==1.44.0 grpc-v1.58.3: grpcio==1.58.3 - grpc-v1.70.0: grpcio==1.70.0 - grpc-v1.71.0rc2: grpcio==1.71.0rc2 + grpc-v1.71.0: grpcio==1.71.0 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -729,7 +729,7 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.9: trytond==7.0.9 - trytond-v7.4.7: trytond==7.4.7 + trytond-v7.4.8: trytond==7.4.8 trytond: werkzeug trytond-v4.6.9: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 11abdd2dba162a44cf4e2d4357752aae69f7ab04 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 19 Mar 2025 08:48:25 +0100 Subject: [PATCH 190/591] Handle loguru msg levels that are not supported by Sentry (#4147) Loguru has two message levels `TRACE` and `SUCCESS` that are not available in Sentry breadcrumbs. This PR maps `TRACE` to `debug` and `SUCCESS` to `info` in Sentry so those breadcrumbs do not show a confusing error message in the Sentry UI. Fixes #2759 --- sentry_sdk/integrations/loguru.py | 36 ++++++++++++++++++++++-- tests/integrations/loguru/test_loguru.py | 23 +++++++-------- 2 files changed, 45 insertions(+), 14 deletions(-) diff --git a/sentry_sdk/integrations/loguru.py b/sentry_sdk/integrations/loguru.py index da99dfc4d6..5b76ea812a 100644 --- a/sentry_sdk/integrations/loguru.py +++ b/sentry_sdk/integrations/loguru.py @@ -11,7 +11,7 @@ if TYPE_CHECKING: from logging import LogRecord - from typing import Optional, Tuple + from typing import Optional, Tuple, Any try: import loguru @@ -31,6 +31,16 @@ class LoggingLevels(enum.IntEnum): CRITICAL = 50 +SENTRY_LEVEL_FROM_LOGURU_LEVEL = { + "TRACE": "DEBUG", + "DEBUG": "DEBUG", + "INFO": "INFO", + "SUCCESS": "INFO", + "WARNING": "WARNING", + "ERROR": "ERROR", + "CRITICAL": "CRITICAL", +} + DEFAULT_LEVEL = LoggingLevels.INFO.value DEFAULT_EVENT_LEVEL = LoggingLevels.ERROR.value # We need to save the handlers to be able to remove them later @@ -87,14 +97,34 @@ class _LoguruBaseHandler(_BaseHandler): def _logging_to_event_level(self, record): # type: (LogRecord) -> str try: - return LoggingLevels(record.levelno).name.lower() - except ValueError: + return SENTRY_LEVEL_FROM_LOGURU_LEVEL[ + LoggingLevels(record.levelno).name + ].lower() + except (ValueError, KeyError): return record.levelname.lower() if record.levelname else "" class LoguruEventHandler(_LoguruBaseHandler, EventHandler): """Modified version of :class:`sentry_sdk.integrations.logging.EventHandler` to use loguru's level names.""" + def __init__(self, *args, **kwargs): + # type: (*Any, **Any) -> None + if kwargs.get("level"): + kwargs["level"] = SENTRY_LEVEL_FROM_LOGURU_LEVEL.get( + kwargs.get("level", ""), DEFAULT_LEVEL + ) + + super().__init__(*args, **kwargs) + class LoguruBreadcrumbHandler(_LoguruBaseHandler, BreadcrumbHandler): """Modified version of :class:`sentry_sdk.integrations.logging.BreadcrumbHandler` to use loguru's level names.""" + + def __init__(self, *args, **kwargs): + # type: (*Any, **Any) -> None + if kwargs.get("level"): + kwargs["level"] = SENTRY_LEVEL_FROM_LOGURU_LEVEL.get( + kwargs.get("level", ""), DEFAULT_LEVEL + ) + + super().__init__(*args, **kwargs) diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 6030108de1..64e9f22ba5 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -8,18 +8,18 @@ @pytest.mark.parametrize( - "level,created_event", + "level,created_event,expected_sentry_level", [ # None - no breadcrumb # False - no event # True - event created - (LoggingLevels.TRACE, None), - (LoggingLevels.DEBUG, None), - (LoggingLevels.INFO, False), - (LoggingLevels.SUCCESS, False), - (LoggingLevels.WARNING, False), - (LoggingLevels.ERROR, True), - (LoggingLevels.CRITICAL, True), + (LoggingLevels.TRACE, None, "debug"), + (LoggingLevels.DEBUG, None, "debug"), + (LoggingLevels.INFO, False, "info"), + (LoggingLevels.SUCCESS, False, "info"), + (LoggingLevels.WARNING, False, "warning"), + (LoggingLevels.ERROR, True, "error"), + (LoggingLevels.CRITICAL, True, "critical"), ], ) @pytest.mark.parametrize("disable_breadcrumbs", [True, False]) @@ -29,6 +29,7 @@ def test_just_log( capture_events, level, created_event, + expected_sentry_level, disable_breadcrumbs, disable_events, ): @@ -48,7 +49,7 @@ def test_just_log( formatted_message = ( " | " + "{:9}".format(level.name.upper()) - + "| tests.integrations.loguru.test_loguru:test_just_log:46 - test" + + "| tests.integrations.loguru.test_loguru:test_just_log:47 - test" ) if not created_event: @@ -59,7 +60,7 @@ def test_just_log( not disable_breadcrumbs and created_event is not None ): # not None == not TRACE or DEBUG level (breadcrumb,) = breadcrumbs - assert breadcrumb["level"] == level.name.lower() + assert breadcrumb["level"] == expected_sentry_level assert breadcrumb["category"] == "tests.integrations.loguru.test_loguru" assert breadcrumb["message"][23:] == formatted_message else: @@ -72,7 +73,7 @@ def test_just_log( return (event,) = events - assert event["level"] == (level.name.lower()) + assert event["level"] == expected_sentry_level assert event["logger"] == "tests.integrations.loguru.test_loguru" assert event["logentry"]["message"][23:] == formatted_message From 65132ba2e878edf9734fb90d08ea15d000bb934c Mon Sep 17 00:00:00 2001 From: Simone Locci Date: Wed, 19 Mar 2025 11:05:26 +0100 Subject: [PATCH 191/591] style(integrations): Fix captured typo (#4161) Small typo fix --- sentry_sdk/integrations/logging.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 28809de4ab..3777381b83 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -232,10 +232,10 @@ def _emit(self, record): event["logger"] = record.name # Log records from `warnings` module as separate issues - record_caputured_from_warnings_module = ( + record_captured_from_warnings_module = ( record.name == "py.warnings" and record.msg == "%s" ) - if record_caputured_from_warnings_module: + if record_captured_from_warnings_module: # use the actual message and not "%s" as the message # this prevents grouping all warnings under one "%s" issue msg = record.args[0] # type: ignore From 0d3bc3df0f4db5adb1028236d41e951fae17b7e5 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 19 Mar 2025 12:12:59 +0100 Subject: [PATCH 192/591] Reset `DedupeIntegration`'s `last-seen` if `before_send` dropped the event (#4142) Imagine an app throws an exception twice, from different places. The first exception is dropped in the user's `before_send`. The second exception is not. Should the second exception appear in Sentry? The current state is that it won't, since `DedupeIntegration` will take the first, dropped exception into account. When encountering the second exception, it'll consider it a duplicate and will drop it, even though the first exception never made it to Sentry. In this PR, we reset `DedupeIntegration`'s `last-seen` if an event has been dropped by `before_send`, ensuring that the next exception will be reported. Closes https://github.com/getsentry/sentry-python/issues/371 --------- Co-authored-by: Anton Pirker --- sentry_sdk/client.py | 9 +++++++++ sentry_sdk/integrations/dedupe.py | 9 +++++++++ tests/test_basics.py | 31 +++++++++++++++++++++++++++++++ 3 files changed, 49 insertions(+) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 5bbf919c02..0f97394561 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -37,6 +37,7 @@ ClientConstructor, ) from sentry_sdk.integrations import _DEFAULT_INTEGRATIONS, setup_integrations +from sentry_sdk.integrations.dedupe import DedupeIntegration from sentry_sdk.sessions import SessionFlusher from sentry_sdk.envelope import Envelope from sentry_sdk.profiler.continuous_profiler import setup_continuous_profiler @@ -606,6 +607,14 @@ def _prepare_event( self.transport.record_lost_event( "before_send", data_category="error" ) + + # If this is an exception, reset the DedupeIntegration. It still + # remembers the dropped exception as the last exception, meaning + # that if the same exception happens again and is not dropped + # in before_send, it'd get dropped by DedupeIntegration. + if event.get("exception"): + DedupeIntegration.reset_last_seen() + event = new_event before_send_transaction = self.options["before_send_transaction"] diff --git a/sentry_sdk/integrations/dedupe.py b/sentry_sdk/integrations/dedupe.py index be6d9311a3..a115e35292 100644 --- a/sentry_sdk/integrations/dedupe.py +++ b/sentry_sdk/integrations/dedupe.py @@ -40,3 +40,12 @@ def processor(event, hint): return None integration._last_seen.set(exc) return event + + @staticmethod + def reset_last_seen(): + # type: () -> None + integration = sentry_sdk.get_client().get_integration(DedupeIntegration) + if integration is None: + return + + integration._last_seen.set(None) diff --git a/tests/test_basics.py b/tests/test_basics.py index ad20bb9fd5..d1c3bce2be 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -710,6 +710,37 @@ def test_dedupe_event_processor_drop_records_client_report( assert lost_event_call == ("event_processor", "error", None, 1) +def test_dedupe_doesnt_take_into_account_dropped_exception(sentry_init, capture_events): + # Two exceptions happen one after another. The first one is dropped in the + # user's before_send. The second one isn't. + # Originally, DedupeIntegration would drop the second exception. This test + # is making sure that that is no longer the case -- i.e., DedupeIntegration + # doesn't consider exceptions dropped in before_send. + count = 0 + + def before_send(event, hint): + nonlocal count + count += 1 + if count == 1: + return None + return event + + sentry_init(before_send=before_send) + events = capture_events() + + exc = ValueError("aha!") + for _ in range(2): + # The first ValueError will be dropped by before_send. The second + # ValueError will be accepted by before_send, and should be sent to + # Sentry. + try: + raise exc + except Exception: + capture_exception() + + assert len(events) == 1 + + def test_event_processor_drop_records_client_report( sentry_init, capture_events, capture_record_lost_event_calls ): From f6db98104c1a8aa002bd2ef31a1447e5c79df675 Mon Sep 17 00:00:00 2001 From: viglia Date: Wed, 19 Mar 2025 14:01:40 +0100 Subject: [PATCH 193/591] feat(profiling): reverse profile_session start/stop methods deprecation (#4162) Revert back to using `start_profiler` and `stop_profiler` function names and deprecate the `*_session` ones instead. Prior PR that introduced the change we're undoing: https://github.com/getsentry/sentry-python/pull/4056 --- sentry_sdk/profiler/__init__.py | 8 ++++---- sentry_sdk/profiler/continuous_profiler.py | 20 ++++++++++---------- 2 files changed, 14 insertions(+), 14 deletions(-) diff --git a/sentry_sdk/profiler/__init__.py b/sentry_sdk/profiler/__init__.py index d8d4e076d5..0bc63e3a6d 100644 --- a/sentry_sdk/profiler/__init__.py +++ b/sentry_sdk/profiler/__init__.py @@ -25,10 +25,10 @@ ) __all__ = [ - "start_profile_session", - "start_profiler", # TODO: Deprecate this in favor of `start_profile_session` - "stop_profile_session", - "stop_profiler", # TODO: Deprecate this in favor of `stop_profile_session` + "start_profile_session", # TODO: Deprecate this in favor of `start_profiler` + "start_profiler", + "stop_profile_session", # TODO: Deprecate this in favor of `stop_profiler` + "stop_profiler", # DEPRECATED: The following was re-exported for backwards compatibility. It # will be removed from sentry_sdk.profiler in a future release. "MAX_PROFILE_DURATION_NS", diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 9e2aa35fc1..47f63d8f59 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -145,32 +145,32 @@ def try_profile_lifecycle_trace_start(): def start_profiler(): # type: () -> None + if _scheduler is None: + return - # TODO: deprecate this as it'll be replaced by `start_profile_session` - start_profile_session() + _scheduler.manual_start() def start_profile_session(): # type: () -> None - if _scheduler is None: - return - _scheduler.manual_start() + # TODO: deprecate this as it'll be replaced by `start_profiler` + start_profiler() def stop_profiler(): # type: () -> None + if _scheduler is None: + return - # TODO: deprecate this as it'll be replaced by `stop_profile_session` - stop_profile_session() + _scheduler.manual_stop() def stop_profile_session(): # type: () -> None - if _scheduler is None: - return - _scheduler.manual_stop() + # TODO: deprecate this as it'll be replaced by `stop_profiler` + stop_profiler() def teardown_continuous_profiler(): From eb189effda67f6ba06f092cb993847ebf0e7347c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 20 Mar 2025 11:37:25 +0100 Subject: [PATCH 194/591] chore(profiler): Add deprecation warning for session functions (#4171) We're deprecating the short-lived `start_profile_session` and `stop_profile_session` functions in favor of `start_profiler` and `stop_profiler`, respectively. The functions will be dropped in 3.x, see https://github.com/getsentry/sentry-python/pull/4170 --- sentry_sdk/profiler/continuous_profiler.py | 13 +++++++++++-- 1 file changed, 11 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 47f63d8f59..77ba60dbda 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -5,6 +5,7 @@ import threading import time import uuid +import warnings from collections import deque from datetime import datetime, timezone @@ -154,7 +155,11 @@ def start_profiler(): def start_profile_session(): # type: () -> None - # TODO: deprecate this as it'll be replaced by `start_profiler` + warnings.warn( + "The `start_profile_session` function is deprecated. Please use `start_profile` instead.", + DeprecationWarning, + stacklevel=2, + ) start_profiler() @@ -169,7 +174,11 @@ def stop_profiler(): def stop_profile_session(): # type: () -> None - # TODO: deprecate this as it'll be replaced by `stop_profiler` + warnings.warn( + "The `stop_profile_session` function is deprecated. Please use `stop_profile` instead.", + DeprecationWarning, + stacklevel=2, + ) stop_profiler() From f76528fa612bc19469813f09612b7dcb448c5b63 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 20 Mar 2025 12:12:20 +0100 Subject: [PATCH 195/591] Fixed flaky test (#4165) The URL www.squirrelchasers.com is actually existing, so we should not access it in our tests. Hope this make the test more stable. --- tests/integrations/stdlib/test_httplib.py | 25 ++++++++--------------- 1 file changed, 8 insertions(+), 17 deletions(-) diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 892e07980b..908a22dc6c 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -398,25 +398,16 @@ def test_http_timeout(monkeypatch, sentry_init, capture_envelopes): envelopes = capture_envelopes() - with start_transaction(op="op", name="name"): - try: - conn = HTTPSConnection("www.squirrelchasers.com") - conn.request("GET", "/top-chasers") + with pytest.raises(TimeoutError): + with start_transaction(op="op", name="name"): + conn = HTTPSConnection("www.example.com") + conn.request("GET", "/bla") conn.getresponse() - except Exception: - pass - - items = [ - item - for envelope in envelopes - for item in envelope.items - if item.type == "transaction" - ] - assert len(items) == 1 - - transaction = items[0].payload.json + + (transaction_envelope,) = envelopes + transaction = transaction_envelope.get_transaction_event() assert len(transaction["spans"]) == 1 span = transaction["spans"][0] assert span["op"] == "http.client" - assert span["description"] == "GET https://www.squirrelchasers.com/top-chasers" + assert span["description"] == "GET https://www.example.com/bla" From 2579cb28e24b5a75a7b8b76fb8849539726ae032 Mon Sep 17 00:00:00 2001 From: Emmanuel Ferdman Date: Thu, 20 Mar 2025 15:05:03 +0200 Subject: [PATCH 196/591] Update scripts sources (#4166) # PR Summary Small PR - Commit d4f4130ad9e2c5c24c06c50855aa0b55fa407a11 moved scripts. This PR adjusts sources to changes. Signed-off-by: Emmanuel Ferdman --- CONTRIBUTING.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 085dbd6075..024a374f85 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -182,14 +182,14 @@ You need to have an AWS account and AWS CLI installed and setup. We put together two helper functions that can help you with development: -- `./scripts/aws-deploy-local-layer.sh` +- `./scripts/aws/aws-deploy-local-layer.sh` - This script [scripts/aws-deploy-local-layer.sh](scripts/aws-deploy-local-layer.sh) will take the code you have checked out locally, create a Lambda layer out of it and deploy it to the `eu-central-1` region of your configured AWS account using `aws` CLI. + This script [scripts/aws/aws-deploy-local-layer.sh](scripts/aws/aws-deploy-local-layer.sh) will take the code you have checked out locally, create a Lambda layer out of it and deploy it to the `eu-central-1` region of your configured AWS account using `aws` CLI. The Lambda layer will have the name `SentryPythonServerlessSDK-local-dev` -- `./scripts/aws-attach-layer-to-lambda-function.sh` +- `./scripts/aws/aws-attach-layer-to-lambda-function.sh` - You can use this script [scripts/aws-attach-layer-to-lambda-function.sh](scripts/aws-attach-layer-to-lambda-function.sh) to attach the Lambda layer you just deployed (using the first script) onto one of your existing Lambda functions. You will have to give the name of the Lambda function to attach onto as an argument. (See the script for details.) + You can use this script [scripts/aws/aws-attach-layer-to-lambda-function.sh](scripts/aws/aws-attach-layer-to-lambda-function.sh) to attach the Lambda layer you just deployed (using the first script) onto one of your existing Lambda functions. You will have to give the name of the Lambda function to attach onto as an argument. (See the script for details.) With these two helper scripts it should be easy to rapidly iterate your development on the Lambda layer. From 5715734eac1c5fb4b6ec61ef459080c74fa777b5 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 20 Mar 2025 14:06:10 +0100 Subject: [PATCH 197/591] Fix memory leak by not piling up breadcrumbs forever in Spark workers. (#4167) We now clear all existing breadcrumbs when a job is started. If an error happens in a job, only breadcrumbs created in this job will be shown. Fixes #1245. --- sentry_sdk/integrations/spark/spark_driver.py | 12 +++++++++--- 1 file changed, 9 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/integrations/spark/spark_driver.py b/sentry_sdk/integrations/spark/spark_driver.py index 701ba12d89..fac985357f 100644 --- a/sentry_sdk/integrations/spark/spark_driver.py +++ b/sentry_sdk/integrations/spark/spark_driver.py @@ -31,9 +31,13 @@ def _set_app_properties(): spark_context = SparkContext._active_spark_context if spark_context: - spark_context.setLocalProperty("sentry_app_name", spark_context.appName) spark_context.setLocalProperty( - "sentry_application_id", spark_context.applicationId + "sentry_app_name", + spark_context.appName, + ) + spark_context.setLocalProperty( + "sentry_application_id", + spark_context.applicationId, ) @@ -231,12 +235,14 @@ def _add_breadcrumb( data=None, # type: Optional[dict[str, Any]] ): # type: (...) -> None - sentry_sdk.get_global_scope().add_breadcrumb( + sentry_sdk.get_isolation_scope().add_breadcrumb( level=level, message=message, data=data ) def onJobStart(self, jobStart): # noqa: N802,N803 # type: (Any) -> None + sentry_sdk.get_isolation_scope().clear_breadcrumbs() + message = "Job {} Started".format(jobStart.jobId()) self._add_breadcrumb(level="info", message=message) _set_app_properties() From 12b3ca39ca48dc611207a77c63659b3a93d88445 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 20 Mar 2025 17:31:21 +0100 Subject: [PATCH 198/591] fix(tracing): Fix `InvalidOperation` (#4179) `InvalidOperation` can occur when using tracing if the `Decimal` class's global context has been modified to set the precision below 6. This change fixes this bug by setting a custom context for our `quantize` call. Fixes #4177 --- sentry_sdk/tracing_utils.py | 8 ++++++-- tests/tracing/test_sample_rand.py | 26 ++++++++++++++++++++++++++ 2 files changed, 32 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 6aa4e4882a..ba56695740 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -5,7 +5,7 @@ import sys from collections.abc import Mapping from datetime import timedelta -from decimal import ROUND_DOWN, Decimal +from decimal import ROUND_DOWN, Context, Decimal from functools import wraps from random import Random from urllib.parse import quote, unquote @@ -871,7 +871,11 @@ def _generate_sample_rand( sample_rand = rng.uniform(lower, upper) # Round down to exactly six decimal-digit precision. - return Decimal(sample_rand).quantize(Decimal("0.000001"), rounding=ROUND_DOWN) + # Setting the context is needed to avoid an InvalidOperation exception + # in case the user has changed the default precision. + return Decimal(sample_rand).quantize( + Decimal("0.000001"), rounding=ROUND_DOWN, context=Context(prec=6) + ) def _sample_rand_range(parent_sampled, sample_rate): diff --git a/tests/tracing/test_sample_rand.py b/tests/tracing/test_sample_rand.py index b8f5c042ed..ef277a3dec 100644 --- a/tests/tracing/test_sample_rand.py +++ b/tests/tracing/test_sample_rand.py @@ -1,3 +1,4 @@ +import decimal from unittest import mock import pytest @@ -53,3 +54,28 @@ def test_transaction_uses_incoming_sample_rand( # Transaction event captured if sample_rand < sample_rate, indicating that # sample_rand is used to make the sampling decision. assert len(events) == int(sample_rand < sample_rate) + + +def test_decimal_context(sentry_init, capture_events): + """ + Ensure that having a decimal context with a precision below 6 + does not cause an InvalidOperation exception. + """ + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + old_prec = decimal.getcontext().prec + decimal.getcontext().prec = 2 + + try: + with mock.patch( + "sentry_sdk.tracing_utils.Random.uniform", return_value=0.123456789 + ): + with sentry_sdk.start_transaction() as transaction: + assert ( + transaction.get_baggage().sentry_items["sample_rand"] == "0.123456" + ) + finally: + decimal.getcontext().prec = old_prec + + assert len(events) == 1 From a3356d7808d3f07ce68a9362efb8d226d080310a Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Fri, 21 Mar 2025 08:59:21 +0000 Subject: [PATCH 199/591] release: 2.24.0 --- CHANGELOG.md | 16 ++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 19 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2bf4da0e29..95ae3f3e96 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,21 @@ # Changelog +## 2.24.0 + +### Various fixes & improvements + +- fix(tracing): Fix `InvalidOperation` (#4179) by @szokeasaurusrex +- Fix memory leak by not piling up breadcrumbs forever in Spark workers. (#4167) by @antonpirker +- Update scripts sources (#4166) by @emmanuel-ferdman +- Fixed flaky test (#4165) by @antonpirker +- chore(profiler): Add deprecation warning for session functions (#4171) by @sentrivana +- feat(profiling): reverse profile_session start/stop methods deprecation (#4162) by @viglia +- Reset `DedupeIntegration`'s `last-seen` if `before_send` dropped the event (#4142) by @sentrivana +- style(integrations): Fix captured typo (#4161) by @pimuzzo +- Handle loguru msg levels that are not supported by Sentry (#4147) by @antonpirker +- feat(tests): Update tox.ini (#4146) by @sentrivana +- Support Starlette/FastAPI `app.host` (#4157) by @sentrivana + ## 2.23.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 9408338941..38772762e1 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.23.1" +release = "2.24.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index a24903e0ff..d20badf9ed 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -965,4 +965,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.23.1" +VERSION = "2.24.0" diff --git a/setup.py b/setup.py index a134913fe4..9c33703ac8 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.23.1", + version="2.24.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From c295047b8540e9da8d0eccecf7c927922af92525 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 21 Mar 2025 10:30:35 +0100 Subject: [PATCH 200/591] meta: Add CODEOWNERS (#4182) Ref #4183 --- .github/CODEOWNERS | 1 + 1 file changed, 1 insertion(+) create mode 100644 .github/CODEOWNERS diff --git a/.github/CODEOWNERS b/.github/CODEOWNERS new file mode 100644 index 0000000000..1dc1a4882f --- /dev/null +++ b/.github/CODEOWNERS @@ -0,0 +1 @@ +* @getsentry/owners-python-sdk From 8ad0d012eeee457b5683d4e32b339a4b39d4dd4e Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 21 Mar 2025 11:04:27 +0100 Subject: [PATCH 201/591] ci: Move `mypy` config into `pyproject.toml` (#4181) First step to consolidate configuration into `pyproject.toml`. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- mypy.ini | 84 ------------------------------- pyproject.toml | 134 ++++++++++++++++++++++++++++++++++++++++++++++++- 2 files changed, 133 insertions(+), 85 deletions(-) delete mode 100644 mypy.ini diff --git a/mypy.ini b/mypy.ini deleted file mode 100644 index 63fa7f334f..0000000000 --- a/mypy.ini +++ /dev/null @@ -1,84 +0,0 @@ -[mypy] -python_version = 3.11 -allow_redefinition = True -check_untyped_defs = True -; disallow_any_decorated = True -; disallow_any_explicit = True -; disallow_any_expr = True -disallow_any_generics = True -; disallow_any_unimported = True -disallow_incomplete_defs = True -disallow_subclassing_any = True -; disallow_untyped_calls = True -disallow_untyped_decorators = True -disallow_untyped_defs = True -no_implicit_optional = True -strict_equality = True -strict_optional = True -warn_redundant_casts = True -; warn_return_any = True -warn_unused_configs = True -warn_unused_ignores = True - - -; Relaxations for code written before mypy was introduced -; -; Do not use wildcards in module paths, otherwise added modules will -; automatically have the same set of relaxed rules as the rest -[mypy-cohere.*] -ignore_missing_imports = True -[mypy-django.*] -ignore_missing_imports = True -[mypy-pyramid.*] -ignore_missing_imports = True -[mypy-psycopg2.*] -ignore_missing_imports = True -[mypy-pytest.*] -ignore_missing_imports = True -[mypy-aiohttp.*] -ignore_missing_imports = True -[mypy-anthropic.*] -ignore_missing_imports = True -[mypy-sanic.*] -ignore_missing_imports = True -[mypy-tornado.*] -ignore_missing_imports = True -[mypy-fakeredis.*] -ignore_missing_imports = True -[mypy-rq.*] -ignore_missing_imports = True -[mypy-pyspark.*] -ignore_missing_imports = True -[mypy-asgiref.*] -ignore_missing_imports = True -[mypy-langchain_core.*] -ignore_missing_imports = True -[mypy-executing.*] -ignore_missing_imports = True -[mypy-asttokens.*] -ignore_missing_imports = True -[mypy-pure_eval.*] -ignore_missing_imports = True -[mypy-blinker.*] -ignore_missing_imports = True -[mypy-sentry_sdk._queue] -ignore_missing_imports = True -disallow_untyped_defs = False -[mypy-sentry_sdk._lru_cache] -disallow_untyped_defs = False -[mypy-celery.app.trace] -ignore_missing_imports = True -[mypy-flask.signals] -ignore_missing_imports = True -[mypy-huey.*] -ignore_missing_imports = True -[mypy-openai.*] -ignore_missing_imports = True -[mypy-openfeature.*] -ignore_missing_imports = True -[mypy-huggingface_hub.*] -ignore_missing_imports = True -[mypy-arq.*] -ignore_missing_imports = True -[mypy-grpc.*] -ignore_missing_imports = True diff --git a/pyproject.toml b/pyproject.toml index 7823c17a7e..37d3a35151 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -20,4 +20,136 @@ omit = [ [tool.coverage.report] exclude_also = [ "if TYPE_CHECKING:", -] \ No newline at end of file +] + +[tool.mypy] +allow_redefinition = true +check_untyped_defs = true +disallow_any_generics = true +disallow_incomplete_defs = true +disallow_subclassing_any = true +disallow_untyped_decorators = true +disallow_untyped_defs = true +no_implicit_optional = true +python_version = "3.11" +strict_equality = true +strict_optional = true +warn_redundant_casts = true +warn_unused_configs = true +warn_unused_ignores = true + +# Relaxations for code written before mypy was introduced +# Do not use wildcards in module paths, otherwise added modules will +# automatically have the same set of relaxed rules as the rest +[[tool.mypy.overrides]] +module = "cohere.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "django.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "pyramid.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "psycopg2.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "pytest.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "aiohttp.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "anthropic.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "sanic.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "tornado.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "fakeredis.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "rq.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "pyspark.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "asgiref.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "langchain_core.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "executing.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "asttokens.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "pure_eval.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "blinker.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "sentry_sdk._queue" +ignore_missing_imports = true +disallow_untyped_defs = false + +[[tool.mypy.overrides]] +module = "sentry_sdk._lru_cache" +disallow_untyped_defs = false + +[[tool.mypy.overrides]] +module = "celery.app.trace" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "flask.signals" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "huey.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "openai.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "openfeature.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "huggingface_hub.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "arq.*" +ignore_missing_imports = true + +[[tool.mypy.overrides]] +module = "grpc.*" +ignore_missing_imports = true From ce9d784aa13de38cbabf0764c3db85dcd6dd4763 Mon Sep 17 00:00:00 2001 From: viglia Date: Fri, 21 Mar 2025 11:17:46 +0100 Subject: [PATCH 202/591] feat(profiling): add platform header to the chunk item-type in the envelope (#4178) We need to send the platform as part of the headers in the chunk item-type as this is the header that relay is checking to manage rate limiting. --- sentry_sdk/envelope.py | 6 +++++- tests/profiler/test_continuous_profiler.py | 21 +++++++++++++-------- 2 files changed, 18 insertions(+), 9 deletions(-) diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index 5f61e689c5..044d282005 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -79,7 +79,11 @@ def add_profile_chunk( ): # type: (...) -> None self.add_item( - Item(payload=PayloadRef(json=profile_chunk), type="profile_chunk") + Item( + payload=PayloadRef(json=profile_chunk), + type="profile_chunk", + headers={"platform": profile_chunk.get("platform", "python")}, + ) ) def add_checkin( diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 78335d7b87..991f8bda5d 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -141,6 +141,11 @@ def assert_single_transaction_with_profile_chunks( if max_chunks is not None: assert len(items["profile_chunk"]) <= max_chunks + for chunk_item in items["profile_chunk"]: + chunk = chunk_item.payload.json + headers = chunk_item.headers + assert chunk["platform"] == headers["platform"] + transaction = items["transaction"][0].payload.json trace_context = transaction["contexts"]["trace"] @@ -215,12 +220,12 @@ def assert_single_transaction_without_profile_chunks(envelopes): pytest.param( start_profile_session, stop_profile_session, - id="start_profile_session/stop_profile_session", + id="start_profile_session/stop_profile_session (deprecated)", ), pytest.param( start_profiler, stop_profiler, - id="start_profiler/stop_profiler (deprecated)", + id="start_profiler/stop_profiler", ), ], ) @@ -292,12 +297,12 @@ def test_continuous_profiler_auto_start_and_manual_stop( pytest.param( start_profile_session, stop_profile_session, - id="start_profile_session/stop_profile_session", + id="start_profile_session/stop_profile_session (deprecated)", ), pytest.param( start_profiler, stop_profiler, - id="start_profiler/stop_profiler (deprecated)", + id="start_profiler/stop_profiler", ), ], ) @@ -374,12 +379,12 @@ def test_continuous_profiler_manual_start_and_stop_sampled( pytest.param( start_profile_session, stop_profile_session, - id="start_profile_session/stop_profile_session", + id="start_profile_session/stop_profile_session (deprecated)", ), pytest.param( start_profiler, stop_profiler, - id="start_profiler/stop_profiler (deprecated)", + id="start_profiler/stop_profiler", ), ], ) @@ -544,12 +549,12 @@ def test_continuous_profiler_auto_start_and_stop_unsampled( pytest.param( start_profile_session, stop_profile_session, - id="start_profile_session/stop_profile_session", + id="start_profile_session/stop_profile_session (deprecated)", ), pytest.param( start_profiler, stop_profiler, - id="start_profiler/stop_profiler (deprecated)", + id="start_profiler/stop_profiler", ), ], ) From aefa34d878b9729bd4261fd5bc74201c65417214 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 21 Mar 2025 11:23:32 +0100 Subject: [PATCH 203/591] ci: Move `pytest` config into `pyproject.toml` (#4184) Consolidate configuration into `pyproject.toml`. --- pyproject.toml | 12 ++++++++++++ pytest.ini | 12 ------------ requirements-devenv.txt | 3 ++- requirements-testing.txt | 3 ++- 4 files changed, 16 insertions(+), 14 deletions(-) delete mode 100644 pytest.ini diff --git a/pyproject.toml b/pyproject.toml index 37d3a35151..25d9b84860 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -22,6 +22,18 @@ exclude_also = [ "if TYPE_CHECKING:", ] +[tool.pytest.ini_options] +addopts = "-vvv -rfEs -s --durations=5 --cov=./sentry_sdk --cov-branch --cov-report= --tb=short --junitxml=.junitxml" +asyncio_mode = "strict" +asyncio_default_fixture_loop_scope = "function" +markers = [ + "tests_internal_exceptions: Handle internal exceptions just as the SDK does, to test it. (Otherwise internal exceptions are recorded and reraised.)", +] + +[tool.pytest-watch] +verbose = true +nobeep = true + [tool.mypy] allow_redefinition = true check_untyped_defs = true diff --git a/pytest.ini b/pytest.ini deleted file mode 100644 index 7edd6127b9..0000000000 --- a/pytest.ini +++ /dev/null @@ -1,12 +0,0 @@ -[pytest] -addopts = -vvv -rfEs -s --durations=5 --cov=./sentry_sdk --cov-branch --cov-report= --tb=short --junitxml=.junitxml -asyncio_mode = strict -asyncio_default_fixture_loop_scope = function -markers = - tests_internal_exceptions: Handle internal exceptions just as the SDK does, to test it. (Otherwise internal exceptions are recorded and reraised.) - -[pytest-watch] -verbose = True -nobeep = True -; Enable this to drop into pdb on errors -; pdb = True diff --git a/requirements-devenv.txt b/requirements-devenv.txt index c0fa5cf245..e5be6c7d77 100644 --- a/requirements-devenv.txt +++ b/requirements-devenv.txt @@ -1,5 +1,6 @@ -r requirements-linting.txt -r requirements-testing.txt mockupdb # required by `pymongo` tests that are enabled by `pymongo` from linter requirements -pytest +pytest>=6.0.0 +tomli;python_version<"3.11" # Only needed for pytest on Python < 3.11 pytest-asyncio diff --git a/requirements-testing.txt b/requirements-testing.txt index cbc515eec2..221863f4ab 100644 --- a/requirements-testing.txt +++ b/requirements-testing.txt @@ -1,5 +1,6 @@ pip -pytest +pytest>=6.0.0 +tomli;python_version<"3.11" # Only needed for pytest on Python < 3.11 pytest-cov pytest-forked pytest-localserver From f8ec5723338d822ff9808cb3d813826b5a23fc64 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 21 Mar 2025 14:56:48 +0100 Subject: [PATCH 204/591] ci: Move `flake8` config into `pyproject.toml` (#4185) Consolidate configuration into `pyproject.toml`. --- .flake8 | 21 ------------------ pyproject.toml | 47 +++++++++++++++++++++++++++++++++++++++- requirements-linting.txt | 7 +++--- 3 files changed, 50 insertions(+), 25 deletions(-) delete mode 100644 .flake8 diff --git a/.flake8 b/.flake8 deleted file mode 100644 index 8610e09241..0000000000 --- a/.flake8 +++ /dev/null @@ -1,21 +0,0 @@ -[flake8] -extend-ignore = - # Handled by black (Whitespace before ':' -- handled by black) - E203, - # Handled by black (Line too long) - E501, - # Sometimes not possible due to execution order (Module level import is not at top of file) - E402, - # I don't care (Do not assign a lambda expression, use a def) - E731, - # does not apply to Python 2 (redundant exception types by flake8-bugbear) - B014, - # I don't care (Lowercase imported as non-lowercase by pep8-naming) - N812, - # is a worse version of and conflicts with B902 (first argument of a classmethod should be named cls) - N804, -extend-exclude=checkouts,lol* -exclude = - # gRCP generated files - grpc_test_service_pb2.py - grpc_test_service_pb2_grpc.py diff --git a/pyproject.toml b/pyproject.toml index 25d9b84860..5e16b30793 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,3 +1,7 @@ +# +# Tool: Black +# + [tool.black] # 'extend-exclude' excludes files or directories in addition to the defaults extend-exclude = ''' @@ -9,6 +13,11 @@ extend-exclude = ''' ) ''' + +# +# Tool: Coverage +# + [tool.coverage.run] branch = true omit = [ @@ -22,6 +31,10 @@ exclude_also = [ "if TYPE_CHECKING:", ] +# +# Tool: Pytest +# + [tool.pytest.ini_options] addopts = "-vvv -rfEs -s --durations=5 --cov=./sentry_sdk --cov-branch --cov-report= --tb=short --junitxml=.junitxml" asyncio_mode = "strict" @@ -34,6 +47,10 @@ markers = [ verbose = true nobeep = true +# +# Tool: Mypy +# + [tool.mypy] allow_redefinition = true check_untyped_defs = true @@ -43,7 +60,7 @@ disallow_subclassing_any = true disallow_untyped_decorators = true disallow_untyped_defs = true no_implicit_optional = true -python_version = "3.11" +python_version = "3.11" strict_equality = true strict_optional = true warn_redundant_casts = true @@ -165,3 +182,31 @@ ignore_missing_imports = true [[tool.mypy.overrides]] module = "grpc.*" ignore_missing_imports = true + +# +# Tool: Flake8 +# + +[tool.flake8] +extend-ignore = [ + # Handled by black (Whitespace before ':' -- handled by black) + "E203", + # Handled by black (Line too long) + "E501", + # Sometimes not possible due to execution order (Module level import is not at top of file) + "E402", + # I don't care (Do not assign a lambda expression, use a def) + "E731", + # does not apply to Python 2 (redundant exception types by flake8-bugbear) + "B014", + # I don't care (Lowercase imported as non-lowercase by pep8-naming) + "N812", + # is a worse version of and conflicts with B902 (first argument of a classmethod should be named cls) + "N804", +] +extend-exclude = ["checkouts", "lol*"] +exclude = [ + # gRCP generated files + "grpc_test_service_pb2.py", + "grpc_test_service_pb2_grpc.py", +] diff --git a/requirements-linting.txt b/requirements-linting.txt index 4255685b5e..20db2151d0 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -1,6 +1,9 @@ mypy black -flake8==5.0.4 # flake8 depends on pyflakes>=3.0.0 and this dropped support for Python 2 "# type:" comments +flake8==5.0.4 +flake8-pyproject # Flake8 plugin to support configuration in pyproject.toml +flake8-bugbear # Flake8 plugin +pep8-naming # Flake8 plugin types-certifi types-protobuf types-gevent @@ -11,8 +14,6 @@ types-webob opentelemetry-distro pymongo # There is no separate types module. loguru # There is no separate types module. -flake8-bugbear -pep8-naming pre-commit # local linting httpcore launchdarkly-server-sdk From 4fbcbf05ec7ce2e3f7a644647045de8bec8ab163 Mon Sep 17 00:00:00 2001 From: Orhan Hirsch Date: Mon, 24 Mar 2025 09:51:47 +0100 Subject: [PATCH 205/591] Broader except in django parsed_body (#4189) We are seeing internal errors in the Sentry SDK if `self.request.data` fails. Specifically, it recently failed with `rest_framework.exceptions.UnsupportedMediaType: Unsupported media type "" in request.`. This exception should not prevent sentry from reporting the original error. Similar to a previous fix I made https://github.com/getsentry/sentry-python/pull/4001 --- sentry_sdk/integrations/django/__init__.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/django/__init__.py b/sentry_sdk/integrations/django/__init__.py index a9477d9954..ff67b3e39b 100644 --- a/sentry_sdk/integrations/django/__init__.py +++ b/sentry_sdk/integrations/django/__init__.py @@ -584,7 +584,7 @@ def parsed_body(self): # type: () -> Optional[Dict[str, Any]] try: return self.request.data - except AttributeError: + except Exception: return RequestExtractor.parsed_body(self) From fafe8f6267738daa52a5823bd0adda05417c3fc4 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Mon, 24 Mar 2025 08:58:37 +0000 Subject: [PATCH 206/591] fix: Always set _spotlight_url (#4186) The conditional early exit in `SpotlightMiddleware` may cause attribute access errors when trying to check if `_spotlight_url` is set or not. This patch sets it to `None` explicitly at class level. --- sentry_sdk/spotlight.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index a783b155a1..c2473b77e9 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -82,6 +82,7 @@ def capture_envelope(self, envelope): class SpotlightMiddleware(MiddlewareMixin): # type: ignore[misc] _spotlight_script = None # type: Optional[str] + _spotlight_url = None # type: Optional[str] def __init__(self, get_response): # type: (Self, Callable[..., HttpResponse]) -> None @@ -103,7 +104,7 @@ def __init__(self, get_response): @property def spotlight_script(self): # type: (Self) -> Optional[str] - if self._spotlight_script is None: + if self._spotlight_url is not None and self._spotlight_script is None: try: spotlight_js_url = urllib.parse.urljoin( self._spotlight_url, SPOTLIGHT_JS_ENTRY_PATH @@ -173,7 +174,7 @@ def process_response(self, _request, response): def process_exception(self, _request, exception): # type: (Self, HttpRequest, Exception) -> Optional[HttpResponseServerError] - if not settings.DEBUG: + if not settings.DEBUG or not self._spotlight_url: return None try: From 2d8ae875d940d26c06a45603630c7884e18f5724 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 24 Mar 2025 09:59:03 +0100 Subject: [PATCH 207/591] build(deps): bump actions/create-github-app-token from 1.11.6 to 1.11.7 (#4188) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 1.11.6 to 1.11.7.
Release notes

Sourced from actions/create-github-app-token's releases.

v1.11.7

1.11.7 (2025-03-20)

Bug Fixes

  • deps: bump undici from 5.28.4 to 7.5.0 (#214) (a24b46a)
Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=1.11.6&new-version=1.11.7)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index c1861ce182..86558d1f18 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@21cfef2b496dd8ef5b904c159339626a10ad380e # v1.11.6 + uses: actions/create-github-app-token@af35edadc00be37caa72ed9f3e6d5f7801bfdf09 # v1.11.7 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From 44238c52b8f851f986b6e731c2190c20fca5591d Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 24 Mar 2025 09:20:00 +0000 Subject: [PATCH 208/591] release: 2.24.1 --- CHANGELOG.md | 13 +++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 16 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 95ae3f3e96..23611595a7 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Changelog +## 2.24.1 + +### Various fixes & improvements + +- build(deps): bump actions/create-github-app-token from 1.11.6 to 1.11.7 (#4188) by @dependabot +- fix: Always set _spotlight_url (#4186) by @BYK +- Broader except in django parsed_body (#4189) by @orhanhenrik +- ci: Move `flake8` config into `pyproject.toml` (#4185) by @antonpirker +- ci: Move `pytest` config into `pyproject.toml` (#4184) by @antonpirker +- feat(profiling): add platform header to the chunk item-type in the envelope (#4178) by @viglia +- ci: Move `mypy` config into `pyproject.toml` (#4181) by @antonpirker +- meta: Add CODEOWNERS (#4182) by @sentrivana + ## 2.24.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 38772762e1..1d80de1231 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.24.0" +release = "2.24.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index d20badf9ed..f9317242cd 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -965,4 +965,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.24.0" +VERSION = "2.24.1" diff --git a/setup.py b/setup.py index 9c33703ac8..cfa9a5a8c1 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.24.0", + version="2.24.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From f60cc78cb0130d5c22f7cb9addaf165898d77160 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 24 Mar 2025 10:21:51 +0100 Subject: [PATCH 209/591] Update CHANGELOG.md --- CHANGELOG.md | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 23611595a7..3999e6fe70 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,14 +4,14 @@ ### Various fixes & improvements -- build(deps): bump actions/create-github-app-token from 1.11.6 to 1.11.7 (#4188) by @dependabot -- fix: Always set _spotlight_url (#4186) by @BYK -- Broader except in django parsed_body (#4189) by @orhanhenrik -- ci: Move `flake8` config into `pyproject.toml` (#4185) by @antonpirker -- ci: Move `pytest` config into `pyproject.toml` (#4184) by @antonpirker -- feat(profiling): add platform header to the chunk item-type in the envelope (#4178) by @viglia -- ci: Move `mypy` config into `pyproject.toml` (#4181) by @antonpirker -- meta: Add CODEOWNERS (#4182) by @sentrivana +- Always set `_spotlight_url` (#4186) by @BYK +- Broader except in Django `parsed_body` (#4189) by @orhanhenrik +- Add platform header to the `chunk` item-type in the envelope (#4178) by @viglia +- Move `mypy` config into `pyproject.toml` (#4181) by @antonpirker +- Move `flake8` config into `pyproject.toml` (#4185) by @antonpirker +- Move `pytest` config into `pyproject.toml` (#4184) by @antonpirker +- Bump `actions/create-github-app-token` from `1.11.6` to `1.11.7` (#4188) by @dependabot +- Add `CODEOWNERS` (#4182) by @sentrivana ## 2.24.0 From 08bbe00f34c5c9455ee1e4064785385f8594a984 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 25 Mar 2025 10:00:47 +0100 Subject: [PATCH 210/591] Added flake8 plugings to pre-commit call of flake8 (#4190) --- .pre-commit-config.yaml | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 775167c10f..9787e136bb 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -17,6 +17,12 @@ repos: rev: 5.0.4 hooks: - id: flake8 + additional_dependencies: + [ + flake8-pyproject, + flake8-bugbear, + pep8-naming, + ] # Disabled for now, because it lists a lot of problems. #- repo: https://github.com/pre-commit/mirrors-mypy From 984f29a1e2007eaabd5c46d53e8efc86038de2d9 Mon Sep 17 00:00:00 2001 From: timdrijvers Date: Tue, 25 Mar 2025 15:04:28 +0100 Subject: [PATCH 211/591] fix(integrations/dramatiq): use set_transaction_name (#4175) The Dramatiq integration is using a deprecated method to set the scope's transaction name, use set_transaction_name instead. "Assigning to scope.transaction directly is deprecated: use scope.set_transaction_name() instead." --- sentry_sdk/integrations/dramatiq.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/dramatiq.py b/sentry_sdk/integrations/dramatiq.py index f9ef13e20b..a756b4c669 100644 --- a/sentry_sdk/integrations/dramatiq.py +++ b/sentry_sdk/integrations/dramatiq.py @@ -95,7 +95,7 @@ def before_process_message(self, broker, message): message._scope_manager.__enter__() scope = sentry_sdk.get_current_scope() - scope.transaction = message.actor_name + scope.set_transaction_name(message.actor_name) scope.set_extra("dramatiq_message_id", message.message_id) scope.add_event_processor(_make_message_event_processor(message, integration)) From ce0727f84111e6f5defd8bf377e64524b0f1b2d8 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 26 Mar 2025 10:26:35 +0100 Subject: [PATCH 212/591] Fix flaky test (#4198) There's a test in `test_utils.py` that flakes very often, but only on Python 3.8 and only in CI (locally it's all fine). I've tried a couple of ways to fix it but at this point it's not worth the effort, so just skipping it on 3.8. --- tests/test_utils.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/tests/test_utils.py b/tests/test_utils.py index 6083ad7ad2..b731c3e3ab 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -7,6 +7,7 @@ import pytest import sentry_sdk +from sentry_sdk._compat import PY38 from sentry_sdk.integrations import Integration from sentry_sdk._queue import Queue from sentry_sdk.utils import ( @@ -901,6 +902,7 @@ def target(): assert (main_thread.ident, main_thread.name) == results.get(timeout=1) +@pytest.mark.skipif(PY38, reason="Flakes a lot on 3.8 in CI.") def test_get_current_thread_meta_failed_to_get_main_thread(): results = Queue(maxsize=1) From 7406113dfd012ce35b52e18b7c1e1b711555d5e0 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 26 Mar 2025 10:35:14 +0100 Subject: [PATCH 213/591] chore: Deprecate Scope.user (#4194) The docstring for `Scope.user` says it's deprecated in favor of `Scope.set_user()`, but there is no user-facing warning. Add one so that we can [drop the property](https://github.com/getsentry/sentry-python/pull/4193) in the next major. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/scope.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 6a5e70a6eb..ce6037e6b6 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -794,6 +794,11 @@ def set_transaction_name(self, name, source=None): def user(self, value): # type: (Optional[Dict[str, Any]]) -> None """When set a specific user is bound to the scope. Deprecated in favor of set_user.""" + warnings.warn( + "The `Scope.user` setter is deprecated in favor of `Scope.set_user()`.", + DeprecationWarning, + stacklevel=2, + ) self.set_user(value) def set_user(self, value): From d394ef6c74f9e5ab5b4b0a3f9663c408ec9fcbed Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 26 Mar 2025 11:17:12 +0100 Subject: [PATCH 214/591] tests: Move Litestar under toxgen (#4197) Remove hardcoded Litestar entries from `tox.ini`/`tox.jinja` and let `toxgen` handle it. (the pymongo update was pulled in by rerunning the script) --- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/populate_tox/config.py | 7 ++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 17 -------- tox.ini | 39 +++++++++---------- 5 files changed, 27 insertions(+), 39 deletions(-) diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index a06ad23b32..93e5569489 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index b5da928d80..b0b1a410da 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -69,6 +69,13 @@ "launchdarkly": { "package": "launchdarkly-server-sdk", }, + "litestar": { + "package": "litestar", + "deps": { + "*": ["pytest-asyncio", "python-multipart", "requests", "cryptography"], + "<2.7": ["httpx<0.28"], + }, + }, "loguru": { "package": "loguru", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 544d4bdcb1..8c6be59450 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -73,7 +73,6 @@ "huggingface_hub", "langchain", "langchain_notiktoken", - "litestar", "openai", "openai_notiktoken", "pure_eval", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 5f1a26ac5e..292590299a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -115,12 +115,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # Litestar - {py3.8,py3.11}-litestar-v{2.0} - {py3.8,py3.11,py3.12}-litestar-v{2.6} - {py3.8,py3.11,py3.12}-litestar-v{2.12} - {py3.8,py3.11,py3.12}-litestar-latest - # OpenAI {py3.9,py3.11,py3.12}-openai-v1.0 {py3.9,py3.11,py3.12}-openai-v1.22 @@ -347,17 +341,6 @@ deps = langchain-{latest,notiktoken}: openai>=1.6.1 langchain-latest: tiktoken~=0.6.0 - # Litestar - litestar: pytest-asyncio - litestar: python-multipart - litestar: requests - litestar: cryptography - litestar-v{2.0,2.6}: httpx<0.28 - litestar-v2.0: litestar~=2.0.0 - litestar-v2.6: litestar~=2.6.0 - litestar-v2.12: litestar~=2.12.0 - litestar-latest: litestar - # OpenAI openai: pytest-asyncio openai-v1.0: openai~=1.0.0 diff --git a/tox.ini b/tox.ini index 40cbf74475..7828007990 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-03-18T10:29:17.585636+00:00 +# Last generated: 2025-03-25T13:14:20.133361+00:00 [tox] requires = @@ -115,12 +115,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # Litestar - {py3.8,py3.11}-litestar-v{2.0} - {py3.8,py3.11,py3.12}-litestar-v{2.6} - {py3.8,py3.11,py3.12}-litestar-v{2.12} - {py3.8,py3.11,py3.12}-litestar-latest - # OpenAI {py3.9,py3.11,py3.12}-openai-v1.0 {py3.9,py3.11,py3.12}-openai-v1.22 @@ -178,7 +172,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.11.2 + {py3.9,py3.12,py3.13}-pymongo-v4.11.3 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -271,6 +265,11 @@ envlist = {py3.6,py3.11,py3.12}-falcon-v3.1.3 {py3.8,py3.11,py3.12}-falcon-v4.0.2 + {py3.8,py3.10,py3.11}-litestar-v2.0.1 + {py3.8,py3.11,py3.12}-litestar-v2.5.5 + {py3.8,py3.11,py3.12}-litestar-v2.10.0 + {py3.8,py3.12,py3.13}-litestar-v2.15.1 + {py3.6}-pyramid-v1.8.6 {py3.6,py3.8,py3.9}-pyramid-v1.10.8 {py3.6,py3.10,py3.11}-pyramid-v2.0.2 @@ -464,17 +463,6 @@ deps = langchain-{latest,notiktoken}: openai>=1.6.1 langchain-latest: tiktoken~=0.6.0 - # Litestar - litestar: pytest-asyncio - litestar: python-multipart - litestar: requests - litestar: cryptography - litestar-v{2.0,2.6}: httpx<0.28 - litestar-v2.0: litestar~=2.0.0 - litestar-v2.6: litestar~=2.6.0 - litestar-v2.12: litestar~=2.12.0 - litestar-latest: litestar - # OpenAI openai: pytest-asyncio openai-v1.0: openai~=1.0.0 @@ -568,7 +556,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.11.2: pymongo==4.11.2 + pymongo-v4.11.3: pymongo==4.11.3 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -694,6 +682,17 @@ deps = falcon-v3.1.3: falcon==3.1.3 falcon-v4.0.2: falcon==4.0.2 + litestar-v2.0.1: litestar==2.0.1 + litestar-v2.5.5: litestar==2.5.5 + litestar-v2.10.0: litestar==2.10.0 + litestar-v2.15.1: litestar==2.15.1 + litestar: pytest-asyncio + litestar: python-multipart + litestar: requests + litestar: cryptography + litestar-v2.0.1: httpx<0.28 + litestar-v2.5.5: httpx<0.28 + pyramid-v1.8.6: pyramid==1.8.6 pyramid-v1.10.8: pyramid==1.10.8 pyramid-v2.0.2: pyramid==2.0.2 From 6f49bfb9fe4f4c7b18db668f0bac79d7be917bb3 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 26 Mar 2025 11:26:14 +0100 Subject: [PATCH 215/591] toxgen: Make it clearer which suites can be migrated (#4196) ...also, `cohere` was in the `IGNORE` list twice, apparently. --- scripts/populate_tox/populate_tox.py | 12 ++++++++---- 1 file changed, 8 insertions(+), 4 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 8c6be59450..d1e6cbca71 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -49,22 +49,26 @@ # suites over to this script. Some entries will probably stay forever # as they don't fit the mold (e.g. common, asgi, which don't have a 3rd party # pypi package to install in different versions). + # + # Test suites that will have to remain hardcoded since they don't fit the + # toxgen usecase + "asgi", + "aws_lambda", + "cloud_resource_context", "common", "gevent", "opentelemetry", "potel", + # Integrations that can be migrated -- we should eventually remove all + # of these from the IGNORE list "aiohttp", "anthropic", "arq", - "asgi", "asyncpg", - "aws_lambda", "beam", "boto3", "chalice", "cohere", - "cloud_resource_context", - "cohere", "django", "fastapi", "gcp", From 2f4b0280048d103d95120ad5f802ec39157e3bc8 Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Thu, 27 Mar 2025 04:52:13 -0400 Subject: [PATCH 216/591] feat(logs): Make the `logging` integration send Sentry logs (#4143) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit We have integrations that make the python logger create breadcrumbs and issues. This adds a third handler which creates Sentry logs on `logger.log` statements. Enable the logger with: ```python sentry_sdk.init( ... _experiments={ "enable_sentry_logs": True } ) some_logger = logging.Logger("some-logger") some_logger.info('Finished sending answer! #chunks=%s', chunks) ``` ![Screenshot 2025-03-17 at 4 12 27 PM](https://github.com/user-attachments/assets/0e8dcd46-6361-47c0-8662-389fcb924969) Refs #4150 --------- Co-authored-by: Anton Pirker --- sentry_sdk/_experimental_logger.py | 23 ++++- sentry_sdk/client.py | 57 ++++-------- sentry_sdk/consts.py | 1 + sentry_sdk/integrations/logging.py | 110 +++++++++++++++++++++- tests/test_logs.py | 141 +++++++++++++++++++---------- 5 files changed, 241 insertions(+), 91 deletions(-) diff --git a/sentry_sdk/_experimental_logger.py b/sentry_sdk/_experimental_logger.py index 1f3cd5e443..d28ff69483 100644 --- a/sentry_sdk/_experimental_logger.py +++ b/sentry_sdk/_experimental_logger.py @@ -1,5 +1,6 @@ # NOTE: this is the logger sentry exposes to users, not some generic logger. import functools +import time from typing import Any from sentry_sdk import get_client, get_current_scope @@ -9,7 +10,27 @@ def _capture_log(severity_text, severity_number, template, **kwargs): # type: (str, int, str, **Any) -> None client = get_client() scope = get_current_scope() - client.capture_log(scope, severity_text, severity_number, template, **kwargs) + + attrs = { + "sentry.message.template": template, + } # type: dict[str, str | bool | float | int] + if "attributes" in kwargs: + attrs.update(kwargs.pop("attributes")) + for k, v in kwargs.items(): + attrs[f"sentry.message.parameters.{k}"] = v + + # noinspection PyProtectedMember + client._capture_experimental_log( + scope, + { + "severity_text": severity_text, + "severity_number": severity_number, + "attributes": attrs, + "body": template.format(**kwargs), + "time_unix_nano": time.time_ns(), + "trace_id": None, + }, + ) trace = functools.partial(_capture_log, "trace", 1) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 0f97394561..df6764a508 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -1,6 +1,5 @@ import json import os -import time import uuid import random import socket @@ -210,8 +209,8 @@ def capture_event(self, *args, **kwargs): # type: (*Any, **Any) -> Optional[str] return None - def capture_log(self, scope, severity_text, severity_number, template, **kwargs): - # type: (Scope, str, int, str, **Any) -> None + def _capture_experimental_log(self, scope, log): + # type: (Scope, Log) -> None pass def capture_session(self, *args, **kwargs): @@ -863,47 +862,36 @@ def capture_event( return return_value - def capture_log(self, scope, severity_text, severity_number, template, **kwargs): - # type: (Scope, str, int, str, **Any) -> None + def _capture_experimental_log(self, current_scope, log): + # type: (Scope, Log) -> None logs_enabled = self.options["_experiments"].get("enable_sentry_logs", False) if not logs_enabled: return + isolation_scope = current_scope.get_isolation_scope() headers = { "sent_at": format_timestamp(datetime.now(timezone.utc)), } # type: dict[str, object] - attrs = { - "sentry.message.template": template, - } # type: dict[str, str | bool | float | int] - - kwargs_attributes = kwargs.get("attributes") - if kwargs_attributes is not None: - attrs.update(kwargs_attributes) - environment = self.options.get("environment") - if environment is not None: - attrs["sentry.environment"] = environment + if environment is not None and "sentry.environment" not in log["attributes"]: + log["attributes"]["sentry.environment"] = environment release = self.options.get("release") - if release is not None: - attrs["sentry.release"] = release + if release is not None and "sentry.release" not in log["attributes"]: + log["attributes"]["sentry.release"] = release - span = scope.span - if span is not None: - attrs["sentry.trace.parent_span_id"] = span.span_id + span = current_scope.span + if span is not None and "sentry.trace.parent_span_id" not in log["attributes"]: + log["attributes"]["sentry.trace.parent_span_id"] = span.span_id - for k, v in kwargs.items(): - attrs[f"sentry.message.parameters.{k}"] = v - - log = { - "severity_text": severity_text, - "severity_number": severity_number, - "body": template.format(**kwargs), - "attributes": attrs, - "time_unix_nano": time.time_ns(), - "trace_id": None, - } # type: Log + if log.get("trace_id") is None: + transaction = current_scope.transaction + propagation_context = isolation_scope.get_active_propagation_context() + if transaction is not None: + log["trace_id"] = transaction.trace_id + elif propagation_context is not None: + log["trace_id"] = propagation_context.trace_id # If debug is enabled, log the log to the console debug = self.options.get("debug", False) @@ -917,15 +905,10 @@ def capture_log(self, scope, severity_text, severity_number, template, **kwargs) "fatal": logging.CRITICAL, } logger.log( - severity_text_to_logging_level.get(severity_text, logging.DEBUG), + severity_text_to_logging_level.get(log["severity_text"], logging.DEBUG), f'[Sentry Logs] {log["body"]}', ) - propagation_context = scope.get_active_propagation_context() - if propagation_context is not None: - headers["trace_id"] = propagation_context.trace_id - log["trace_id"] = propagation_context.trace_id - envelope = Envelope(headers=headers) before_emit_log = self.options["_experiments"].get("before_emit_log") diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index f9317242cd..e4f156256a 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -78,6 +78,7 @@ class CompressionAlgo(Enum): Callable[[str, MetricValue, MeasurementUnit, MetricTags], bool] ], "metric_code_locations": Optional[bool], + "enable_sentry_logs": Optional[bool], }, total=False, ) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 3777381b83..2114f4867a 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -1,8 +1,10 @@ +import json import logging from datetime import datetime, timezone from fnmatch import fnmatch import sentry_sdk +from sentry_sdk.client import BaseClient from sentry_sdk.utils import ( to_string, event_from_exception, @@ -11,7 +13,7 @@ ) from sentry_sdk.integrations import Integration -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, Tuple if TYPE_CHECKING: from collections.abc import MutableMapping @@ -61,14 +63,23 @@ def ignore_logger( class LoggingIntegration(Integration): identifier = "logging" - def __init__(self, level=DEFAULT_LEVEL, event_level=DEFAULT_EVENT_LEVEL): - # type: (Optional[int], Optional[int]) -> None + def __init__( + self, + level=DEFAULT_LEVEL, + event_level=DEFAULT_EVENT_LEVEL, + sentry_logs_level=DEFAULT_LEVEL, + ): + # type: (Optional[int], Optional[int], Optional[int]) -> None self._handler = None self._breadcrumb_handler = None + self._sentry_logs_handler = None if level is not None: self._breadcrumb_handler = BreadcrumbHandler(level=level) + if sentry_logs_level is not None: + self._sentry_logs_handler = SentryLogsHandler(level=sentry_logs_level) + if event_level is not None: self._handler = EventHandler(level=event_level) @@ -83,6 +94,12 @@ def _handle_record(self, record): ): self._breadcrumb_handler.handle(record) + if ( + self._sentry_logs_handler is not None + and record.levelno >= self._sentry_logs_handler.level + ): + self._sentry_logs_handler.handle(record) + @staticmethod def setup_once(): # type: () -> None @@ -296,3 +313,90 @@ def _breadcrumb_from_record(self, record): "timestamp": datetime.fromtimestamp(record.created, timezone.utc), "data": self._extra_from_record(record), } + + +def _python_level_to_otel(record_level): + # type: (int) -> Tuple[int, str] + for py_level, otel_severity_number, otel_severity_text in [ + (50, 21, "fatal"), + (40, 17, "error"), + (30, 13, "warn"), + (20, 9, "info"), + (10, 5, "debug"), + (5, 1, "trace"), + ]: + if record_level >= py_level: + return otel_severity_number, otel_severity_text + return 0, "default" + + +class SentryLogsHandler(_BaseHandler): + """ + A logging handler that records Sentry logs for each Python log record. + + Note that you do not have to use this class if the logging integration is enabled, which it is by default. + """ + + def emit(self, record): + # type: (LogRecord) -> Any + with capture_internal_exceptions(): + self.format(record) + if not self._can_record(record): + return + + client = sentry_sdk.get_client() + if not client.is_active(): + return + + if not client.options["_experiments"].get("enable_sentry_logs", False): + return + + SentryLogsHandler._capture_log_from_record(client, record) + + @staticmethod + def _capture_log_from_record(client, record): + # type: (BaseClient, LogRecord) -> None + scope = sentry_sdk.get_current_scope() + otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) + attrs = { + "sentry.message.template": ( + record.msg if isinstance(record.msg, str) else json.dumps(record.msg) + ), + } # type: dict[str, str | bool | float | int] + if record.args is not None: + if isinstance(record.args, tuple): + for i, arg in enumerate(record.args): + attrs[f"sentry.message.parameters.{i}"] = ( + arg if isinstance(arg, str) else json.dumps(arg) + ) + if record.lineno: + attrs["code.line.number"] = record.lineno + if record.pathname: + attrs["code.file.path"] = record.pathname + if record.funcName: + attrs["code.function.name"] = record.funcName + + if record.thread: + attrs["thread.id"] = record.thread + if record.threadName: + attrs["thread.name"] = record.threadName + + if record.process: + attrs["process.pid"] = record.process + if record.processName: + attrs["process.executable.name"] = record.processName + if record.name: + attrs["logger.name"] = record.name + + # noinspection PyProtectedMember + client._capture_experimental_log( + scope, + { + "severity_text": otel_severity_text, + "severity_number": otel_severity_number, + "body": record.message, + "attributes": attrs, + "time_unix_nano": int(record.created * 1e9), + "trace_id": None, + }, + ) diff --git a/tests/test_logs.py b/tests/test_logs.py index 173a4028d6..9527fb9807 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -1,19 +1,28 @@ +import logging import sys +from typing import List, Any from unittest import mock import pytest import sentry_sdk from sentry_sdk import _experimental_logger as sentry_logger - +from sentry_sdk.integrations.logging import LoggingIntegration minimum_python_37 = pytest.mark.skipif( sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" ) +def otel_attributes_to_dict(otel_attrs: List[Any]): + return {item["key"]: item["value"] for item in otel_attrs} + + @minimum_python_37 def test_logs_disabled_by_default(sentry_init, capture_envelopes): sentry_init() + + python_logger = logging.Logger("some-logger") + envelopes = capture_envelopes() sentry_logger.trace("This is a 'trace' log.") @@ -22,6 +31,7 @@ def test_logs_disabled_by_default(sentry_init, capture_envelopes): sentry_logger.warn("This is a 'warn' log...") sentry_logger.error("This is a 'error' log...") sentry_logger.fatal("This is a 'fatal' log...") + python_logger.warning("sad") assert len(envelopes) == 0 @@ -64,14 +74,14 @@ def test_logs_basics(sentry_init, capture_envelopes): @minimum_python_37 def test_logs_before_emit_log(sentry_init, capture_envelopes): def _before_log(record, hint): - assert list(record.keys()) == [ + assert set(record.keys()) == { "severity_text", "severity_number", "body", "attributes", "time_unix_nano", "trace_id", - ] + } if record["severity_text"] in ["fatal", "error"]: return None @@ -123,34 +133,14 @@ def test_logs_attributes(sentry_init, capture_envelopes): log_item = envelopes[0].items[0].payload.json assert log_item["body"]["stringValue"] == "The recorded value was 'some value'" - assert log_item["attributes"][1] == { - "key": "attr_int", - "value": {"intValue": "1"}, - } # TODO: this is strange. - assert log_item["attributes"][2] == { - "key": "attr_float", - "value": {"doubleValue": 2.0}, - } - assert log_item["attributes"][3] == { - "key": "attr_bool", - "value": {"boolValue": True}, - } - assert log_item["attributes"][4] == { - "key": "attr_string", - "value": {"stringValue": "string attribute"}, - } - assert log_item["attributes"][5] == { - "key": "sentry.environment", - "value": {"stringValue": "production"}, - } - assert log_item["attributes"][6] == { - "key": "sentry.release", - "value": {"stringValue": mock.ANY}, - } - assert log_item["attributes"][7] == { - "key": "sentry.message.parameters.my_var", - "value": {"stringValue": "some value"}, - } + attrs = otel_attributes_to_dict(log_item["attributes"]) + assert attrs["attr_int"] == {"intValue": "1"} + assert attrs["attr_float"] == {"doubleValue": 2.0} + assert attrs["attr_bool"] == {"boolValue": True} + assert attrs["attr_string"] == {"stringValue": "string attribute"} + assert attrs["sentry.environment"] == {"stringValue": "production"} + assert attrs["sentry.release"] == {"stringValue": mock.ANY} + assert attrs["sentry.message.parameters.my_var"] == {"stringValue": "some value"} @minimum_python_37 @@ -172,37 +162,33 @@ def test_logs_message_params(sentry_init, capture_envelopes): envelopes[0].items[0].payload.json["body"]["stringValue"] == "The recorded value was '1'" ) - assert envelopes[0].items[0].payload.json["attributes"][-1] == { - "key": "sentry.message.parameters.int_var", - "value": {"intValue": "1"}, - } # TODO: this is strange. + assert otel_attributes_to_dict(envelopes[0].items[0].payload.json["attributes"])[ + "sentry.message.parameters.int_var" + ] == {"intValue": "1"} assert ( envelopes[1].items[0].payload.json["body"]["stringValue"] == "The recorded value was '2.0'" ) - assert envelopes[1].items[0].payload.json["attributes"][-1] == { - "key": "sentry.message.parameters.float_var", - "value": {"doubleValue": 2.0}, - } + assert otel_attributes_to_dict(envelopes[1].items[0].payload.json["attributes"])[ + "sentry.message.parameters.float_var" + ] == {"doubleValue": 2.0} assert ( envelopes[2].items[0].payload.json["body"]["stringValue"] == "The recorded value was 'False'" ) - assert envelopes[2].items[0].payload.json["attributes"][-1] == { - "key": "sentry.message.parameters.bool_var", - "value": {"boolValue": False}, - } + assert otel_attributes_to_dict(envelopes[2].items[0].payload.json["attributes"])[ + "sentry.message.parameters.bool_var" + ] == {"boolValue": False} assert ( envelopes[3].items[0].payload.json["body"]["stringValue"] == "The recorded value was 'some string value'" ) - assert envelopes[3].items[0].payload.json["attributes"][-1] == { - "key": "sentry.message.parameters.string_var", - "value": {"stringValue": "some string value"}, - } + assert otel_attributes_to_dict(envelopes[3].items[0].payload.json["attributes"])[ + "sentry.message.parameters.string_var" + ] == {"stringValue": "some string value"} @minimum_python_37 @@ -235,8 +221,63 @@ def test_logs_tied_to_spans(sentry_init, capture_envelopes): with sentry_sdk.start_span(description="test-span") as span: sentry_logger.warn("This is a log tied to a span") + attrs = otel_attributes_to_dict(envelopes[0].items[0].payload.json["attributes"]) + assert attrs["sentry.trace.parent_span_id"] == {"stringValue": span.span_id} + + +@minimum_python_37 +def test_logger_integration_warning(sentry_init, capture_envelopes): + """ + The python logger module should create 'warn' sentry logs if the flag is on. + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("this is %s a template %s", "1", "2") + log_entry = envelopes[0].items[0].payload.json - assert log_entry["attributes"][-1] == { - "key": "sentry.trace.parent_span_id", - "value": {"stringValue": span.span_id}, + attrs = otel_attributes_to_dict(log_entry["attributes"]) + assert attrs["sentry.message.template"] == { + "stringValue": "this is %s a template %s" } + assert "code.file.path" in attrs + assert "code.line.number" in attrs + assert attrs["logger.name"] == {"stringValue": "test-logger"} + assert attrs["sentry.environment"] == {"stringValue": "production"} + assert attrs["sentry.message.parameters.0"] == {"stringValue": "1"} + assert attrs["sentry.message.parameters.1"] + assert log_entry["severityNumber"] == 13 + assert log_entry["severityText"] == "warn" + + +@minimum_python_37 +def test_logger_integration_debug(sentry_init, capture_envelopes): + """ + The python logger module should not create 'debug' sentry logs if the flag is on by default + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.debug("this is %s a template %s", "1", "2") + + assert len(envelopes) == 0 + + +@minimum_python_37 +def test_no_log_infinite_loop(sentry_init, capture_envelopes): + """ + If 'debug' mode is true, and you set a low log level in the logging integration, there should be no infinite loops. + """ + sentry_init( + _experiments={"enable_sentry_logs": True}, + integrations=[LoggingIntegration(sentry_logs_level=logging.DEBUG)], + debug=True, + ) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.debug("this is %s a template %s", "1", "2") + + assert len(envelopes) == 1 From e432fb46684ad2cd2ec3cc350ec89ab746a741d3 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Fri, 28 Mar 2025 09:59:05 +0100 Subject: [PATCH 217/591] fix: Don't hang when capturing long stacktrace (#4191) Fixes #2764 --------- Co-authored-by: Anton Pirker --- sentry_sdk/_types.py | 11 +++++++---- sentry_sdk/client.py | 2 ++ sentry_sdk/utils.py | 36 ++++++++++++++++++++++++++++++++---- tests/test_basics.py | 44 ++++++++++++++++++++++++++++++++++++++++++++ 4 files changed, 85 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index bc730719d2..22b91b202f 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -47,11 +47,14 @@ def removed_because_raw_data(cls): ) @classmethod - def removed_because_over_size_limit(cls): - # type: () -> AnnotatedValue - """The actual value was removed because the size of the field exceeded the configured maximum size (specified with the max_request_body_size sdk option)""" + def removed_because_over_size_limit(cls, value=""): + # type: (Any) -> AnnotatedValue + """ + The actual value was removed because the size of the field exceeded the configured maximum size, + for example specified with the max_request_body_size sdk option. + """ return AnnotatedValue( - value="", + value=value, metadata={ "rem": [ # Remark [ diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index df6764a508..980e7179d9 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -755,6 +755,8 @@ def _update_session_from_event( if exceptions: errored = True for error in exceptions: + if isinstance(error, AnnotatedValue): + error = error.value or {} mechanism = error.get("mechanism") if isinstance(mechanism, Mapping) and mechanism.get("handled") is False: crashed = True diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 89b2354c52..595bbe0cf3 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -77,6 +77,15 @@ FALSY_ENV_VALUES = frozenset(("false", "f", "n", "no", "off", "0")) TRUTHY_ENV_VALUES = frozenset(("true", "t", "y", "yes", "on", "1")) +MAX_STACK_FRAMES = 2000 +"""Maximum number of stack frames to send to Sentry. + +If we have more than this number of stack frames, we will stop processing +the stacktrace to avoid getting stuck in a long-lasting loop. This value +exceeds the default sys.getrecursionlimit() of 1000, so users will only +be affected by this limit if they have a custom recursion limit. +""" + def env_to_bool(value, *, strict=False): # type: (Any, Optional[bool]) -> bool | None @@ -732,10 +741,23 @@ def single_exception_from_error_tuple( max_value_length=max_value_length, custom_repr=custom_repr, ) - for tb in iter_stacks(tb) + # Process at most MAX_STACK_FRAMES + 1 frames, to avoid hanging on + # processing a super-long stacktrace. + for tb, _ in zip(iter_stacks(tb), range(MAX_STACK_FRAMES + 1)) ] # type: List[Dict[str, Any]] - if frames: + if len(frames) > MAX_STACK_FRAMES: + # If we have more frames than the limit, we remove the stacktrace completely. + # We don't trim the stacktrace here because we have not processed the whole + # thing (see above, we stop at MAX_STACK_FRAMES + 1). Normally, Relay would + # intelligently trim by removing frames in the middle of the stacktrace, but + # since we don't have the whole stacktrace, we can't do that. Instead, we + # drop the entire stacktrace. + exception_value["stacktrace"] = AnnotatedValue.removed_because_over_size_limit( + value=None + ) + + elif frames: if not full_stack: new_frames = frames else: @@ -941,7 +963,7 @@ def to_string(value): def iter_event_stacktraces(event): - # type: (Event) -> Iterator[Dict[str, Any]] + # type: (Event) -> Iterator[Annotated[Dict[str, Any]]] if "stacktrace" in event: yield event["stacktrace"] if "threads" in event: @@ -950,13 +972,16 @@ def iter_event_stacktraces(event): yield thread["stacktrace"] if "exception" in event: for exception in event["exception"].get("values") or (): - if "stacktrace" in exception: + if isinstance(exception, dict) and "stacktrace" in exception: yield exception["stacktrace"] def iter_event_frames(event): # type: (Event) -> Iterator[Dict[str, Any]] for stacktrace in iter_event_stacktraces(event): + if isinstance(stacktrace, AnnotatedValue): + stacktrace = stacktrace.value or {} + for frame in stacktrace.get("frames") or (): yield frame @@ -964,6 +989,9 @@ def iter_event_frames(event): def handle_in_app(event, in_app_exclude=None, in_app_include=None, project_root=None): # type: (Event, Optional[List[str]], Optional[List[str]], Optional[str]) -> Event for stacktrace in iter_event_stacktraces(event): + if isinstance(stacktrace, AnnotatedValue): + stacktrace = stacktrace.value or {} + set_in_app_in_frames( stacktrace.get("frames"), in_app_exclude=in_app_exclude, diff --git a/tests/test_basics.py b/tests/test_basics.py index d1c3bce2be..e16956979a 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -1065,3 +1065,47 @@ def __str__(self): (event,) = events assert event["exception"]["values"][0]["value"] == "aha!\nnote 1\nnote 3" + + +@pytest.mark.skipif( + sys.version_info < (3, 11), + reason="this test appears to cause a segfault on Python < 3.11", +) +def test_stacktrace_big_recursion(sentry_init, capture_events): + """ + Ensure that if the recursion limit is increased, the full stacktrace is not captured, + as it would take too long to process the entire stack trace. + Also, ensure that the capturing does not take too long. + """ + sentry_init() + events = capture_events() + + def recurse(): + recurse() + + old_recursion_limit = sys.getrecursionlimit() + + try: + sys.setrecursionlimit(100_000) + recurse() + except RecursionError as e: + capture_start_time = time.perf_counter_ns() + sentry_sdk.capture_exception(e) + capture_end_time = time.perf_counter_ns() + finally: + sys.setrecursionlimit(old_recursion_limit) + + (event,) = events + + assert event["exception"]["values"][0]["stacktrace"] is None + assert event["_meta"] == { + "exception": { + "values": {"0": {"stacktrace": {"": {"rem": [["!config", "x"]]}}}} + } + } + + # On my machine, it takes about 100-200ms to capture the exception, + # so this limit should be generous enough. + assert ( + capture_end_time - capture_start_time < 10**9 + ), "stacktrace capture took too long, check that frame limit is set correctly" From 3d2f04469050b6469f6454465b9e0f4c6fecbb8a Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 28 Mar 2025 10:10:22 +0100 Subject: [PATCH 218/591] ci: Fix GraphQL failures (#4208) Looks like strawberry is not compatible with the latest pydantic release (2.11.0). Restrict the version of pydantic used in strawberry tests for now. sqlalchemy apparently released a new version which made it in by rerunning toxgen. --- scripts/populate_tox/config.py | 1 + tox.ini | 10 +++++++--- 2 files changed, 8 insertions(+), 3 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index b0b1a410da..3e8f6cf898 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -148,6 +148,7 @@ "package": "strawberry-graphql[fastapi,flask]", "deps": { "*": ["httpx"], + "<=0.262.5": ["pydantic<2.11"], }, }, "tornado": { diff --git a/tox.ini b/tox.ini index 7828007990..f4b25848fc 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-03-25T13:14:20.133361+00:00 +# Last generated: 2025-03-28T08:54:21.617802+00:00 [tox] requires = @@ -181,7 +181,7 @@ envlist = {py3.6,py3.7}-sqlalchemy-v1.3.9 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 {py3.7,py3.10,py3.11}-sqlalchemy-v2.0.9 - {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.39 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.40 # ~~~ Flags ~~~ @@ -566,7 +566,7 @@ deps = sqlalchemy-v1.3.9: sqlalchemy==1.3.9 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 sqlalchemy-v2.0.9: sqlalchemy==2.0.9 - sqlalchemy-v2.0.39: sqlalchemy==2.0.39 + sqlalchemy-v2.0.40: sqlalchemy==2.0.40 # ~~~ Flags ~~~ @@ -613,6 +613,10 @@ deps = strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 strawberry-v0.262.5: strawberry-graphql[fastapi,flask]==0.262.5 strawberry: httpx + strawberry-v0.209.8: pydantic<2.11 + strawberry-v0.227.7: pydantic<2.11 + strawberry-v0.245.0: pydantic<2.11 + strawberry-v0.262.5: pydantic<2.11 # ~~~ Network ~~~ From 4aaadf4f2daee72c7d792f1b82bdb701254ca37b Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 28 Mar 2025 11:18:01 +0100 Subject: [PATCH 219/591] Update Ubuntu in Github test runners (#4204) The runner `ubuntu-20.04` will be removed on April 1st, 2025. --- .github/workflows/test-integrations-ai.yml | 12 ++++++++--- .github/workflows/test-integrations-cloud.yml | 12 ++++++++--- .../workflows/test-integrations-common.yml | 7 +++++-- .github/workflows/test-integrations-dbs.yml | 20 ++++++++++++------- .github/workflows/test-integrations-flags.yml | 7 +++++-- .../workflows/test-integrations-gevent.yml | 7 +++++-- .../workflows/test-integrations-graphql.yml | 7 +++++-- .github/workflows/test-integrations-misc.yml | 7 +++++-- .../workflows/test-integrations-network.yml | 12 ++++++++--- .github/workflows/test-integrations-tasks.yml | 12 ++++++++--- .github/workflows/test-integrations-web-1.yml | 16 ++++++++++----- .github/workflows/test-integrations-web-2.yml | 12 ++++++++--- .../templates/check_required.jinja | 2 +- .../templates/test_group.jinja | 10 ++++++---- .../test_celery_beat_cron_monitoring.py | 4 ++++ 15 files changed, 105 insertions(+), 42 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 2b2e13059b..10171ce196 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -106,10 +109,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -171,7 +177,7 @@ jobs: needs: test-ai-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-ai-pinned.result, 'failure') || contains(needs.test-ai-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 0468518ec6..1d728f3486 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -34,14 +34,17 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: docker: image: docker:dind # Required for Docker network management options: --privileged # Required for Docker-in-Docker operations + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -110,14 +113,17 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: docker: image: docker:dind # Required for Docker network management options: --privileged # Required for Docker-in-Docker operations + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -179,7 +185,7 @@ jobs: needs: test-cloud-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-cloud-pinned.result, 'failure') || contains(needs.test-cloud-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index b1bdc564f3..4fa12607eb 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -83,7 +86,7 @@ jobs: needs: test-common-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-common-pinned.result, 'failure') || contains(needs.test-common-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index ed35630da6..435ec9d7bb 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -34,7 +34,7 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: postgres: image: postgres @@ -50,17 +50,20 @@ jobs: ports: - 5432:5432 env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: localhost + SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true - name: "Setup ClickHouse Server" - uses: getsentry/action-clickhouse-in-ci@v1.5 + uses: getsentry/action-clickhouse-in-ci@v1.6 - name: Setup Test Env run: | pip install "coverage[toml]" tox @@ -130,7 +133,7 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: postgres: image: postgres @@ -146,17 +149,20 @@ jobs: ports: - 5432:5432 env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: localhost + SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true - name: "Setup ClickHouse Server" - uses: getsentry/action-clickhouse-in-ci@v1.5 + uses: getsentry/action-clickhouse-in-ci@v1.6 - name: Setup Test Env run: | pip install "coverage[toml]" tox @@ -219,7 +225,7 @@ jobs: needs: test-dbs-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-dbs-pinned.result, 'failure') || contains(needs.test-dbs-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index d3ec53de62..f2fdfd5473 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -95,7 +98,7 @@ jobs: needs: test-flags-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-flags-pinned.result, 'failure') || contains(needs.test-flags-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index e9c64d568b..eb6aa1297f 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -83,7 +86,7 @@ jobs: needs: test-gevent-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-gevent-pinned.result, 'failure') || contains(needs.test-gevent-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 235e660474..9713f80c25 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -95,7 +98,7 @@ jobs: needs: test-graphql-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-graphql-pinned.result, 'failure') || contains(needs.test-graphql-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 0db363c3c1..607835ee94 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -103,7 +106,7 @@ jobs: needs: test-misc-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-misc-pinned.result, 'failure') || contains(needs.test-misc-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 96ecdbe5ad..b51c7bfb07 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -98,10 +101,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -155,7 +161,7 @@ jobs: needs: test-network-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-network-pinned.result, 'failure') || contains(needs.test-network-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index a5ed395f32..a27c13278f 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -120,10 +123,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -199,7 +205,7 @@ jobs: needs: test-tasks-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-tasks-pinned.result, 'failure') || contains(needs.test-tasks-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 72cc958308..a294301dbc 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -34,7 +34,7 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: postgres: image: postgres @@ -50,12 +50,15 @@ jobs: ports: - 5432:5432 env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: localhost + SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -120,7 +123,7 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] services: postgres: image: postgres @@ -136,12 +139,15 @@ jobs: ports: - 5432:5432 env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: localhost + SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -199,7 +205,7 @@ jobs: needs: test-web_1-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-web_1-pinned.result, 'failure') || contains(needs.test-web_1-pinned.result, 'skipped') diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 93e5569489..3d3d6e7c84 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -34,10 +34,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -126,10 +129,13 @@ jobs: # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] + # Use Docker container only for Python 3.6 + container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} allow-prereleases: true @@ -211,7 +217,7 @@ jobs: needs: test-web_2-pinned # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-web_2-pinned.result, 'failure') || contains(needs.test-web_2-pinned.result, 'skipped') diff --git a/scripts/split_tox_gh_actions/templates/check_required.jinja b/scripts/split_tox_gh_actions/templates/check_required.jinja index ddb47cddf1..a2ca2db26e 100644 --- a/scripts/split_tox_gh_actions/templates/check_required.jinja +++ b/scripts/split_tox_gh_actions/templates/check_required.jinja @@ -5,7 +5,7 @@ {% endif %} # Always run this, even if a dependent job failed if: always() - runs-on: ubuntu-20.04 + runs-on: ubuntu-22.04 steps: - name: Check for failures if: contains(needs.test-{{ lowercase_group }}-pinned.result, 'failure') || contains(needs.test-{{ lowercase_group }}-pinned.result, 'skipped') diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 5ff68e37dc..91849beff4 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -10,7 +10,7 @@ # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-20.04] + os: [ubuntu-22.04] {% if needs_docker %} services: @@ -34,21 +34,23 @@ ports: - 5432:5432 env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: localhost + SENTRY_PYTHON_TEST_POSTGRES_HOST: {% raw %}${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }}{% endraw %} SENTRY_PYTHON_TEST_POSTGRES_USER: postgres SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry {% endif %} - + # Use Docker container only for Python 3.6 + {% raw %}container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }}{% endraw %} steps: - uses: actions/checkout@v4.2.2 - uses: actions/setup-python@v5 + {% raw %}if: ${{ matrix.python-version != '3.6' }}{% endraw %} with: python-version: {% raw %}${{ matrix.python-version }}{% endraw %} allow-prereleases: true {% if needs_clickhouse %} - name: "Setup ClickHouse Server" - uses: getsentry/action-clickhouse-in-ci@v1.5 + uses: getsentry/action-clickhouse-in-ci@v1.6 {% endif %} {% if needs_redis %} diff --git a/tests/integrations/celery/integration_tests/test_celery_beat_cron_monitoring.py b/tests/integrations/celery/integration_tests/test_celery_beat_cron_monitoring.py index 53f2f63215..e7d8197439 100644 --- a/tests/integrations/celery/integration_tests/test_celery_beat_cron_monitoring.py +++ b/tests/integrations/celery/integration_tests/test_celery_beat_cron_monitoring.py @@ -1,4 +1,5 @@ import os +import sys import pytest from celery.contrib.testing.worker import start_worker @@ -52,6 +53,7 @@ def inner(propagate_traces=True, monitor_beat_tasks=False, **kwargs): return inner +@pytest.mark.skipif(sys.version_info < (3, 7), reason="Requires Python 3.7+") @pytest.mark.forked def test_explanation(celery_init, capture_envelopes): """ @@ -90,6 +92,7 @@ def test_task(): assert len(envelopes) >= 0 +@pytest.mark.skipif(sys.version_info < (3, 7), reason="Requires Python 3.7+") @pytest.mark.forked def test_beat_task_crons_success(celery_init, capture_envelopes): app = celery_init( @@ -122,6 +125,7 @@ def test_task(): assert check_in["status"] == "ok" +@pytest.mark.skipif(sys.version_info < (3, 7), reason="Requires Python 3.7+") @pytest.mark.forked def test_beat_task_crons_error(celery_init, capture_envelopes): app = celery_init( From 3b28649994cb27944b96c81706c97cc1d9cc3301 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Fri, 28 Mar 2025 11:05:38 +0000 Subject: [PATCH 220/591] feat: Sample everything 100% w/ Spotlight & no DSN set (#4207) This patch makes Spotlight easier to setup by turning all sampling to 100% when no DSN is set and Spotlight is enabled. I consider this a non-breaking and a safe change as these only apply when no DSN is set so it should have no production or billing implications. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/client.py | 12 +++++++----- 1 file changed, 7 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 980e7179d9..0cdf0f7717 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -417,6 +417,12 @@ def _capture_envelope(envelope): if self.options.get("spotlight"): self.spotlight = setup_spotlight(self.options) + if not self.options["dsn"]: + sample_all = lambda *_args, **_kwargs: 1.0 + self.options["send_default_pii"] = True + self.options["error_sampler"] = sample_all + self.options["traces_sampler"] = sample_all + self.options["profiles_sampler"] = sample_all sdk_name = get_sdk_name(list(self.integrations.keys())) SDK_INFO["name"] = sdk_name @@ -468,11 +474,7 @@ def should_send_default_pii(self): Returns whether the client should send default PII (Personally Identifiable Information) data to Sentry. """ - result = self.options.get("send_default_pii") - if result is None: - result = not self.options["dsn"] and self.spotlight is not None - - return result + return self.options.get("send_default_pii") or False @property def dsn(self): From 8841b1fd72c0018edb48f53b206390ca245d3999 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 31 Mar 2025 08:57:34 +0000 Subject: [PATCH 221/591] release: 2.25.0 --- CHANGELOG.md | 16 ++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 19 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 3999e6fe70..5c96ff7bdc 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,21 @@ # Changelog +## 2.25.0 + +### Various fixes & improvements + +- feat: Sample everything 100% w/ Spotlight & no DSN set (#4207) by @BYK +- Update Ubuntu in Github test runners (#4204) by @antonpirker +- ci: Fix GraphQL failures (#4208) by @sentrivana +- fix: Don't hang when capturing long stacktrace (#4191) by @szokeasaurusrex +- feat(logs): Make the `logging` integration send Sentry logs (#4143) by @colin-sentry +- toxgen: Make it clearer which suites can be migrated (#4196) by @sentrivana +- tests: Move Litestar under toxgen (#4197) by @sentrivana +- chore: Deprecate Scope.user (#4194) by @sentrivana +- Fix flaky test (#4198) by @sentrivana +- fix(integrations/dramatiq): use set_transaction_name (#4175) by @timdrijvers +- Added flake8 plugings to pre-commit call of flake8 (#4190) by @antonpirker + ## 2.24.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 1d80de1231..6a85b141cf 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.24.1" +release = "2.25.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index e4f156256a..6c663b6ff2 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -966,4 +966,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.24.1" +VERSION = "2.25.0" diff --git a/setup.py b/setup.py index cfa9a5a8c1..3e04ced1da 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.24.1", + version="2.25.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 711816b0a828835ae729b84fafd749ef669cf932 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 31 Mar 2025 11:18:54 +0200 Subject: [PATCH 222/591] Updated changelog --- CHANGELOG.md | 48 +++++++++++++++++++++++++++++++++++++++--------- 1 file changed, 39 insertions(+), 9 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5c96ff7bdc..c3da3d3003 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,17 +4,47 @@ ### Various fixes & improvements -- feat: Sample everything 100% w/ Spotlight & no DSN set (#4207) by @BYK -- Update Ubuntu in Github test runners (#4204) by @antonpirker -- ci: Fix GraphQL failures (#4208) by @sentrivana -- fix: Don't hang when capturing long stacktrace (#4191) by @szokeasaurusrex -- feat(logs): Make the `logging` integration send Sentry logs (#4143) by @colin-sentry +- **New Beta Feature** Enable Sentry logs in `logging` Integration (#4143) by @colin-sentry + + You can now send existing log messages to the new Sentry Logs feature. + + For more information see: https://github.com/getsentry/sentry/discussions/86804 + + This is how you can use it (Sentry Logs is in beta right now so the API can still change): + + ```python + import sentry_sdk + from sentry_sdk.integrations.logging import LoggingIntegration + + # Setup Sentry SDK to send log messages with a level of "error" or higher to Sentry. + sentry_sdk.init( + dsn="...", + _experiments={ + "enable_sentry_logs": True + } + integrations=[ + LoggingIntegration(sentry_logs_level="error"), + ] + ) + + # Your existing logging setup + import logging + some_logger = logging.Logger("some-logger") + + some_logger.info('In this example info events will not be sent to Sentry logs. my_value=%s', my_value) + some_logger.error('But error events will be sent to Sentry logs. my_value=%s', my_value) + ``` + +- Spotlight: Sample everything 100% w/ Spotlight & no DSN set (#4207) by @BYK +- Dramatiq: use set_transaction_name (#4175) by @timdrijvers - toxgen: Make it clearer which suites can be migrated (#4196) by @sentrivana -- tests: Move Litestar under toxgen (#4197) by @sentrivana -- chore: Deprecate Scope.user (#4194) by @sentrivana -- Fix flaky test (#4198) by @sentrivana -- fix(integrations/dramatiq): use set_transaction_name (#4175) by @timdrijvers +- Move Litestar under toxgen (#4197) by @sentrivana - Added flake8 plugings to pre-commit call of flake8 (#4190) by @antonpirker +- Deprecate Scope.user (#4194) by @sentrivana +- Fix hanging when capturing long stacktrace (#4191) by @szokeasaurusrex +- Fix GraphQL failures (#4208) by @sentrivana +- Fix flaky test (#4198) by @sentrivana +- Update Ubuntu in Github test runners (#4204) by @antonpirker ## 2.24.1 From fae17b384cb1867d4c02267682e5113c48ffedc0 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 31 Mar 2025 14:04:46 +0200 Subject: [PATCH 223/591] Pin `fakeredis` until `rq` can work with the new version (#4216) This is breaking our test suite right now. The eco system should stabilize in the next couple of days/weeks, then we can remove the pin. --- .github/CODEOWNERS | 2 +- scripts/populate_tox/tox.jinja | 4 ++-- tox.ini | 11 +++++------ 3 files changed, 8 insertions(+), 9 deletions(-) diff --git a/.github/CODEOWNERS b/.github/CODEOWNERS index 1dc1a4882f..e5d24f170c 100644 --- a/.github/CODEOWNERS +++ b/.github/CODEOWNERS @@ -1 +1 @@ -* @getsentry/owners-python-sdk +* @getsentry/team-web-sdk-backend diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 292590299a..1514ff197a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -400,9 +400,9 @@ deps = rq-v{0.6}: fakeredis<1.0 rq-v{0.6}: redis<3.2.2 rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 - rq-v{1.15,1.16}: fakeredis + rq-v{1.15,1.16}: fakeredis<2.28.0 {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-latest: fakeredis + rq-latest: fakeredis<2.28.0 {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 rq-v0.6: rq~=0.6.0 rq-v0.13: rq~=0.13.0 diff --git a/tox.ini b/tox.ini index f4b25848fc..a093b4de00 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-03-28T08:54:21.617802+00:00 +# Last generated: 2025-03-31T10:49:05.789167+00:00 [tox] requires = @@ -217,7 +217,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.227.7 {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.262.5 + {py3.9,py3.12,py3.13}-strawberry-v0.262.6 # ~~~ Network ~~~ @@ -522,9 +522,9 @@ deps = rq-v{0.6}: fakeredis<1.0 rq-v{0.6}: redis<3.2.2 rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 - rq-v{1.15,1.16}: fakeredis + rq-v{1.15,1.16}: fakeredis<2.28.0 {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-latest: fakeredis + rq-latest: fakeredis<2.28.0 {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 rq-v0.6: rq~=0.6.0 rq-v0.13: rq~=0.13.0 @@ -611,12 +611,11 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.262.5: strawberry-graphql[fastapi,flask]==0.262.5 + strawberry-v0.262.6: strawberry-graphql[fastapi,flask]==0.262.6 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.227.7: pydantic<2.11 strawberry-v0.245.0: pydantic<2.11 - strawberry-v0.262.5: pydantic<2.11 # ~~~ Network ~~~ From 4dcd538d086c3646634a00c953d962cf0987bcbd Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 31 Mar 2025 20:41:17 +0200 Subject: [PATCH 224/591] fixed code snippet (#4218) --- CHANGELOG.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index c3da3d3003..e9f27fed3a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -13,6 +13,8 @@ This is how you can use it (Sentry Logs is in beta right now so the API can still change): ```python + import logging + import sentry_sdk from sentry_sdk.integrations.logging import LoggingIntegration @@ -23,12 +25,11 @@ "enable_sentry_logs": True } integrations=[ - LoggingIntegration(sentry_logs_level="error"), + LoggingIntegration(sentry_logs_level=logging.ERROR), ] ) # Your existing logging setup - import logging some_logger = logging.Logger("some-logger") some_logger.info('In this example info events will not be sent to Sentry logs. my_value=%s', my_value) From d0d70a50b1ab3c7a8c2961ffc8e8a3f4524c5ea8 Mon Sep 17 00:00:00 2001 From: Burak Yigit Kaya Date: Tue, 1 Apr 2025 11:33:07 +0300 Subject: [PATCH 225/591] feat: Do not spam sentry_sdk.warnings logger w/ Spotlight (#4219) Sometimes one may have Spotlight turned on in the SDK but not have the sidecar running or reachable. In that case we spam the console with every event as they fail to reach Spotlight. This patch limits the fail warnings to 3: the first 2 are actual errors and the final one is a note about shutting up. --- sentry_sdk/spotlight.py | 15 ++++++++++++--- 1 file changed, 12 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/spotlight.py b/sentry_sdk/spotlight.py index c2473b77e9..4ac427b9c1 100644 --- a/sentry_sdk/spotlight.py +++ b/sentry_sdk/spotlight.py @@ -38,7 +38,7 @@ def __init__(self, url): # type: (str) -> None self.url = url self.http = urllib3.PoolManager() - self.tries = 0 + self.fails = 0 def capture_envelope(self, envelope): # type: (Envelope) -> None @@ -54,9 +54,18 @@ def capture_envelope(self, envelope): }, ) req.close() + self.fails = 0 except Exception as e: - # TODO: Implement buffering and retrying with exponential backoff - sentry_logger.warning(str(e)) + if self.fails < 2: + sentry_logger.warning(str(e)) + self.fails += 1 + elif self.fails == 2: + self.fails += 1 + sentry_logger.warning( + "Looks like Spotlight is not running, will keep trying to send events but will not log errors." + ) + # omitting self.fails += 1 in the `else:` case intentionally + # to avoid overflowing the variable if Spotlight never becomes reachable try: From 2dde2fe4480d8be18799542b4500015b97233189 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 1 Apr 2025 13:10:22 +0000 Subject: [PATCH 226/591] build(deps): bump actions/create-github-app-token from 1.11.7 to 1.12.0 (#4214) --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 86558d1f18..ed8b3e4094 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@af35edadc00be37caa72ed9f3e6d5f7801bfdf09 # v1.11.7 + uses: actions/create-github-app-token@d72941d797fd3113feb6b93fd0dec494b13a2547 # v1.12.0 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From 8b40aa04f9aa6b08d44b036ea31a3a5ca5505470 Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Wed, 2 Apr 2025 08:07:10 -0400 Subject: [PATCH 227/591] fix(ourlogs): Use repr instead of json for message and arguments (#4227) Currently if you do something like ``` python_logger = logging.Logger("test-logger") python_logger.error(Exception("test exc")) ``` It will error, because Exception is not JSON serializable. --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/logging.py | 12 ++++------ tests/test_logs.py | 38 ++++++++++++++++++++++++++++++ 2 files changed, 43 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 2114f4867a..7822608de8 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -1,4 +1,3 @@ -import json import logging from datetime import datetime, timezone from fnmatch import fnmatch @@ -6,6 +5,7 @@ import sentry_sdk from sentry_sdk.client import BaseClient from sentry_sdk.utils import ( + safe_repr, to_string, event_from_exception, current_stacktrace, @@ -358,16 +358,14 @@ def _capture_log_from_record(client, record): # type: (BaseClient, LogRecord) -> None scope = sentry_sdk.get_current_scope() otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) - attrs = { - "sentry.message.template": ( - record.msg if isinstance(record.msg, str) else json.dumps(record.msg) - ), - } # type: dict[str, str | bool | float | int] + attrs = {} # type: dict[str, str | bool | float | int] + if isinstance(record.msg, str): + attrs["sentry.message.template"] = record.msg if record.args is not None: if isinstance(record.args, tuple): for i, arg in enumerate(record.args): attrs[f"sentry.message.parameters.{i}"] = ( - arg if isinstance(arg, str) else json.dumps(arg) + arg if isinstance(arg, str) else safe_repr(arg) ) if record.lineno: attrs["code.line.number"] = record.lineno diff --git a/tests/test_logs.py b/tests/test_logs.py index 9527fb9807..7ef708ceb1 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -281,3 +281,41 @@ def test_no_log_infinite_loop(sentry_init, capture_envelopes): python_logger.debug("this is %s a template %s", "1", "2") assert len(envelopes) == 1 + + +@minimum_python_37 +def test_logging_errors(sentry_init, capture_envelopes): + """ + The python logger module should be able to log errors without erroring + """ + sentry_init(_experiments={"enable_sentry_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.error(Exception("test exc 1")) + python_logger.error("error is %s", Exception("test exc 2")) + + error_event_1 = envelopes[0].items[0].payload.json + assert error_event_1["level"] == "error" + + log_event_1 = envelopes[1].items[0].payload.json + assert log_event_1["severityText"] == "error" + # When only logging an exception, there is no "sentry.message.template" or "sentry.message.parameters.0" + assert len(log_event_1["attributes"]) == 10 + assert log_event_1["attributes"][0]["key"] == "code.line.number" + + error_event_2 = envelopes[2].items[0].payload.json + assert error_event_2["level"] == "error" + + log_event_2 = envelopes[3].items[0].payload.json + assert log_event_2["severityText"] == "error" + assert len(log_event_2["attributes"]) == 12 + assert log_event_2["attributes"][0]["key"] == "sentry.message.template" + assert log_event_2["attributes"][0]["value"] == {"stringValue": "error is %s"} + assert log_event_2["attributes"][1]["key"] == "sentry.message.parameters.0" + assert log_event_2["attributes"][1]["value"] == { + "stringValue": "Exception('test exc 2')" + } + assert log_event_2["attributes"][2]["key"] == "code.line.number" + + assert len(envelopes) == 4 From e4b8dae2b99d92567c42493eb34b56087708e051 Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Wed, 2 Apr 2025 09:25:03 -0400 Subject: [PATCH 228/591] fix(ai): Do not consume anthropic streaming stop (#4232) The old functionality wouldn't re-emit the `stop` message for streaming Anthropic calls. --- sentry_sdk/integrations/anthropic.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 4cb54309c8..76a3bb9f13 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -184,8 +184,7 @@ def new_iterator(): input_tokens, output_tokens, content_blocks = _collect_ai_data( event, input_tokens, output_tokens, content_blocks ) - if event.type != "message_stop": - yield event + yield event _add_ai_data_to_span( span, integration, input_tokens, output_tokens, content_blocks @@ -202,8 +201,7 @@ async def new_iterator_async(): input_tokens, output_tokens, content_blocks = _collect_ai_data( event, input_tokens, output_tokens, content_blocks ) - if event.type != "message_stop": - yield event + yield event _add_ai_data_to_span( span, integration, input_tokens, output_tokens, content_blocks From 438ee01c18cfe7f0a821b6e54844965822547405 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 2 Apr 2025 16:27:36 +0200 Subject: [PATCH 229/591] Debug output from Sentry logs should always be `debug` level. (#4224) Prevent emitting too many log messages. --- sentry_sdk/client.py | 14 ++------------ 1 file changed, 2 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 0cdf0f7717..3b47123e3b 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -3,7 +3,6 @@ import uuid import random import socket -import logging from collections.abc import Mapping from datetime import datetime, timezone from importlib import import_module @@ -900,17 +899,8 @@ def _capture_experimental_log(self, current_scope, log): # If debug is enabled, log the log to the console debug = self.options.get("debug", False) if debug: - severity_text_to_logging_level = { - "trace": logging.DEBUG, - "debug": logging.DEBUG, - "info": logging.INFO, - "warn": logging.WARNING, - "error": logging.ERROR, - "fatal": logging.CRITICAL, - } - logger.log( - severity_text_to_logging_level.get(log["severity_text"], logging.DEBUG), - f'[Sentry Logs] {log["body"]}', + logger.debug( + f'[Sentry Logs] [{log.get("severity_text")}] {log.get("body")}' ) envelope = Envelope(headers=headers) From c254ba4309b2c0dab3b356c2eeab7b555b34797f Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Wed, 2 Apr 2025 10:31:21 -0400 Subject: [PATCH 230/591] feat(ourlogs): Add a class which batches groups of logs together. (#4229) Currently, sentry logs create a new envelope per-log, which is inefficient. This changes the behavior to batch a large chunk of logs to be sent all at once. Fixes https://github.com/getsentry/sentry-python/issues/4155 Fixes https://github.com/getsentry/sentry-python/issues/4225 Fixes https://github.com/getsentry/sentry-python/issues/4152 --------- Co-authored-by: Anton Pirker --- sentry_sdk/__init__.py | 2 +- sentry_sdk/_log_batcher.py | 142 ++++++++ sentry_sdk/client.py | 62 +--- sentry_sdk/consts.py | 2 +- sentry_sdk/integrations/logging.py | 9 +- .../{_experimental_logger.py => logger.py} | 17 +- sentry_sdk/types.py | 5 +- tests/test_logs.py | 342 +++++++++++------- 8 files changed, 397 insertions(+), 184 deletions(-) create mode 100644 sentry_sdk/_log_batcher.py rename sentry_sdk/{_experimental_logger.py => logger.py} (75%) diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index e7e069e377..b4859cc5d2 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -45,7 +45,7 @@ "start_transaction", "trace", "monitor", - "_experimental_logger", + "logger", ] # Initialize the debug support after everything is loaded diff --git a/sentry_sdk/_log_batcher.py b/sentry_sdk/_log_batcher.py new file mode 100644 index 0000000000..77efe29a2c --- /dev/null +++ b/sentry_sdk/_log_batcher.py @@ -0,0 +1,142 @@ +import os +import random +import threading +from datetime import datetime, timezone +from typing import Optional, List, Callable, TYPE_CHECKING, Any + +from sentry_sdk.utils import format_timestamp, safe_repr +from sentry_sdk.envelope import Envelope + +if TYPE_CHECKING: + from sentry_sdk._types import Log + + +class LogBatcher: + MAX_LOGS_BEFORE_FLUSH = 100 + FLUSH_WAIT_TIME = 5.0 + + def __init__( + self, + capture_func, # type: Callable[[Envelope], None] + ): + # type: (...) -> None + self._log_buffer = [] # type: List[Log] + self._capture_func = capture_func + self._running = True + self._lock = threading.Lock() + + self._flush_event = threading.Event() # type: threading.Event + + self._flusher = None # type: Optional[threading.Thread] + self._flusher_pid = None # type: Optional[int] + + def _ensure_thread(self): + # type: (...) -> bool + """For forking processes we might need to restart this thread. + This ensures that our process actually has that thread running. + """ + if not self._running: + return False + + pid = os.getpid() + if self._flusher_pid == pid: + return True + + with self._lock: + # Recheck to make sure another thread didn't get here and start the + # the flusher in the meantime + if self._flusher_pid == pid: + return True + + self._flusher_pid = pid + + self._flusher = threading.Thread(target=self._flush_loop) + self._flusher.daemon = True + + try: + self._flusher.start() + except RuntimeError: + # Unfortunately at this point the interpreter is in a state that no + # longer allows us to spawn a thread and we have to bail. + self._running = False + return False + + return True + + def _flush_loop(self): + # type: (...) -> None + while self._running: + self._flush_event.wait(self.FLUSH_WAIT_TIME + random.random()) + self._flush_event.clear() + self._flush() + + def add( + self, + log, # type: Log + ): + # type: (...) -> None + if not self._ensure_thread() or self._flusher is None: + return None + + with self._lock: + self._log_buffer.append(log) + if len(self._log_buffer) >= self.MAX_LOGS_BEFORE_FLUSH: + self._flush_event.set() + + def kill(self): + # type: (...) -> None + if self._flusher is None: + return + + self._running = False + self._flush_event.set() + self._flusher = None + + def flush(self): + # type: (...) -> None + self._flush() + + @staticmethod + def _log_to_otel(log): + # type: (Log) -> Any + def format_attribute(key, val): + # type: (str, int | float | str | bool) -> Any + if isinstance(val, bool): + return {"key": key, "value": {"boolValue": val}} + if isinstance(val, int): + return {"key": key, "value": {"intValue": str(val)}} + if isinstance(val, float): + return {"key": key, "value": {"doubleValue": val}} + if isinstance(val, str): + return {"key": key, "value": {"stringValue": val}} + return {"key": key, "value": {"stringValue": safe_repr(val)}} + + otel_log = { + "severityText": log["severity_text"], + "severityNumber": log["severity_number"], + "body": {"stringValue": log["body"]}, + "timeUnixNano": str(log["time_unix_nano"]), + "attributes": [ + format_attribute(k, v) for (k, v) in log["attributes"].items() + ], + } + + if "trace_id" in log: + otel_log["traceId"] = log["trace_id"] + + return otel_log + + def _flush(self): + # type: (...) -> Optional[Envelope] + + envelope = Envelope( + headers={"sent_at": format_timestamp(datetime.now(timezone.utc))} + ) + with self._lock: + for log in self._log_buffer: + envelope.add_log(self._log_to_otel(log)) + self._log_buffer.clear() + if envelope.items: + self._capture_func(envelope) + return envelope + return None diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 3b47123e3b..3350c1372a 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -1,4 +1,3 @@ -import json import os import uuid import random @@ -64,6 +63,7 @@ from sentry_sdk.session import Session from sentry_sdk.spotlight import SpotlightClient from sentry_sdk.transport import Transport + from sentry_sdk._log_batcher import LogBatcher I = TypeVar("I", bound=Integration) # noqa: E741 @@ -177,6 +177,7 @@ def __init__(self, options=None): self.transport = None # type: Optional[Transport] self.monitor = None # type: Optional[Monitor] self.metrics_aggregator = None # type: Optional[MetricsAggregator] + self.log_batcher = None # type: Optional[LogBatcher] def __getstate__(self, *args, **kwargs): # type: (*Any, **Any) -> Any @@ -374,6 +375,12 @@ def _capture_envelope(envelope): "Metrics not supported on Python 3.6 and lower with gevent." ) + self.log_batcher = None + if experiments.get("enable_logs", False): + from sentry_sdk._log_batcher import LogBatcher + + self.log_batcher = LogBatcher(capture_func=_capture_envelope) + max_request_body_size = ("always", "never", "small", "medium") if self.options["max_request_body_size"] not in max_request_body_size: raise ValueError( @@ -450,6 +457,7 @@ def _capture_envelope(envelope): if ( self.monitor or self.metrics_aggregator + or self.log_batcher or has_profiling_enabled(self.options) or isinstance(self.transport, BaseHttpTransport) ): @@ -867,15 +875,11 @@ def capture_event( def _capture_experimental_log(self, current_scope, log): # type: (Scope, Log) -> None - logs_enabled = self.options["_experiments"].get("enable_sentry_logs", False) + logs_enabled = self.options["_experiments"].get("enable_logs", False) if not logs_enabled: return isolation_scope = current_scope.get_isolation_scope() - headers = { - "sent_at": format_timestamp(datetime.now(timezone.utc)), - } # type: dict[str, object] - environment = self.options.get("environment") if environment is not None and "sentry.environment" not in log["attributes"]: log["attributes"]["sentry.environment"] = environment @@ -903,46 +907,14 @@ def _capture_experimental_log(self, current_scope, log): f'[Sentry Logs] [{log.get("severity_text")}] {log.get("body")}' ) - envelope = Envelope(headers=headers) - - before_emit_log = self.options["_experiments"].get("before_emit_log") - if before_emit_log is not None: - log = before_emit_log(log, {}) + before_send_log = self.options["_experiments"].get("before_send_log") + if before_send_log is not None: + log = before_send_log(log, {}) if log is None: return - def format_attribute(key, val): - # type: (str, int | float | str | bool) -> Any - if isinstance(val, bool): - return {"key": key, "value": {"boolValue": val}} - if isinstance(val, int): - return {"key": key, "value": {"intValue": str(val)}} - if isinstance(val, float): - return {"key": key, "value": {"doubleValue": val}} - if isinstance(val, str): - return {"key": key, "value": {"stringValue": val}} - return {"key": key, "value": {"stringValue": json.dumps(val)}} - - otel_log = { - "severityText": log["severity_text"], - "severityNumber": log["severity_number"], - "body": {"stringValue": log["body"]}, - "timeUnixNano": str(log["time_unix_nano"]), - "attributes": [ - format_attribute(k, v) for (k, v) in log["attributes"].items() - ], - } - - if "trace_id" in log: - otel_log["traceId"] = log["trace_id"] - - envelope.add_log(otel_log) # TODO: batch these - - if self.spotlight: - self.spotlight.capture_envelope(envelope) - - if self.transport is not None: - self.transport.capture_envelope(envelope) + if self.log_batcher: + self.log_batcher.add(log) def capture_session( self, session # type: Session @@ -996,6 +968,8 @@ def close( self.session_flusher.kill() if self.metrics_aggregator is not None: self.metrics_aggregator.kill() + if self.log_batcher is not None: + self.log_batcher.kill() if self.monitor: self.monitor.kill() self.transport.kill() @@ -1020,6 +994,8 @@ def flush( self.session_flusher.flush() if self.metrics_aggregator is not None: self.metrics_aggregator.flush() + if self.log_batcher is not None: + self.log_batcher.flush() self.transport.flush(timeout=timeout, callback=callback) def __enter__(self): diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 6c663b6ff2..05942b6071 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -78,7 +78,7 @@ class CompressionAlgo(Enum): Callable[[str, MetricValue, MeasurementUnit, MetricTags], bool] ], "metric_code_locations": Optional[bool], - "enable_sentry_logs": Optional[bool], + "enable_logs": Optional[bool], }, total=False, ) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 7822608de8..ba6e6581b7 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -348,7 +348,7 @@ def emit(self, record): if not client.is_active(): return - if not client.options["_experiments"].get("enable_sentry_logs", False): + if not client.options["_experiments"].get("enable_logs", False): return SentryLogsHandler._capture_log_from_record(client, record) @@ -365,7 +365,12 @@ def _capture_log_from_record(client, record): if isinstance(record.args, tuple): for i, arg in enumerate(record.args): attrs[f"sentry.message.parameters.{i}"] = ( - arg if isinstance(arg, str) else safe_repr(arg) + arg + if isinstance(arg, str) + or isinstance(arg, float) + or isinstance(arg, int) + or isinstance(arg, bool) + else safe_repr(arg) ) if record.lineno: attrs["code.line.number"] = record.lineno diff --git a/sentry_sdk/_experimental_logger.py b/sentry_sdk/logger.py similarity index 75% rename from sentry_sdk/_experimental_logger.py rename to sentry_sdk/logger.py index d28ff69483..1fa31b786b 100644 --- a/sentry_sdk/_experimental_logger.py +++ b/sentry_sdk/logger.py @@ -4,6 +4,7 @@ from typing import Any from sentry_sdk import get_client, get_current_scope +from sentry_sdk.utils import safe_repr def _capture_log(severity_text, severity_number, template, **kwargs): @@ -19,6 +20,20 @@ def _capture_log(severity_text, severity_number, template, **kwargs): for k, v in kwargs.items(): attrs[f"sentry.message.parameters.{k}"] = v + attrs = { + k: ( + v + if ( + isinstance(v, str) + or isinstance(v, int) + or isinstance(v, bool) + or isinstance(v, float) + ) + else safe_repr(v) + ) + for (k, v) in attrs.items() + } + # noinspection PyProtectedMember client._capture_experimental_log( scope, @@ -36,6 +51,6 @@ def _capture_log(severity_text, severity_number, template, **kwargs): trace = functools.partial(_capture_log, "trace", 1) debug = functools.partial(_capture_log, "debug", 5) info = functools.partial(_capture_log, "info", 9) -warn = functools.partial(_capture_log, "warn", 13) +warning = functools.partial(_capture_log, "warning", 13) error = functools.partial(_capture_log, "error", 17) fatal = functools.partial(_capture_log, "fatal", 21) diff --git a/sentry_sdk/types.py b/sentry_sdk/types.py index a81be8f1c1..2b9f04c097 100644 --- a/sentry_sdk/types.py +++ b/sentry_sdk/types.py @@ -11,7 +11,7 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from sentry_sdk._types import Event, EventDataCategory, Hint + from sentry_sdk._types import Event, EventDataCategory, Hint, Log else: from typing import Any @@ -20,5 +20,6 @@ Event = Any EventDataCategory = Any Hint = Any + Log = Any -__all__ = ("Event", "EventDataCategory", "Hint") +__all__ = ("Event", "EventDataCategory", "Hint", "Log") diff --git a/tests/test_logs.py b/tests/test_logs.py index 7ef708ceb1..1305f243de 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -1,20 +1,60 @@ +import json import logging import sys -from typing import List, Any -from unittest import mock +import time +from typing import List, Any, Mapping, Union import pytest import sentry_sdk -from sentry_sdk import _experimental_logger as sentry_logger +import sentry_sdk.logger +from sentry_sdk import get_client +from sentry_sdk.envelope import Envelope from sentry_sdk.integrations.logging import LoggingIntegration +from sentry_sdk.types import Log minimum_python_37 = pytest.mark.skipif( sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" ) -def otel_attributes_to_dict(otel_attrs: List[Any]): - return {item["key"]: item["value"] for item in otel_attrs} +def otel_attributes_to_dict(otel_attrs): + # type: (List[Mapping[str, Any]]) -> Mapping[str, Any] + def _convert_attr(attr): + # type: (Mapping[str, Union[str, float, bool]]) -> Any + if "boolValue" in attr: + return bool(attr["boolValue"]) + if "doubleValue" in attr: + return float(attr["doubleValue"]) + if "intValue" in attr: + return int(attr["intValue"]) + if attr["stringValue"].startswith("{"): + try: + return json.loads(attr["stringValue"]) + except ValueError: + pass + return str(attr["stringValue"]) + + return {item["key"]: _convert_attr(item["value"]) for item in otel_attrs} + + +def envelopes_to_logs(envelopes: List[Envelope]) -> List[Log]: + res = [] # type: List[Log] + for envelope in envelopes: + for item in envelope.items: + if item.type == "otel_log": + log_json = item.payload.json + log = { + "severity_text": log_json["severityText"], + "severity_number": log_json["severityNumber"], + "body": log_json["body"]["stringValue"], + "attributes": otel_attributes_to_dict(log_json["attributes"]), + "time_unix_nano": int(log_json["timeUnixNano"]), + "trace_id": None, + } # type: Log + if "traceId" in log_json: + log["trace_id"] = log_json["traceId"] + res.append(log) + return res @minimum_python_37 @@ -25,12 +65,12 @@ def test_logs_disabled_by_default(sentry_init, capture_envelopes): envelopes = capture_envelopes() - sentry_logger.trace("This is a 'trace' log.") - sentry_logger.debug("This is a 'debug' log...") - sentry_logger.info("This is a 'info' log...") - sentry_logger.warn("This is a 'warn' log...") - sentry_logger.error("This is a 'error' log...") - sentry_logger.fatal("This is a 'fatal' log...") + sentry_sdk.logger.trace("This is a 'trace' log.") + sentry_sdk.logger.debug("This is a 'debug' log...") + sentry_sdk.logger.info("This is a 'info' log...") + sentry_sdk.logger.warning("This is a 'warning' log...") + sentry_sdk.logger.error("This is a 'error' log...") + sentry_sdk.logger.fatal("This is a 'fatal' log...") python_logger.warning("sad") assert len(envelopes) == 0 @@ -38,41 +78,41 @@ def test_logs_disabled_by_default(sentry_init, capture_envelopes): @minimum_python_37 def test_logs_basics(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() - sentry_logger.trace("This is a 'trace' log...") - sentry_logger.debug("This is a 'debug' log...") - sentry_logger.info("This is a 'info' log...") - sentry_logger.warn("This is a 'warn' log...") - sentry_logger.error("This is a 'error' log...") - sentry_logger.fatal("This is a 'fatal' log...") + sentry_sdk.logger.trace("This is a 'trace' log...") + sentry_sdk.logger.debug("This is a 'debug' log...") + sentry_sdk.logger.info("This is a 'info' log...") + sentry_sdk.logger.warning("This is a 'warn' log...") + sentry_sdk.logger.error("This is a 'error' log...") + sentry_sdk.logger.fatal("This is a 'fatal' log...") - assert ( - len(envelopes) == 6 - ) # We will batch those log items into a single envelope at some point - - assert envelopes[0].items[0].payload.json["severityText"] == "trace" - assert envelopes[0].items[0].payload.json["severityNumber"] == 1 + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert logs[0].get("severity_text") == "trace" + assert logs[0].get("severity_number") == 1 - assert envelopes[1].items[0].payload.json["severityText"] == "debug" - assert envelopes[1].items[0].payload.json["severityNumber"] == 5 + assert logs[1].get("severity_text") == "debug" + assert logs[1].get("severity_number") == 5 - assert envelopes[2].items[0].payload.json["severityText"] == "info" - assert envelopes[2].items[0].payload.json["severityNumber"] == 9 + assert logs[2].get("severity_text") == "info" + assert logs[2].get("severity_number") == 9 - assert envelopes[3].items[0].payload.json["severityText"] == "warn" - assert envelopes[3].items[0].payload.json["severityNumber"] == 13 + assert logs[3].get("severity_text") == "warning" + assert logs[3].get("severity_number") == 13 - assert envelopes[4].items[0].payload.json["severityText"] == "error" - assert envelopes[4].items[0].payload.json["severityNumber"] == 17 + assert logs[4].get("severity_text") == "error" + assert logs[4].get("severity_number") == 17 - assert envelopes[5].items[0].payload.json["severityText"] == "fatal" - assert envelopes[5].items[0].payload.json["severityNumber"] == 21 + assert logs[5].get("severity_text") == "fatal" + assert logs[5].get("severity_number") == 21 @minimum_python_37 -def test_logs_before_emit_log(sentry_init, capture_envelopes): +def test_logs_before_send_log(sentry_init, capture_envelopes): + before_log_called = [False] + def _before_log(record, hint): assert set(record.keys()) == { "severity_text", @@ -86,29 +126,34 @@ def _before_log(record, hint): if record["severity_text"] in ["fatal", "error"]: return None + before_log_called[0] = True + return record sentry_init( _experiments={ - "enable_sentry_logs": True, - "before_emit_log": _before_log, + "enable_logs": True, + "before_send_log": _before_log, } ) envelopes = capture_envelopes() - sentry_logger.trace("This is a 'trace' log...") - sentry_logger.debug("This is a 'debug' log...") - sentry_logger.info("This is a 'info' log...") - sentry_logger.warn("This is a 'warn' log...") - sentry_logger.error("This is a 'error' log...") - sentry_logger.fatal("This is a 'fatal' log...") + sentry_sdk.logger.trace("This is a 'trace' log...") + sentry_sdk.logger.debug("This is a 'debug' log...") + sentry_sdk.logger.info("This is a 'info' log...") + sentry_sdk.logger.warning("This is a 'warning' log...") + sentry_sdk.logger.error("This is a 'error' log...") + sentry_sdk.logger.fatal("This is a 'fatal' log...") - assert len(envelopes) == 4 + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert len(logs) == 4 - assert envelopes[0].items[0].payload.json["severityText"] == "trace" - assert envelopes[1].items[0].payload.json["severityText"] == "debug" - assert envelopes[2].items[0].payload.json["severityText"] == "info" - assert envelopes[3].items[0].payload.json["severityText"] == "warn" + assert logs[0]["severity_text"] == "trace" + assert logs[1]["severity_text"] == "debug" + assert logs[2]["severity_text"] == "info" + assert logs[3]["severity_text"] == "warning" + assert before_log_called[0] @minimum_python_37 @@ -116,7 +161,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): """ Passing arbitrary attributes to log messages. """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() attrs = { @@ -126,21 +171,19 @@ def test_logs_attributes(sentry_init, capture_envelopes): "attr_string": "string attribute", } - sentry_logger.warn( + sentry_sdk.logger.warning( "The recorded value was '{my_var}'", my_var="some value", attributes=attrs ) - log_item = envelopes[0].items[0].payload.json - assert log_item["body"]["stringValue"] == "The recorded value was 'some value'" + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert logs[0]["body"] == "The recorded value was 'some value'" - attrs = otel_attributes_to_dict(log_item["attributes"]) - assert attrs["attr_int"] == {"intValue": "1"} - assert attrs["attr_float"] == {"doubleValue": 2.0} - assert attrs["attr_bool"] == {"boolValue": True} - assert attrs["attr_string"] == {"stringValue": "string attribute"} - assert attrs["sentry.environment"] == {"stringValue": "production"} - assert attrs["sentry.release"] == {"stringValue": mock.ANY} - assert attrs["sentry.message.parameters.my_var"] == {"stringValue": "some value"} + for k, v in attrs.items(): + assert logs[0]["attributes"][k] == v + assert logs[0]["attributes"]["sentry.environment"] == "production" + assert "sentry.release" in logs[0]["attributes"] + assert logs[0]["attributes"]["sentry.message.parameters.my_var"] == "some value" @minimum_python_37 @@ -148,47 +191,42 @@ def test_logs_message_params(sentry_init, capture_envelopes): """ This is the official way of how to pass vars to log messages. """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() - sentry_logger.warn("The recorded value was '{int_var}'", int_var=1) - sentry_logger.warn("The recorded value was '{float_var}'", float_var=2.0) - sentry_logger.warn("The recorded value was '{bool_var}'", bool_var=False) - sentry_logger.warn( + sentry_sdk.logger.warning("The recorded value was '{int_var}'", int_var=1) + sentry_sdk.logger.warning("The recorded value was '{float_var}'", float_var=2.0) + sentry_sdk.logger.warning("The recorded value was '{bool_var}'", bool_var=False) + sentry_sdk.logger.warning( "The recorded value was '{string_var}'", string_var="some string value" ) - - assert ( - envelopes[0].items[0].payload.json["body"]["stringValue"] - == "The recorded value was '1'" + sentry_sdk.logger.error( + "The recorded error was '{error}'", error=Exception("some error") ) - assert otel_attributes_to_dict(envelopes[0].items[0].payload.json["attributes"])[ - "sentry.message.parameters.int_var" - ] == {"intValue": "1"} - assert ( - envelopes[1].items[0].payload.json["body"]["stringValue"] - == "The recorded value was '2.0'" - ) - assert otel_attributes_to_dict(envelopes[1].items[0].payload.json["attributes"])[ - "sentry.message.parameters.float_var" - ] == {"doubleValue": 2.0} + get_client().flush() + logs = envelopes_to_logs(envelopes) + + assert logs[0]["body"] == "The recorded value was '1'" + assert logs[0]["attributes"]["sentry.message.parameters.int_var"] == 1 + assert logs[1]["body"] == "The recorded value was '2.0'" + assert logs[1]["attributes"]["sentry.message.parameters.float_var"] == 2.0 + + assert logs[2]["body"] == "The recorded value was 'False'" + assert logs[2]["attributes"]["sentry.message.parameters.bool_var"] is False + + assert logs[3]["body"] == "The recorded value was 'some string value'" assert ( - envelopes[2].items[0].payload.json["body"]["stringValue"] - == "The recorded value was 'False'" + logs[3]["attributes"]["sentry.message.parameters.string_var"] + == "some string value" ) - assert otel_attributes_to_dict(envelopes[2].items[0].payload.json["attributes"])[ - "sentry.message.parameters.bool_var" - ] == {"boolValue": False} + assert logs[4]["body"] == "The recorded error was 'some error'" assert ( - envelopes[3].items[0].payload.json["body"]["stringValue"] - == "The recorded value was 'some string value'" + logs[4]["attributes"]["sentry.message.parameters.error"] + == "Exception('some error')" ) - assert otel_attributes_to_dict(envelopes[3].items[0].payload.json["attributes"])[ - "sentry.message.parameters.string_var" - ] == {"stringValue": "some string value"} @minimum_python_37 @@ -196,17 +234,15 @@ def test_logs_tied_to_transactions(sentry_init, capture_envelopes): """ Log messages are also tied to transactions. """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction") as trx: - sentry_logger.warn("This is a log tied to a transaction") + sentry_sdk.logger.warning("This is a log tied to a transaction") - log_entry = envelopes[0].items[0].payload.json - assert log_entry["attributes"][-1] == { - "key": "sentry.trace.parent_span_id", - "value": {"stringValue": trx.span_id}, - } + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert logs[0]["attributes"]["sentry.trace.parent_span_id"] == trx.span_id @minimum_python_37 @@ -214,15 +250,16 @@ def test_logs_tied_to_spans(sentry_init, capture_envelopes): """ Log messages are also tied to spans. """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction"): - with sentry_sdk.start_span(description="test-span") as span: - sentry_logger.warn("This is a log tied to a span") + with sentry_sdk.start_span(name="test-span") as span: + sentry_sdk.logger.warning("This is a log tied to a span") - attrs = otel_attributes_to_dict(envelopes[0].items[0].payload.json["attributes"]) - assert attrs["sentry.trace.parent_span_id"] == {"stringValue": span.span_id} + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert logs[0]["attributes"]["sentry.trace.parent_span_id"] == span.span_id @minimum_python_37 @@ -230,25 +267,24 @@ def test_logger_integration_warning(sentry_init, capture_envelopes): """ The python logger module should create 'warn' sentry logs if the flag is on. """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") python_logger.warning("this is %s a template %s", "1", "2") - log_entry = envelopes[0].items[0].payload.json - attrs = otel_attributes_to_dict(log_entry["attributes"]) - assert attrs["sentry.message.template"] == { - "stringValue": "this is %s a template %s" - } + get_client().flush() + logs = envelopes_to_logs(envelopes) + attrs = logs[0]["attributes"] + assert attrs["sentry.message.template"] == "this is %s a template %s" assert "code.file.path" in attrs assert "code.line.number" in attrs - assert attrs["logger.name"] == {"stringValue": "test-logger"} - assert attrs["sentry.environment"] == {"stringValue": "production"} - assert attrs["sentry.message.parameters.0"] == {"stringValue": "1"} - assert attrs["sentry.message.parameters.1"] - assert log_entry["severityNumber"] == 13 - assert log_entry["severityText"] == "warn" + assert attrs["logger.name"] == "test-logger" + assert attrs["sentry.environment"] == "production" + assert attrs["sentry.message.parameters.0"] == "1" + assert attrs["sentry.message.parameters.1"] == "2" + assert logs[0]["severity_number"] == 13 + assert logs[0]["severity_text"] == "warn" @minimum_python_37 @@ -256,11 +292,12 @@ def test_logger_integration_debug(sentry_init, capture_envelopes): """ The python logger module should not create 'debug' sentry logs if the flag is on by default """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") python_logger.debug("this is %s a template %s", "1", "2") + get_client().flush() assert len(envelopes) == 0 @@ -271,7 +308,7 @@ def test_no_log_infinite_loop(sentry_init, capture_envelopes): If 'debug' mode is true, and you set a low log level in the logging integration, there should be no infinite loops. """ sentry_init( - _experiments={"enable_sentry_logs": True}, + _experiments={"enable_logs": True}, integrations=[LoggingIntegration(sentry_logs_level=logging.DEBUG)], debug=True, ) @@ -279,6 +316,7 @@ def test_no_log_infinite_loop(sentry_init, capture_envelopes): python_logger = logging.Logger("test-logger") python_logger.debug("this is %s a template %s", "1", "2") + get_client().flush() assert len(envelopes) == 1 @@ -288,34 +326,70 @@ def test_logging_errors(sentry_init, capture_envelopes): """ The python logger module should be able to log errors without erroring """ - sentry_init(_experiments={"enable_sentry_logs": True}) + sentry_init(_experiments={"enable_logs": True}) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") python_logger.error(Exception("test exc 1")) python_logger.error("error is %s", Exception("test exc 2")) + get_client().flush() error_event_1 = envelopes[0].items[0].payload.json assert error_event_1["level"] == "error" + error_event_2 = envelopes[1].items[0].payload.json + assert error_event_2["level"] == "error" - log_event_1 = envelopes[1].items[0].payload.json - assert log_event_1["severityText"] == "error" - # When only logging an exception, there is no "sentry.message.template" or "sentry.message.parameters.0" - assert len(log_event_1["attributes"]) == 10 - assert log_event_1["attributes"][0]["key"] == "code.line.number" + print(envelopes) + logs = envelopes_to_logs(envelopes) + assert logs[0]["severity_text"] == "error" + assert "sentry.message.template" not in logs[0]["attributes"] + assert "sentry.message.parameters.0" not in logs[0]["attributes"] + assert "code.line.number" in logs[0]["attributes"] - error_event_2 = envelopes[2].items[0].payload.json - assert error_event_2["level"] == "error" + assert logs[1]["severity_text"] == "error" + assert logs[1]["attributes"]["sentry.message.template"] == "error is %s" + assert ( + logs[1]["attributes"]["sentry.message.parameters.0"] + == "Exception('test exc 2')" + ) + assert "code.line.number" in logs[1]["attributes"] - log_event_2 = envelopes[3].items[0].payload.json - assert log_event_2["severityText"] == "error" - assert len(log_event_2["attributes"]) == 12 - assert log_event_2["attributes"][0]["key"] == "sentry.message.template" - assert log_event_2["attributes"][0]["value"] == {"stringValue": "error is %s"} - assert log_event_2["attributes"][1]["key"] == "sentry.message.parameters.0" - assert log_event_2["attributes"][1]["value"] == { - "stringValue": "Exception('test exc 2')" - } - assert log_event_2["attributes"][2]["key"] == "code.line.number" + assert len(logs) == 2 + + +def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): + """ + If you log >100 logs, it should automatically trigger a flush. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + for i in range(200): + python_logger.warning("log #%d", i) + + for _ in range(500): + time.sleep(1.0 / 100.0) + if len(envelopes) > 0: + return + + raise AssertionError("200 logs were never flushed after five seconds") + + +@minimum_python_37 +def test_auto_flush_logs_after_5s(sentry_init, capture_envelopes): + """ + If you log a single log, it should automatically flush after 5 seconds, at most 10 seconds. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("log #%d", 1) + + for _ in range(100): + time.sleep(1.0 / 10.0) + if len(envelopes) > 0: + return - assert len(envelopes) == 4 + raise AssertionError("1 logs was never flushed after 10 seconds") From d7cf51033025812763cceffc388b58da7123fe50 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Wed, 2 Apr 2025 14:48:04 +0000 Subject: [PATCH 231/591] release: 2.25.1 --- CHANGELOG.md | 12 ++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 15 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index e9f27fed3a..d012353cc7 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,17 @@ # Changelog +## 2.25.1 + +### Various fixes & improvements + +- feat(ourlogs): Add a class which batches groups of logs together. (#4229) by @colin-sentry +- Debug output from Sentry logs should always be `debug` level. (#4224) by @antonpirker +- fix(ai): Do not consume anthropic streaming stop (#4232) by @colin-sentry +- fix(ourlogs): Use repr instead of json for message and arguments (#4227) by @colin-sentry +- build(deps): bump actions/create-github-app-token from 1.11.7 to 1.12.0 (#4214) by @dependabot +- feat: Do not spam sentry_sdk.warnings logger w/ Spotlight (#4219) by @BYK +- fixed code snippet (#4218) by @antonpirker + ## 2.25.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 6a85b141cf..2f575d3097 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.25.0" +release = "2.25.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 05942b6071..c0f6ff66c6 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -966,4 +966,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.25.0" +VERSION = "2.25.1" diff --git a/setup.py b/setup.py index 3e04ced1da..6de160dcfb 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.25.0", + version="2.25.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From d42e63274b38c2e52ac165beea89ac8e43b2f95c Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 2 Apr 2025 16:50:55 +0200 Subject: [PATCH 232/591] Updated changelog --- CHANGELOG.md | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index d012353cc7..a9294eaec1 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,13 +4,13 @@ ### Various fixes & improvements -- feat(ourlogs): Add a class which batches groups of logs together. (#4229) by @colin-sentry -- Debug output from Sentry logs should always be `debug` level. (#4224) by @antonpirker +- fix(logs): Add a class which batches groups of logs together. (#4229) by @colin-sentry +- fix(logs): Use repr instead of json for message and arguments (#4227) by @colin-sentry +- fix(logs): Debug output from Sentry logs should always be `debug` level. (#4224) by @antonpirker - fix(ai): Do not consume anthropic streaming stop (#4232) by @colin-sentry -- fix(ourlogs): Use repr instead of json for message and arguments (#4227) by @colin-sentry +- fix(spotlight): Do not spam sentry_sdk.warnings logger w/ Spotlight (#4219) by @BYK +- fix(docs): fixed code snippet (#4218) by @antonpirker - build(deps): bump actions/create-github-app-token from 1.11.7 to 1.12.0 (#4214) by @dependabot -- feat: Do not spam sentry_sdk.warnings logger w/ Spotlight (#4219) by @BYK -- fixed code snippet (#4218) by @antonpirker ## 2.25.0 From 5f71872c8abf2ee0cd0f4a35e1771f0a097e6938 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 3 Apr 2025 12:38:30 +0200 Subject: [PATCH 233/591] fix(asyncio): Remove shutdown handler (#4237) Remove the shutdown handler from the asyncio integration. It's only purpose was to log a message, but it looks like it has [unintended side effects](https://github.com/getsentry/sentry-python/issues/4234). Closes https://github.com/getsentry/sentry-python/issues/4234 --- sentry_sdk/integrations/asyncio.py | 17 ----------------- 1 file changed, 17 deletions(-) diff --git a/sentry_sdk/integrations/asyncio.py b/sentry_sdk/integrations/asyncio.py index 9326c16e9a..ae580ca038 100644 --- a/sentry_sdk/integrations/asyncio.py +++ b/sentry_sdk/integrations/asyncio.py @@ -1,5 +1,4 @@ import sys -import signal import sentry_sdk from sentry_sdk.consts import OP @@ -37,22 +36,6 @@ def patch_asyncio(): loop = asyncio.get_running_loop() orig_task_factory = loop.get_task_factory() - # Add a shutdown handler to log a helpful message - def shutdown_handler(): - # type: () -> None - logger.info( - "AsyncIO is shutting down. If you see 'Task was destroyed but it is pending!' " - "errors with '_task_with_sentry_span_creation', these are normal during shutdown " - "and not a problem with your code or Sentry." - ) - - try: - loop.add_signal_handler(signal.SIGINT, shutdown_handler) - loop.add_signal_handler(signal.SIGTERM, shutdown_handler) - except (NotImplementedError, AttributeError): - # Signal handlers might not be supported on all platforms - pass - def _sentry_task_factory(loop, coro, **kwargs): # type: (asyncio.AbstractEventLoop, Coroutine[Any, Any, Any], Any) -> asyncio.Future[Any] From 2b3b82d492ece2634e23ffeb2dd589dcce284c10 Mon Sep 17 00:00:00 2001 From: Mahmoodreza <47904885+moodix@users.noreply.github.com> Date: Thu, 3 Apr 2025 17:49:47 +0300 Subject: [PATCH 234/591] fix: Handle JSONDecodeError gracefully in StarletteRequestExtractor (#4226) Previously, when encountering malformed JSON in request bodies, the json() method would raise a JSONDecodeError. This change updates the method to catch the exception and return None instead, providing more consistent behavior and preventing unexpected crashes. Added a test case to verify this error handling behavior. --- sentry_sdk/integrations/starlette.py | 7 ++++-- .../integrations/starlette/test_starlette.py | 25 +++++++++++++++++++ 2 files changed, 30 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index dbb47dff58..d0f0bf2045 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -3,6 +3,7 @@ import warnings from collections.abc import Set from copy import deepcopy +from json import JSONDecodeError import sentry_sdk from sentry_sdk.consts import OP @@ -680,8 +681,10 @@ async def json(self): # type: (StarletteRequestExtractor) -> Optional[Dict[str, Any]] if not self.is_json(): return None - - return await self.request.json() + try: + return await self.request.json() + except JSONDecodeError: + return None def _transaction_name_from_router(scope): diff --git a/tests/integrations/starlette/test_starlette.py b/tests/integrations/starlette/test_starlette.py index 3289f69ed6..bc445bf8f2 100644 --- a/tests/integrations/starlette/test_starlette.py +++ b/tests/integrations/starlette/test_starlette.py @@ -1354,3 +1354,28 @@ async def _error(_): client.get("/error") assert len(events) == int(expected_error) + + +@pytest.mark.asyncio +async def test_starletterequestextractor_malformed_json_error_handling(sentry_init): + scope = SCOPE.copy() + scope["headers"] = [ + [b"content-type", b"application/json"], + ] + starlette_request = starlette.requests.Request(scope) + + malformed_json = "{invalid json" + malformed_messages = [ + {"type": "http.request", "body": malformed_json.encode("utf-8")}, + {"type": "http.disconnect"}, + ] + + side_effect = [_mock_receive(msg) for msg in malformed_messages] + starlette_request._receive = mock.Mock(side_effect=side_effect) + + extractor = StarletteRequestExtractor(starlette_request) + + assert extractor.is_json() + + result = await extractor.json() + assert result is None From f1a8db0a654f8a59e8b00afd7a6fd89a508b1a10 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 3 Apr 2025 16:50:27 +0200 Subject: [PATCH 235/591] tests: Move django under toxgen (#4238) --- .github/workflows/test-integrations-web-1.yml | 2 +- scripts/populate_tox/config.py | 19 ++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 44 -------- tox.ini | 101 +++++++++--------- 5 files changed, 68 insertions(+), 99 deletions(-) diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index a294301dbc..6d3e62a78a 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.10","3.12","3.13"] + python-version: ["3.8","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 3e8f6cf898..0bacfcaa7b 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -29,6 +29,25 @@ "clickhouse_driver": { "package": "clickhouse-driver", }, + "django": { + "package": "django", + "deps": { + "*": [ + "psycopg2-binary", + "djangorestframework", + "pytest-django", + "Werkzeug", + ], + ">=3.0": ["pytest-asyncio"], + ">=2.2,<3.1": ["six"], + "<3.3": [ + "djangorestframework>=3.0,<4.0", + "Werkzeug<2.1.0", + ], + "<3.1": ["pytest-django<4.0"], + ">=2.0": ["channels[daphne]"], + }, + }, "dramatiq": { "package": "dramatiq", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index d1e6cbca71..df45e30ed9 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -69,7 +69,6 @@ "boto3", "chalice", "cohere", - "django", "fastapi", "gcp", "httpx", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 1514ff197a..e599f45436 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -80,21 +80,6 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5 {py3.9,py3.11,py3.12}-cohere-latest - # Django - # - Django 1.x - {py3.6,py3.7}-django-v{1.11} - # - Django 2.x - {py3.6,py3.7}-django-v{2.0} - {py3.6,py3.9}-django-v{2.2} - # - Django 3.x - {py3.6,py3.9}-django-v{3.0} - {py3.6,py3.9,py3.11}-django-v{3.2} - # - Django 4.x - {py3.8,py3.11,py3.12}-django-v{4.0,4.1,4.2} - # - Django 5.x - {py3.10,py3.11,py3.12}-django-v{5.0,5.1} - {py3.10,py3.12,py3.13}-django-latest - # FastAPI {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest @@ -267,35 +252,6 @@ deps = cohere-v5: cohere~=5.3.3 cohere-latest: cohere - # Django - django: psycopg2-binary - django-v{1.11,2.0,2.1,2.2,3.0,3.1,3.2}: djangorestframework>=3.0.0,<4.0.0 - django-v{2.0,2.2,3.0,3.2,4.0,4.1,4.2,5.0,5.1}: channels[daphne] - django-v{2.2,3.0}: six - django-v{1.11,2.0,2.2,3.0,3.2}: Werkzeug<2.1.0 - django-v{1.11,2.0,2.2,3.0}: pytest-django<4.0 - django-v{3.2,4.0,4.1,4.2,5.0,5.1}: pytest-django - django-v{4.0,4.1,4.2,5.0,5.1}: djangorestframework - django-v{4.0,4.1,4.2,5.0,5.1}: pytest-asyncio - django-v{4.0,4.1,4.2,5.0,5.1}: Werkzeug - django-latest: djangorestframework - django-latest: pytest-asyncio - django-latest: pytest-django - django-latest: Werkzeug - django-latest: channels[daphne] - - django-v1.11: Django~=1.11.0 - django-v2.0: Django~=2.0.0 - django-v2.2: Django~=2.2.0 - django-v3.0: Django~=3.0.0 - django-v3.2: Django~=3.2.0 - django-v4.0: Django~=4.0.0 - django-v4.1: Django~=4.1.0 - django-v4.2: Django~=4.2.0 - django-v5.0: Django~=5.0.0 - django-v5.1: Django==5.1rc1 - django-latest: Django - # FastAPI fastapi: httpx # (this is a dependency of httpx) diff --git a/tox.ini b/tox.ini index a093b4de00..1854b0f711 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-03-31T10:49:05.789167+00:00 +# Last generated: 2025-04-03T11:46:44.595900+00:00 [tox] requires = @@ -80,21 +80,6 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5 {py3.9,py3.11,py3.12}-cohere-latest - # Django - # - Django 1.x - {py3.6,py3.7}-django-v{1.11} - # - Django 2.x - {py3.6,py3.7}-django-v{2.0} - {py3.6,py3.9}-django-v{2.2} - # - Django 3.x - {py3.6,py3.9}-django-v{3.0} - {py3.6,py3.9,py3.11}-django-v{3.2} - # - Django 4.x - {py3.8,py3.11,py3.12}-django-v{4.0,4.1,4.2} - # - Django 5.x - {py3.10,py3.11,py3.12}-django-v{5.0,5.1} - {py3.10,py3.12,py3.13}-django-latest - # FastAPI {py3.7,py3.10}-fastapi-v{0.79} {py3.8,py3.12,py3.13}-fastapi-latest @@ -217,7 +202,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.227.7 {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.262.6 + {py3.9,py3.12,py3.13}-strawberry-v0.263.0 # ~~~ Network ~~~ @@ -230,8 +215,7 @@ envlist = # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 - {py3.8,py3.11,py3.12}-celery-v5.4.0 - {py3.8,py3.12,py3.13}-celery-v5.5.0rc5 + {py3.8,py3.12,py3.13}-celery-v5.5.0 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -245,6 +229,14 @@ envlist = # ~~~ Web 1 ~~~ + {py3.6}-django-v1.11.9 + {py3.6,py3.7}-django-v1.11.29 + {py3.6,py3.8,py3.9}-django-v2.2.28 + {py3.6,py3.9,py3.10}-django-v3.2.25 + {py3.8,py3.11,py3.12}-django-v4.2.20 + {py3.10,py3.11,py3.12}-django-v5.0.9 + {py3.10,py3.12,py3.13}-django-v5.2 + {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 {py3.8,py3.12,py3.13}-flask-v3.0.3 @@ -293,7 +285,7 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.9 - {py3.8,py3.11,py3.12}-trytond-v7.4.8 + {py3.8,py3.11,py3.12}-trytond-v7.4.9 {py3.7,py3.12,py3.13}-typer-v0.15.2 @@ -389,35 +381,6 @@ deps = cohere-v5: cohere~=5.3.3 cohere-latest: cohere - # Django - django: psycopg2-binary - django-v{1.11,2.0,2.1,2.2,3.0,3.1,3.2}: djangorestframework>=3.0.0,<4.0.0 - django-v{2.0,2.2,3.0,3.2,4.0,4.1,4.2,5.0,5.1}: channels[daphne] - django-v{2.2,3.0}: six - django-v{1.11,2.0,2.2,3.0,3.2}: Werkzeug<2.1.0 - django-v{1.11,2.0,2.2,3.0}: pytest-django<4.0 - django-v{3.2,4.0,4.1,4.2,5.0,5.1}: pytest-django - django-v{4.0,4.1,4.2,5.0,5.1}: djangorestframework - django-v{4.0,4.1,4.2,5.0,5.1}: pytest-asyncio - django-v{4.0,4.1,4.2,5.0,5.1}: Werkzeug - django-latest: djangorestframework - django-latest: pytest-asyncio - django-latest: pytest-django - django-latest: Werkzeug - django-latest: channels[daphne] - - django-v1.11: Django~=1.11.0 - django-v2.0: Django~=2.0.0 - django-v2.2: Django~=2.2.0 - django-v3.0: Django~=3.0.0 - django-v3.2: Django~=3.2.0 - django-v4.0: Django~=4.0.0 - django-v4.1: Django~=4.1.0 - django-v4.2: Django~=4.2.0 - django-v5.0: Django~=5.0.0 - django-v5.1: Django==5.1rc1 - django-latest: Django - # FastAPI fastapi: httpx # (this is a dependency of httpx) @@ -611,7 +574,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.262.6: strawberry-graphql[fastapi,flask]==0.262.6 + strawberry-v0.263.0: strawberry-graphql[fastapi,flask]==0.263.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.227.7: pydantic<2.11 @@ -632,8 +595,7 @@ deps = # ~~~ Tasks ~~~ celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 - celery-v5.4.0: celery==5.4.0 - celery-v5.5.0rc5: celery==5.5.0rc5 + celery-v5.5.0: celery==5.5.0 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 @@ -650,6 +612,39 @@ deps = # ~~~ Web 1 ~~~ + django-v1.11.9: django==1.11.9 + django-v1.11.29: django==1.11.29 + django-v2.2.28: django==2.2.28 + django-v3.2.25: django==3.2.25 + django-v4.2.20: django==4.2.20 + django-v5.0.9: django==5.0.9 + django-v5.2: django==5.2 + django: psycopg2-binary + django: djangorestframework + django: pytest-django + django: Werkzeug + django-v3.2.25: pytest-asyncio + django-v4.2.20: pytest-asyncio + django-v5.0.9: pytest-asyncio + django-v5.2: pytest-asyncio + django-v2.2.28: six + django-v1.11.9: djangorestframework>=3.0,<4.0 + django-v1.11.9: Werkzeug<2.1.0 + django-v1.11.29: djangorestframework>=3.0,<4.0 + django-v1.11.29: Werkzeug<2.1.0 + django-v2.2.28: djangorestframework>=3.0,<4.0 + django-v2.2.28: Werkzeug<2.1.0 + django-v3.2.25: djangorestframework>=3.0,<4.0 + django-v3.2.25: Werkzeug<2.1.0 + django-v1.11.9: pytest-django<4.0 + django-v1.11.29: pytest-django<4.0 + django-v2.2.28: pytest-django<4.0 + django-v2.2.28: channels[daphne] + django-v3.2.25: channels[daphne] + django-v4.2.20: channels[daphne] + django-v5.0.9: channels[daphne] + django-v5.2: channels[daphne] + flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 flask-v3.0.3: flask==3.0.3 @@ -731,7 +726,7 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.9: trytond==7.0.9 - trytond-v7.4.8: trytond==7.4.8 + trytond-v7.4.9: trytond==7.4.9 trytond: werkzeug trytond-v4.6.9: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 5147ab9fdf3e1a8a42fefbd665743ae01998ba66 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 3 Apr 2025 16:56:15 +0200 Subject: [PATCH 236/591] feat(breadcrumbs): add `_meta` information for truncation of breadcrumbs (#4007) - Implements annotations for breadcrumbs - Adds an `int` field to `Scope` to track the number of truncated breadcrumbs - When scopes are merged, the number of breadcrumbs that were removed are added - If breadcrumbs were truncated, add the original number of breadcrumbs to `_meta` - Closes https://github.com/getsentry/projects/issues/593 --------- Co-authored-by: Anton Pirker --- sentry_sdk/_types.py | 15 +++++++++++++-- sentry_sdk/client.py | 16 +++++++++++++++- sentry_sdk/scope.py | 30 +++++++++++++++++++++++------- sentry_sdk/scrubber.py | 5 ++++- tests/test_scrubber.py | 20 ++++++++++++++------ 5 files changed, 69 insertions(+), 17 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 22b91b202f..9bcb5a61f9 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -30,6 +30,17 @@ def __eq__(self, other): return self.value == other.value and self.metadata == other.metadata + def __str__(self): + # type: (AnnotatedValue) -> str + return str({"value": str(self.value), "metadata": str(self.metadata)}) + + def __len__(self): + # type: (AnnotatedValue) -> int + if self.value is not None: + return len(self.value) + else: + return 0 + @classmethod def removed_because_raw_data(cls): # type: () -> AnnotatedValue @@ -152,8 +163,8 @@ class SDKInfo(TypedDict): Event = TypedDict( "Event", { - "breadcrumbs": dict[ - Literal["values"], list[dict[str, Any]] + "breadcrumbs": Annotated[ + dict[Literal["values"], list[dict[str, Any]]] ], # TODO: We can expand on this type "check_in_id": str, "contexts": dict[str, dict[str, object]], diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 3350c1372a..4dfccb3132 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -498,6 +498,7 @@ def _prepare_event( # type: (...) -> Optional[Event] previous_total_spans = None # type: Optional[int] + previous_total_breadcrumbs = None # type: Optional[int] if event.get("timestamp") is None: event["timestamp"] = datetime.now(timezone.utc) @@ -534,6 +535,16 @@ def _prepare_event( dropped_spans = event.pop("_dropped_spans", 0) + spans_delta # type: int if dropped_spans > 0: previous_total_spans = spans_before + dropped_spans + if scope._n_breadcrumbs_truncated > 0: + breadcrumbs = event.get("breadcrumbs", {}) + values = ( + breadcrumbs.get("values", []) + if not isinstance(breadcrumbs, AnnotatedValue) + else [] + ) + previous_total_breadcrumbs = ( + len(values) + scope._n_breadcrumbs_truncated + ) if ( self.options["attach_stacktrace"] @@ -586,7 +597,10 @@ def _prepare_event( event["spans"] = AnnotatedValue( event.get("spans", []), {"len": previous_total_spans} ) - + if previous_total_breadcrumbs is not None: + event["breadcrumbs"] = AnnotatedValue( + event.get("breadcrumbs", []), {"len": previous_total_breadcrumbs} + ) # Postprocess the event here so that annotated types do # generally not surface in before_send if event is not None: diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index ce6037e6b6..f346569255 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -9,6 +9,7 @@ from functools import wraps from itertools import chain +from sentry_sdk._types import AnnotatedValue from sentry_sdk.attachments import Attachment from sentry_sdk.consts import DEFAULT_MAX_BREADCRUMBS, FALSE_VALUES, INSTRUMENTER from sentry_sdk.feature_flags import FlagBuffer, DEFAULT_FLAG_CAPACITY @@ -186,6 +187,7 @@ class Scope: "_contexts", "_extras", "_breadcrumbs", + "_n_breadcrumbs_truncated", "_event_processors", "_error_processors", "_should_capture", @@ -210,6 +212,7 @@ def __init__(self, ty=None, client=None): self._name = None # type: Optional[str] self._propagation_context = None # type: Optional[PropagationContext] + self._n_breadcrumbs_truncated = 0 # type: int self.client = NonRecordingClient() # type: sentry_sdk.client.BaseClient @@ -243,6 +246,7 @@ def __copy__(self): rv._extras = dict(self._extras) rv._breadcrumbs = copy(self._breadcrumbs) + rv._n_breadcrumbs_truncated = copy(self._n_breadcrumbs_truncated) rv._event_processors = list(self._event_processors) rv._error_processors = list(self._error_processors) rv._propagation_context = self._propagation_context @@ -916,6 +920,7 @@ def clear_breadcrumbs(self): # type: () -> None """Clears breadcrumb buffer.""" self._breadcrumbs = deque() # type: Deque[Breadcrumb] + self._n_breadcrumbs_truncated = 0 def add_attachment( self, @@ -983,6 +988,7 @@ def add_breadcrumb(self, crumb=None, hint=None, **kwargs): while len(self._breadcrumbs) > max_breadcrumbs: self._breadcrumbs.popleft() + self._n_breadcrumbs_truncated += 1 def start_transaction( self, @@ -1366,17 +1372,23 @@ def _apply_level_to_event(self, event, hint, options): def _apply_breadcrumbs_to_event(self, event, hint, options): # type: (Event, Hint, Optional[Dict[str, Any]]) -> None - event.setdefault("breadcrumbs", {}).setdefault("values", []).extend( - self._breadcrumbs - ) + event.setdefault("breadcrumbs", {}) + + # This check is just for mypy - + if not isinstance(event["breadcrumbs"], AnnotatedValue): + event["breadcrumbs"].setdefault("values", []) + event["breadcrumbs"]["values"].extend(self._breadcrumbs) # Attempt to sort timestamps try: - for crumb in event["breadcrumbs"]["values"]: - if isinstance(crumb["timestamp"], str): - crumb["timestamp"] = datetime_from_isoformat(crumb["timestamp"]) + if not isinstance(event["breadcrumbs"], AnnotatedValue): + for crumb in event["breadcrumbs"]["values"]: + if isinstance(crumb["timestamp"], str): + crumb["timestamp"] = datetime_from_isoformat(crumb["timestamp"]) - event["breadcrumbs"]["values"].sort(key=lambda crumb: crumb["timestamp"]) + event["breadcrumbs"]["values"].sort( + key=lambda crumb: crumb["timestamp"] + ) except Exception as err: logger.debug("Error when sorting breadcrumbs", exc_info=err) pass @@ -1564,6 +1576,10 @@ def update_from_scope(self, scope): self._extras.update(scope._extras) if scope._breadcrumbs: self._breadcrumbs.extend(scope._breadcrumbs) + if scope._n_breadcrumbs_truncated: + self._n_breadcrumbs_truncated = ( + self._n_breadcrumbs_truncated + scope._n_breadcrumbs_truncated + ) if scope._span: self._span = scope._span if scope._attachments: diff --git a/sentry_sdk/scrubber.py b/sentry_sdk/scrubber.py index 1df5573798..b0576c7e95 100644 --- a/sentry_sdk/scrubber.py +++ b/sentry_sdk/scrubber.py @@ -144,7 +144,10 @@ def scrub_breadcrumbs(self, event): # type: (Event) -> None with capture_internal_exceptions(): if "breadcrumbs" in event: - if "values" in event["breadcrumbs"]: + if ( + not isinstance(event["breadcrumbs"], AnnotatedValue) + and "values" in event["breadcrumbs"] + ): for value in event["breadcrumbs"]["values"]: if "data" in value: self.scrub_dict(value["data"]) diff --git a/tests/test_scrubber.py b/tests/test_scrubber.py index 2c462153dd..2cc5f4139f 100644 --- a/tests/test_scrubber.py +++ b/tests/test_scrubber.py @@ -119,25 +119,33 @@ def test_stack_var_scrubbing(sentry_init, capture_events): def test_breadcrumb_extra_scrubbing(sentry_init, capture_events): - sentry_init() + sentry_init(max_breadcrumbs=2) events = capture_events() - - logger.info("bread", extra=dict(foo=42, password="secret")) + logger.info("breadcrumb 1", extra=dict(foo=1, password="secret")) + logger.info("breadcrumb 2", extra=dict(bar=2, auth="secret")) + logger.info("breadcrumb 3", extra=dict(foobar=3, password="secret")) logger.critical("whoops", extra=dict(bar=69, auth="secret")) (event,) = events assert event["extra"]["bar"] == 69 assert event["extra"]["auth"] == "[Filtered]" - assert event["breadcrumbs"]["values"][0]["data"] == { - "foo": 42, + "bar": 2, + "auth": "[Filtered]", + } + assert event["breadcrumbs"]["values"][1]["data"] == { + "foobar": 3, "password": "[Filtered]", } assert event["_meta"]["extra"]["auth"] == {"": {"rem": [["!config", "s"]]}} assert event["_meta"]["breadcrumbs"] == { - "values": {"0": {"data": {"password": {"": {"rem": [["!config", "s"]]}}}}} + "": {"len": 3}, + "values": { + "0": {"data": {"auth": {"": {"rem": [["!config", "s"]]}}}}, + "1": {"data": {"password": {"": {"rem": [["!config", "s"]]}}}}, + }, } From adcfa0f6abf8850f3b007bde609d0f943f621786 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 3 Apr 2025 17:21:41 +0200 Subject: [PATCH 237/591] Trying to prevent the grpc setup from being flaky (#4233) Automatically select a port and not set it by hand also make creating of the channel more stable. --- tests/integrations/grpc/test_grpc.py | 163 ++++++++++--------- tests/integrations/grpc/test_grpc_aio.py | 190 +++++++++++++---------- 2 files changed, 197 insertions(+), 156 deletions(-) diff --git a/tests/integrations/grpc/test_grpc.py b/tests/integrations/grpc/test_grpc.py index a8872ef0b5..8d2698f411 100644 --- a/tests/integrations/grpc/test_grpc.py +++ b/tests/integrations/grpc/test_grpc.py @@ -1,10 +1,8 @@ -import os - import grpc import pytest from concurrent import futures -from typing import List, Optional +from typing import List, Optional, Tuple from unittest.mock import Mock from sentry_sdk import start_span, start_transaction @@ -19,25 +17,36 @@ ) -PORT = 50051 -PORT += os.getpid() % 100 # avoid port conflicts when running tests in parallel - - -def _set_up(interceptors: Optional[List[grpc.ServerInterceptor]] = None): +# Set up in-memory channel instead of network-based +def _set_up( + interceptors: Optional[List[grpc.ServerInterceptor]] = None, +) -> Tuple[grpc.Server, grpc.Channel]: + """ + Sets up a gRPC server and returns both the server and a channel connected to it. + This eliminates network dependencies and makes tests more reliable. + """ + # Create server with thread pool server = grpc.server( futures.ThreadPoolExecutor(max_workers=2), interceptors=interceptors, ) - add_gRPCTestServiceServicer_to_server(TestService(), server) - server.add_insecure_port("[::]:{}".format(PORT)) + # Add our test service to the server + servicer = TestService() + add_gRPCTestServiceServicer_to_server(servicer, server) + + # Use dynamic port allocation instead of hardcoded port + port = server.add_insecure_port("[::]:0") # Let gRPC choose an available port server.start() - return server + # Create channel connected to our server + channel = grpc.insecure_channel(f"localhost:{port}") # noqa: E231 + + return server, channel def _tear_down(server: grpc.Server): - server.stop(None) + server.stop(grace=None) # Immediate shutdown @pytest.mark.forked @@ -45,11 +54,11 @@ def test_grpc_server_starts_transaction(sentry_init, capture_events_forksafe): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) - stub.TestServe(gRPCTestMessage(text="test")) + # Use the provided channel + stub = gRPCTestServiceStub(channel) + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) @@ -76,11 +85,11 @@ def test_grpc_server_other_interceptors(sentry_init, capture_events_forksafe): mock_interceptor = Mock() mock_interceptor.intercept_service.side_effect = mock_intercept - server = _set_up(interceptors=[mock_interceptor]) + server, channel = _set_up(interceptors=[mock_interceptor]) - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) - stub.TestServe(gRPCTestMessage(text="test")) + # Use the provided channel + stub = gRPCTestServiceStub(channel) + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) @@ -103,30 +112,30 @@ def test_grpc_server_continues_transaction(sentry_init, capture_events_forksafe) sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) + # Use the provided channel + stub = gRPCTestServiceStub(channel) - with start_transaction() as transaction: - metadata = ( - ( - "baggage", - "sentry-trace_id={trace_id},sentry-environment=test," - "sentry-transaction=test-transaction,sentry-sample_rate=1.0".format( - trace_id=transaction.trace_id - ), + with start_transaction() as transaction: + metadata = ( + ( + "baggage", + "sentry-trace_id={trace_id},sentry-environment=test," + "sentry-transaction=test-transaction,sentry-sample_rate=1.0".format( + trace_id=transaction.trace_id ), - ( - "sentry-trace", - "{trace_id}-{parent_span_id}-{sampled}".format( - trace_id=transaction.trace_id, - parent_span_id=transaction.span_id, - sampled=1, - ), + ), + ( + "sentry-trace", + "{trace_id}-{parent_span_id}-{sampled}".format( + trace_id=transaction.trace_id, + parent_span_id=transaction.span_id, + sampled=1, ), - ) - stub.TestServe(gRPCTestMessage(text="test"), metadata=metadata) + ), + ) + stub.TestServe(gRPCTestMessage(text="test"), metadata=metadata) _tear_down(server=server) @@ -148,13 +157,13 @@ def test_grpc_client_starts_span(sentry_init, capture_events_forksafe): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) + # Use the provided channel + stub = gRPCTestServiceStub(channel) - with start_transaction(): - stub.TestServe(gRPCTestMessage(text="test")) + with start_transaction(): + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) @@ -183,13 +192,13 @@ def test_grpc_client_unary_stream_starts_span(sentry_init, capture_events_forksa sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) + # Use the provided channel + stub = gRPCTestServiceStub(channel) - with start_transaction(): - [el for el in stub.TestUnaryStream(gRPCTestMessage(text="test"))] + with start_transaction(): + [el for el in stub.TestUnaryStream(gRPCTestMessage(text="test"))] _tear_down(server=server) @@ -227,14 +236,14 @@ def test_grpc_client_other_interceptor(sentry_init, capture_events_forksafe): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - channel = grpc.intercept_channel(channel, MockClientInterceptor()) - stub = gRPCTestServiceStub(channel) + # Intercept the channel + channel = grpc.intercept_channel(channel, MockClientInterceptor()) + stub = gRPCTestServiceStub(channel) - with start_transaction(): - stub.TestServe(gRPCTestMessage(text="test")) + with start_transaction(): + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) @@ -267,13 +276,13 @@ def test_grpc_client_and_servers_interceptors_integration( sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) + # Use the provided channel + stub = gRPCTestServiceStub(channel) - with start_transaction(): - stub.TestServe(gRPCTestMessage(text="test")) + with start_transaction(): + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) @@ -290,13 +299,13 @@ def test_grpc_client_and_servers_interceptors_integration( @pytest.mark.forked def test_stream_stream(sentry_init): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) - response_iterator = stub.TestStreamStream(iter((gRPCTestMessage(text="test"),))) - for response in response_iterator: - assert response.text == "test" + # Use the provided channel + stub = gRPCTestServiceStub(channel) + response_iterator = stub.TestStreamStream(iter((gRPCTestMessage(text="test"),))) + for response in response_iterator: + assert response.text == "test" _tear_down(server=server) @@ -308,12 +317,12 @@ def test_stream_unary(sentry_init): Tracing not supported for it yet. """ sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) - response = stub.TestStreamUnary(iter((gRPCTestMessage(text="test"),))) - assert response.text == "test" + # Use the provided channel + stub = gRPCTestServiceStub(channel) + response = stub.TestStreamUnary(iter((gRPCTestMessage(text="test"),))) + assert response.text == "test" _tear_down(server=server) @@ -323,13 +332,13 @@ def test_span_origin(sentry_init, capture_events_forksafe): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) events = capture_events_forksafe() - server = _set_up() + server, channel = _set_up() - with grpc.insecure_channel("localhost:{}".format(PORT)) as channel: - stub = gRPCTestServiceStub(channel) + # Use the provided channel + stub = gRPCTestServiceStub(channel) - with start_transaction(name="custom_transaction"): - stub.TestServe(gRPCTestMessage(text="test")) + with start_transaction(name="custom_transaction"): + stub.TestServe(gRPCTestMessage(text="test")) _tear_down(server=server) diff --git a/tests/integrations/grpc/test_grpc_aio.py b/tests/integrations/grpc/test_grpc_aio.py index 9ce9aef6a5..96e9a4dba8 100644 --- a/tests/integrations/grpc/test_grpc_aio.py +++ b/tests/integrations/grpc/test_grpc_aio.py @@ -1,5 +1,4 @@ import asyncio -import os import grpc import pytest @@ -17,37 +16,52 @@ gRPCTestServiceStub, ) -AIO_PORT = 50052 -AIO_PORT += os.getpid() % 100 # avoid port conflicts when running tests in parallel - @pytest_asyncio.fixture(scope="function") -async def grpc_server(sentry_init): +async def grpc_server_and_channel(sentry_init): + """ + Creates an async gRPC server and a channel connected to it. + Returns both for use in tests, and cleans up afterward. + """ sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) + + # Create server server = grpc.aio.server() - server.add_insecure_port("[::]:{}".format(AIO_PORT)) + + # Let gRPC choose a free port instead of hardcoding it + port = server.add_insecure_port("[::]:0") + + # Add service implementation add_gRPCTestServiceServicer_to_server(TestService, server) + # Start the server await asyncio.create_task(server.start()) + # Create channel connected to our server + channel = grpc.aio.insecure_channel(f"localhost:{port}") # noqa: E231 + try: - yield server + yield server, channel finally: + # Clean up resources + await channel.close() await server.stop(None) @pytest.mark.asyncio async def test_noop_for_unimplemented_method(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0, integrations=[GRPCIntegration()]) - server = grpc.aio.server() - server.add_insecure_port("[::]:{}".format(AIO_PORT)) + # Create empty server with no services + server = grpc.aio.server() + port = server.add_insecure_port("[::]:0") # Let gRPC choose a free port await asyncio.create_task(server.start()) events = capture_events() + try: async with grpc.aio.insecure_channel( - "localhost:{}".format(AIO_PORT) + f"localhost:{port}" # noqa: E231 ) as channel: stub = gRPCTestServiceStub(channel) with pytest.raises(grpc.RpcError) as exc: @@ -60,12 +74,13 @@ async def test_noop_for_unimplemented_method(sentry_init, capture_events): @pytest.mark.asyncio -async def test_grpc_server_starts_transaction(grpc_server, capture_events): +async def test_grpc_server_starts_transaction(grpc_server_and_channel, capture_events): + _, channel = grpc_server_and_channel events = capture_events() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - await stub.TestServe(gRPCTestMessage(text="test")) + # Use the provided channel + stub = gRPCTestServiceStub(channel) + await stub.TestServe(gRPCTestMessage(text="test")) (event,) = events span = event["spans"][0] @@ -79,32 +94,35 @@ async def test_grpc_server_starts_transaction(grpc_server, capture_events): @pytest.mark.asyncio -async def test_grpc_server_continues_transaction(grpc_server, capture_events): +async def test_grpc_server_continues_transaction( + grpc_server_and_channel, capture_events +): + _, channel = grpc_server_and_channel events = capture_events() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - - with sentry_sdk.start_transaction() as transaction: - metadata = ( - ( - "baggage", - "sentry-trace_id={trace_id},sentry-environment=test," - "sentry-transaction=test-transaction,sentry-sample_rate=1.0".format( - trace_id=transaction.trace_id - ), + # Use the provided channel + stub = gRPCTestServiceStub(channel) + + with sentry_sdk.start_transaction() as transaction: + metadata = ( + ( + "baggage", + "sentry-trace_id={trace_id},sentry-environment=test," + "sentry-transaction=test-transaction,sentry-sample_rate=1.0".format( + trace_id=transaction.trace_id ), - ( - "sentry-trace", - "{trace_id}-{parent_span_id}-{sampled}".format( - trace_id=transaction.trace_id, - parent_span_id=transaction.span_id, - sampled=1, - ), + ), + ( + "sentry-trace", + "{trace_id}-{parent_span_id}-{sampled}".format( + trace_id=transaction.trace_id, + parent_span_id=transaction.span_id, + sampled=1, ), - ) + ), + ) - await stub.TestServe(gRPCTestMessage(text="test"), metadata=metadata) + await stub.TestServe(gRPCTestMessage(text="test"), metadata=metadata) (event, _) = events span = event["spans"][0] @@ -119,16 +137,17 @@ async def test_grpc_server_continues_transaction(grpc_server, capture_events): @pytest.mark.asyncio -async def test_grpc_server_exception(grpc_server, capture_events): +async def test_grpc_server_exception(grpc_server_and_channel, capture_events): + _, channel = grpc_server_and_channel events = capture_events() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - try: - await stub.TestServe(gRPCTestMessage(text="exception")) - raise AssertionError() - except Exception: - pass + # Use the provided channel + stub = gRPCTestServiceStub(channel) + try: + await stub.TestServe(gRPCTestMessage(text="exception")) + raise AssertionError() + except Exception: + pass (event, _) = events @@ -139,28 +158,35 @@ async def test_grpc_server_exception(grpc_server, capture_events): @pytest.mark.asyncio -async def test_grpc_server_abort(grpc_server, capture_events): +async def test_grpc_server_abort(grpc_server_and_channel, capture_events): + _, channel = grpc_server_and_channel events = capture_events() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - try: - await stub.TestServe(gRPCTestMessage(text="abort")) - raise AssertionError() - except Exception: - pass + # Use the provided channel + stub = gRPCTestServiceStub(channel) + try: + await stub.TestServe(gRPCTestMessage(text="abort")) + raise AssertionError() + except Exception: + pass + + # Add a small delay to allow events to be collected + await asyncio.sleep(0.1) assert len(events) == 1 @pytest.mark.asyncio -async def test_grpc_client_starts_span(grpc_server, capture_events_forksafe): +async def test_grpc_client_starts_span( + grpc_server_and_channel, capture_events_forksafe +): + _, channel = grpc_server_and_channel events = capture_events_forksafe() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - with start_transaction(): - await stub.TestServe(gRPCTestMessage(text="test")) + # Use the provided channel + stub = gRPCTestServiceStub(channel) + with start_transaction(): + await stub.TestServe(gRPCTestMessage(text="test")) events.write_file.close() events.read_event() @@ -184,15 +210,16 @@ async def test_grpc_client_starts_span(grpc_server, capture_events_forksafe): @pytest.mark.asyncio async def test_grpc_client_unary_stream_starts_span( - grpc_server, capture_events_forksafe + grpc_server_and_channel, capture_events_forksafe ): + _, channel = grpc_server_and_channel events = capture_events_forksafe() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - with start_transaction(): - response = stub.TestUnaryStream(gRPCTestMessage(text="test")) - [_ async for _ in response] + # Use the provided channel + stub = gRPCTestServiceStub(channel) + with start_transaction(): + response = stub.TestUnaryStream(gRPCTestMessage(text="test")) + [_ async for _ in response] events.write_file.close() local_transaction = events.read_event() @@ -213,38 +240,43 @@ async def test_grpc_client_unary_stream_starts_span( @pytest.mark.asyncio -async def test_stream_stream(grpc_server): +async def test_stream_stream(grpc_server_and_channel): """ Test to verify stream-stream works. Tracing not supported for it yet. """ - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - response = stub.TestStreamStream((gRPCTestMessage(text="test"),)) - async for r in response: - assert r.text == "test" + _, channel = grpc_server_and_channel + + # Use the provided channel + stub = gRPCTestServiceStub(channel) + response = stub.TestStreamStream((gRPCTestMessage(text="test"),)) + async for r in response: + assert r.text == "test" @pytest.mark.asyncio -async def test_stream_unary(grpc_server): +async def test_stream_unary(grpc_server_and_channel): """ Test to verify stream-stream works. Tracing not supported for it yet. """ - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - response = await stub.TestStreamUnary((gRPCTestMessage(text="test"),)) - assert response.text == "test" + _, channel = grpc_server_and_channel + + # Use the provided channel + stub = gRPCTestServiceStub(channel) + response = await stub.TestStreamUnary((gRPCTestMessage(text="test"),)) + assert response.text == "test" @pytest.mark.asyncio -async def test_span_origin(grpc_server, capture_events_forksafe): +async def test_span_origin(grpc_server_and_channel, capture_events_forksafe): + _, channel = grpc_server_and_channel events = capture_events_forksafe() - async with grpc.aio.insecure_channel("localhost:{}".format(AIO_PORT)) as channel: - stub = gRPCTestServiceStub(channel) - with start_transaction(name="custom_transaction"): - await stub.TestServe(gRPCTestMessage(text="test")) + # Use the provided channel + stub = gRPCTestServiceStub(channel) + with start_transaction(name="custom_transaction"): + await stub.TestServe(gRPCTestMessage(text="test")) events.write_file.close() @@ -283,7 +315,7 @@ async def TestServe(cls, request, context): # noqa: N802 raise cls.TestException() if request.text == "abort": - await context.abort(grpc.StatusCode.ABORTED) + await context.abort(grpc.StatusCode.ABORTED, "Aborted!") return gRPCTestMessage(text=request.text) From 8016aab4c5c31702473b492e49cf233baa8961c9 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 8 Apr 2025 14:17:56 +0000 Subject: [PATCH 238/591] build(deps): bump actions/create-github-app-token from 1.12.0 to 2.0.2 (#4248) --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index ed8b3e4094..a0e39a5784 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@d72941d797fd3113feb6b93fd0dec494b13a2547 # v1.12.0 + uses: actions/create-github-app-token@3ff1caaa28b64c9cc276ce0a02e2ff584f3900c5 # v2.0.2 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From 2ba4ed096166bc6f797ffdccc1c8c5e8e3205c12 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 9 Apr 2025 08:54:25 +0200 Subject: [PATCH 239/591] toxgen: Retry & fail if we fail to fetch PyPI data (#4251) - try to refetch data if PyPI returns an error - if we fail after 3 tries, fail the whole script (it doesn't make sense to run it without access to up-to-date PyPI data) --- scripts/populate_tox/populate_tox.py | 56 +++++++++++++++++++--------- tox.ini | 18 ++++----- 2 files changed, 48 insertions(+), 26 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index df45e30ed9..c405a2bc23 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -36,6 +36,8 @@ lstrip_blocks=True, ) +PYPI_COOLDOWN = 0.15 # seconds to wait between requests to PyPI + PYPI_PROJECT_URL = "https://pypi.python.org/pypi/%7Bproject%7D/json" PYPI_VERSION_URL = "https://pypi.python.org/pypi/%7Bproject%7D/%7Bversion%7D/json" CLASSIFIER_PREFIX = "Programming Language :: Python :: " @@ -88,27 +90,34 @@ } -@functools.cache -def fetch_package(package: str) -> dict: - """Fetch package metadata from PyPI.""" - url = PYPI_PROJECT_URL.format(project=package) - pypi_data = requests.get(url) +def fetch_url(/service/url: str) -> Optional[dict]: + for attempt in range(3): + pypi_data = requests.get(url) - if pypi_data.status_code != 200: - print(f"{package} not found") + if pypi_data.status_code == 200: + return pypi_data.json() - return pypi_data.json() + backoff = PYPI_COOLDOWN * 2**attempt + print( + f"{url} returned an error: {pypi_data.status_code}. Attempt {attempt + 1}/3. Waiting {backoff}s" + ) + time.sleep(backoff) + + return None @functools.cache -def fetch_release(package: str, version: Version) -> dict: - url = PYPI_VERSION_URL.format(project=package, version=version) - pypi_data = requests.get(url) +def fetch_package(package: str) -> Optional[dict]: + """Fetch package metadata from PyPI.""" + url = PYPI_PROJECT_URL.format(project=package) + return fetch_https://github.com/url(url) - if pypi_data.status_code != 200: - print(f"{package} not found") - return pypi_data.json() +@functools.cache +def fetch_release(package: str, version: Version) -> Optional[dict]: + """Fetch release metadata from PyPI.""" + url = PYPI_VERSION_URL.format(project=package, version=version) + return fetch_https://github.com/url(url) def _prefilter_releases( @@ -229,8 +238,14 @@ def get_supported_releases( expected_python_versions = SpecifierSet(f">={MIN_PYTHON_VERSION}") def _supports_lowest(release: Version) -> bool: - time.sleep(0.1) # don't DoS PYPI - py_versions = determine_python_versions(fetch_release(package, release)) + time.sleep(PYPI_COOLDOWN) # don't DoS PYPI + + pypi_data = fetch_release(package, release) + if pypi_data is None: + print("Failed to fetch necessary data from PyPI. Aborting.") + sys.exit(1) + + py_versions = determine_python_versions(pypi_data) target_python_versions = TEST_SUITE_CONFIG[integration].get("python") if target_python_versions: target_python_versions = SpecifierSet(target_python_versions) @@ -499,7 +514,11 @@ def _add_python_versions_to_release( integration: str, package: str, release: Version ) -> None: release_pypi_data = fetch_release(package, release) - time.sleep(0.1) # give PYPI some breathing room + if release_pypi_data is None: + print("Failed to fetch necessary data from PyPI. Aborting.") + sys.exit(1) + + time.sleep(PYPI_COOLDOWN) # give PYPI some breathing room target_python_versions = TEST_SUITE_CONFIG[integration].get("python") if target_python_versions: @@ -592,6 +611,9 @@ def main(fail_on_changes: bool = False) -> None: # Fetch data for the main package pypi_data = fetch_package(package) + if pypi_data is None: + print("Failed to fetch necessary data from PyPI. Aborting.") + sys.exit(1) # Get the list of all supported releases diff --git a/tox.ini b/tox.ini index 1854b0f711..c04691e2ac 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-03T11:46:44.595900+00:00 +# Last generated: 2025-04-08T10:33:11.499210+00:00 [tox] requires = @@ -179,7 +179,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 - {py3.7,py3.12,py3.13}-statsig-v0.57.1 + {py3.7,py3.12,py3.13}-statsig-v0.57.2 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -202,7 +202,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.227.7 {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.263.0 + {py3.9,py3.12,py3.13}-strawberry-v0.263.2 # ~~~ Network ~~~ @@ -215,7 +215,7 @@ envlist = # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 - {py3.8,py3.12,py3.13}-celery-v5.5.0 + {py3.8,py3.12,py3.13}-celery-v5.5.1 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -260,7 +260,7 @@ envlist = {py3.8,py3.10,py3.11}-litestar-v2.0.1 {py3.8,py3.11,py3.12}-litestar-v2.5.5 {py3.8,py3.11,py3.12}-litestar-v2.10.0 - {py3.8,py3.12,py3.13}-litestar-v2.15.1 + {py3.8,py3.12,py3.13}-litestar-v2.15.2 {py3.6}-pyramid-v1.8.6 {py3.6,py3.8,py3.9}-pyramid-v1.10.8 @@ -542,7 +542,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 - statsig-v0.57.1: statsig==0.57.1 + statsig-v0.57.2: statsig==0.57.2 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -574,7 +574,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.263.0: strawberry-graphql[fastapi,flask]==0.263.0 + strawberry-v0.263.2: strawberry-graphql[fastapi,flask]==0.263.2 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.227.7: pydantic<2.11 @@ -595,7 +595,7 @@ deps = # ~~~ Tasks ~~~ celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 - celery-v5.5.0: celery==5.5.0 + celery-v5.5.1: celery==5.5.1 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 @@ -683,7 +683,7 @@ deps = litestar-v2.0.1: litestar==2.0.1 litestar-v2.5.5: litestar==2.5.5 litestar-v2.10.0: litestar==2.10.0 - litestar-v2.15.1: litestar==2.15.1 + litestar-v2.15.2: litestar==2.15.2 litestar: pytest-asyncio litestar: python-multipart litestar: requests From 7cb0451865f82f3b6382c574ef57014a68f77c4f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 9 Apr 2025 09:47:59 +0200 Subject: [PATCH 240/591] feat(tests): Add optional cutoff to toxgen (#4243) This will be useful to identify old versions of packages when we're doing a deprecation round. --- scripts/populate_tox/populate_tox.py | 16 ++++++++++++---- 1 file changed, 12 insertions(+), 4 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index c405a2bc23..58dbed0308 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -9,7 +9,7 @@ import time from bisect import bisect_left from collections import defaultdict -from datetime import datetime, timezone +from datetime import datetime, timedelta, timezone # noqa: F401 from importlib.metadata import metadata from packaging.specifiers import SpecifierSet from packaging.version import Version @@ -29,6 +29,10 @@ from split_tox_gh_actions.split_tox_gh_actions import GROUPS +# Set CUTOFF this to a datetime to ignore packages older than CUTOFF +CUTOFF = None +# CUTOFF = datetime.now(tz=timezone.utc) - timedelta(days=365 * 5) + TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" ENV = Environment( loader=FileSystemLoader(Path(__file__).resolve().parent), @@ -162,9 +166,13 @@ def _prefilter_releases( if meta["yanked"]: continue - if older_than is not None: - if datetime.fromisoformat(meta["upload_time_iso_8601"]) > older_than: - continue + uploaded = datetime.fromisoformat(meta["upload_time_iso_8601"]) + + if older_than is not None and uploaded > older_than: + continue + + if CUTOFF is not None and uploaded < CUTOFF: + continue version = Version(release) From 6a1364d4bb27b4d15f829f36dabbb18cb8f32cdf Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Wed, 9 Apr 2025 10:25:43 +0200 Subject: [PATCH 241/591] feat(logs): Add sentry.origin attribute for log handler (#4250) resolves https://linear.app/getsentry/issue/LOGS-13 Docs: https://develop-docs-git-abhi-logs-sdk-developer-documentation.sentry.dev/sdk/telemetry/logs/#default-attributes > If a log is generated by an SDK integration, the SDK should also set the sentry.origin attribute, as per the [Trace Origin](https://develop-docs-git-abhi-logs-sdk-developer-documentation.sentry.dev/sdk/telemetry/logs/traces/trace-origin/) documentation. It is assumed that logs without a sentry.origin attribute are manually created by the user. --- sentry_sdk/integrations/logging.py | 4 +++- tests/test_logs.py | 1 + 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index ba6e6581b7..1fbecb2e08 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -358,7 +358,9 @@ def _capture_log_from_record(client, record): # type: (BaseClient, LogRecord) -> None scope = sentry_sdk.get_current_scope() otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) - attrs = {} # type: dict[str, str | bool | float | int] + attrs = { + "sentry.origin": "auto.logger.log", + } # type: dict[str, str | bool | float | int] if isinstance(record.msg, str): attrs["sentry.message.template"] = record.msg if record.args is not None: diff --git a/tests/test_logs.py b/tests/test_logs.py index 1305f243de..fb824760a8 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -283,6 +283,7 @@ def test_logger_integration_warning(sentry_init, capture_envelopes): assert attrs["sentry.environment"] == "production" assert attrs["sentry.message.parameters.0"] == "1" assert attrs["sentry.message.parameters.1"] == "2" + assert attrs["sentry.origin"] == "auto.logger.log" assert logs[0]["severity_number"] == 13 assert logs[0]["severity_text"] == "warn" From e05ed0aa62cfe2c992b26b07c64c3148f837a609 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 9 Apr 2025 10:57:50 +0200 Subject: [PATCH 242/591] chore: Deprecate `same_process_as_parent` (#4244) Preparing to remove this in https://github.com/getsentry/sentry-python/pull/4201 --- sentry_sdk/tracing.py | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 13d9f63d5e..ab1a7a8fdf 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -323,6 +323,13 @@ def __init__( self.scope = self.scope or hub.scope + if same_process_as_parent is not None: + warnings.warn( + "The `same_process_as_parent` parameter is deprecated.", + DeprecationWarning, + stacklevel=2, + ) + if start_timestamp is None: start_timestamp = datetime.now(timezone.utc) elif isinstance(start_timestamp, float): From acf508cb38c633cbf95561343684e964876dd32c Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Wed, 9 Apr 2025 15:43:48 +0200 Subject: [PATCH 243/591] feat(logs): Add server.address to logs (#4257) Docs: https://develop-docs-git-abhi-logs-sdk-developer-documentation.sentry.dev/sdk/telemetry/logs/#default-attributes > [BACKEND SDKS ONLY] `server.address`: The address of the server that sent the log. Equivalent to server_name we attach to errors and transactions. `server.address` convention docs: https://getsentry.github.io/sentry-conventions/generated/attributes/server.html#serveraddress resolves https://linear.app/getsentry/issue/LOGS-33 --- sentry_sdk/client.py | 5 +++++ tests/test_logs.py | 4 +++- 2 files changed, 8 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 4dfccb3132..102392c61d 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -27,6 +27,7 @@ from sentry_sdk.tracing import trace from sentry_sdk.transport import BaseHttpTransport, make_transport from sentry_sdk.consts import ( + SPANDATA, DEFAULT_MAX_VALUE_LENGTH, DEFAULT_OPTIONS, INSTRUMENTER, @@ -894,6 +895,10 @@ def _capture_experimental_log(self, current_scope, log): return isolation_scope = current_scope.get_isolation_scope() + server_name = self.options.get("server_name") + if server_name is not None and SPANDATA.SERVER_ADDRESS not in log["attributes"]: + log["attributes"][SPANDATA.SERVER_ADDRESS] = server_name + environment = self.options.get("environment") if environment is not None and "sentry.environment" not in log["attributes"]: log["attributes"]["sentry.environment"] = environment diff --git a/tests/test_logs.py b/tests/test_logs.py index fb824760a8..d58aa9acdd 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -11,6 +11,7 @@ from sentry_sdk.envelope import Envelope from sentry_sdk.integrations.logging import LoggingIntegration from sentry_sdk.types import Log +from sentry_sdk.consts import SPANDATA minimum_python_37 = pytest.mark.skipif( sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" @@ -161,7 +162,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): """ Passing arbitrary attributes to log messages. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(_experiments={"enable_logs": True}, server_name="test-server") envelopes = capture_envelopes() attrs = { @@ -184,6 +185,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): assert logs[0]["attributes"]["sentry.environment"] == "production" assert "sentry.release" in logs[0]["attributes"] assert logs[0]["attributes"]["sentry.message.parameters.my_var"] == "some value" + assert logs[0]["attributes"][SPANDATA.SERVER_ADDRESS] == "test-server" @minimum_python_37 From 97c435a82c4ddca2706794ed90b74f6527f8162f Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Wed, 9 Apr 2025 16:00:16 +0200 Subject: [PATCH 244/591] feat(logs): Add sdk name and version as log attributes (#4262) Docs: https://develop-docs-git-abhi-logs-sdk-developer-documentation.sentry.dev/sdk/telemetry/logs/#default-attributes > sentry.sdk.name: The name of the SDK that sent the log > sentry.sdk.version: The version of the SDK that sent the log convention docs: - `sentry.sdk.name`: https://getsentry.github.io/sentry-conventions/generated/attributes/sentry.html#sentrysdkname - `sentry.sdk.version`: https://getsentry.github.io/sentry-conventions/generated/attributes/sentry.html#sentrysdkversion resolves https://linear.app/getsentry/issue/PY-1/ --- sentry_sdk/client.py | 3 +++ tests/test_logs.py | 4 +++- 2 files changed, 6 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 102392c61d..f06166bcc8 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -895,6 +895,9 @@ def _capture_experimental_log(self, current_scope, log): return isolation_scope = current_scope.get_isolation_scope() + log["attributes"]["sentry.sdk.name"] = SDK_INFO["name"] + log["attributes"]["sentry.sdk.version"] = SDK_INFO["version"] + server_name = self.options.get("server_name") if server_name is not None and SPANDATA.SERVER_ADDRESS not in log["attributes"]: log["attributes"][SPANDATA.SERVER_ADDRESS] = server_name diff --git a/tests/test_logs.py b/tests/test_logs.py index d58aa9acdd..1c34d52b20 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -11,7 +11,7 @@ from sentry_sdk.envelope import Envelope from sentry_sdk.integrations.logging import LoggingIntegration from sentry_sdk.types import Log -from sentry_sdk.consts import SPANDATA +from sentry_sdk.consts import SPANDATA, VERSION minimum_python_37 = pytest.mark.skipif( sys.version_info < (3, 7), reason="Asyncio tests need Python >= 3.7" @@ -186,6 +186,8 @@ def test_logs_attributes(sentry_init, capture_envelopes): assert "sentry.release" in logs[0]["attributes"] assert logs[0]["attributes"]["sentry.message.parameters.my_var"] == "some value" assert logs[0]["attributes"][SPANDATA.SERVER_ADDRESS] == "test-server" + assert logs[0]["attributes"]["sentry.sdk.name"] == "sentry.python" + assert logs[0]["attributes"]["sentry.sdk.version"] == VERSION @minimum_python_37 From fb6d3745c8d7aef20142dbca708c884f63f7f821 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 10 Apr 2025 10:49:17 +0200 Subject: [PATCH 245/591] meta: Change CODEOWNERS back to Python SDK owners (#4269) Don't spam the whole backend SDK team on each PR. --- .github/CODEOWNERS | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/CODEOWNERS b/.github/CODEOWNERS index e5d24f170c..1dc1a4882f 100644 --- a/.github/CODEOWNERS +++ b/.github/CODEOWNERS @@ -1 +1 @@ -* @getsentry/team-web-sdk-backend +* @getsentry/owners-python-sdk From 6000f87d2d3ec77fc4a1ec391d357ff3969a873b Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 10 Apr 2025 11:44:10 +0200 Subject: [PATCH 246/591] feat(transport): Add a timeout (#4252) For some reason, we don't define any timeouts in our default transport(s). With this change: - We add a 30s total timeout for the whole connect+read cycle in the default HTTP transport - In the experimental HTTP/2 httpcore-based transport there is no way to set a single timeout, so we set 15s each for getting a connection from the pool, connecting, writing, and reading Backend SDKs in general set wildly different timeouts, from 30s in Go to <5s in Ruby or PHP. I went for the higher end of the range here since this is mainly meant to prevent the SDK preventing process shutdown like described in https://github.com/getsentry/sentry-python/issues/4247 -- we don't want to cut off legitimate requests that are just taking a long time. (I was considering going even higher, maybe to 60s -- but I think 30s is a good first shot at this and we can always change it later.) --- sentry_sdk/transport.py | 13 +++++++++++++ tests/test_transport.py | 36 ++++++++++++++++++++++++++++++++++++ 2 files changed, 49 insertions(+) diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index efc955ca7b..f9a5262903 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -196,6 +196,8 @@ def _parse_rate_limits(header, now=None): class BaseHttpTransport(Transport): """The base HTTP transport.""" + TIMEOUT = 30 # seconds + def __init__(self, options): # type: (Self, Dict[str, Any]) -> None from sentry_sdk.consts import VERSION @@ -621,6 +623,7 @@ def _get_pool_options(self): options = { "num_pools": 2 if num_pools is None else int(num_pools), "cert_reqs": "CERT_REQUIRED", + "timeout": urllib3.Timeout(total=self.TIMEOUT), } socket_options = None # type: Optional[List[Tuple[int, int, int | bytes]]] @@ -736,6 +739,8 @@ def __init__(self, options): class Http2Transport(BaseHttpTransport): # type: ignore """The HTTP2 transport based on httpcore.""" + TIMEOUT = 15 + if TYPE_CHECKING: _pool: Union[ httpcore.SOCKSProxy, httpcore.HTTPProxy, httpcore.ConnectionPool @@ -765,6 +770,14 @@ def _request( self._auth.get_api_url(https://github.com/endpoint_type), content=body, headers=headers, # type: ignore + extensions={ + "timeout": { + "pool": self.TIMEOUT, + "connect": self.TIMEOUT, + "write": self.TIMEOUT, + "read": self.TIMEOUT, + } + }, ) return response diff --git a/tests/test_transport.py b/tests/test_transport.py index d24bea0491..6eb7cdf829 100644 --- a/tests/test_transport.py +++ b/tests/test_transport.py @@ -14,6 +14,11 @@ from pytest_localserver.http import WSGIServer from werkzeug.wrappers import Request, Response +try: + import httpcore +except (ImportError, ModuleNotFoundError): + httpcore = None + try: import gevent except ImportError: @@ -274,6 +279,37 @@ def test_keep_alive_on_by_default(make_client): assert "socket_options" not in options +def test_default_timeout(make_client): + client = make_client() + + options = client.transport._get_pool_options() + assert "timeout" in options + assert options["timeout"].total == client.transport.TIMEOUT + + +@pytest.mark.skipif(not PY38, reason="HTTP2 libraries are only available in py3.8+") +def test_default_timeout_http2(make_client): + client = make_client(_experiments={"transport_http2": True}) + + with mock.patch( + "sentry_sdk.transport.httpcore.ConnectionPool.request", + return_value=httpcore.Response(200), + ) as request_mock: + sentry_sdk.get_global_scope().set_client(client) + capture_message("hi") + client.flush() + + request_mock.assert_called_once() + assert request_mock.call_args.kwargs["extensions"] == { + "timeout": { + "pool": client.transport.TIMEOUT, + "connect": client.transport.TIMEOUT, + "write": client.transport.TIMEOUT, + "read": client.transport.TIMEOUT, + } + } + + @pytest.mark.skipif(not PY38, reason="HTTP2 libraries are only available in py3.8+") def test_http2_with_https_dsn(make_client): client = make_client(_experiments={"transport_http2": True}) From be229121608feba3033dbe84ef1884b6ba6ad3ee Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 14 Apr 2025 10:16:38 +0200 Subject: [PATCH 247/591] test(tracing): Simplify static/classmethod tracing tests (#4278) These tests were causing flakes where the mock method was being called more than once. The tests were also difficult to understand. This change removes the need for mocking (hopefully increasing test stability) and also should hopefully make it easier to understand what these tests are meant to be checking --- tests/test_basics.py | 119 +++++++++++++++++++++++++++++++------------ 1 file changed, 86 insertions(+), 33 deletions(-) diff --git a/tests/test_basics.py b/tests/test_basics.py index e16956979a..94ced5013a 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -9,7 +9,6 @@ import pytest from sentry_sdk.client import Client from sentry_sdk.utils import datetime_from_isoformat -from tests.conftest import patch_start_tracing_child import sentry_sdk import sentry_sdk.scope @@ -935,46 +934,100 @@ def class_(cls, arg): return cls, arg -def test_staticmethod_tracing(sentry_init): - test_staticmethod_name = "tests.test_basics.TracingTestClass.static" +# We need to fork here because the test modifies tests.test_basics.TracingTestClass +@pytest.mark.forked +def test_staticmethod_class_tracing(sentry_init, capture_events): + sentry_init( + debug=True, + traces_sample_rate=1.0, + functions_to_trace=[ + {"qualified_name": "tests.test_basics.TracingTestClass.static"} + ], + ) - assert ( - ".".join( - [ - TracingTestClass.static.__module__, - TracingTestClass.static.__qualname__, - ] - ) - == test_staticmethod_name - ), "The test static method was moved or renamed. Please update the name accordingly" + events = capture_events() - sentry_init(functions_to_trace=[{"qualified_name": test_staticmethod_name}]) + with sentry_sdk.start_transaction(name="test"): + assert TracingTestClass.static(1) == 1 - for instance_or_class in (TracingTestClass, TracingTestClass()): - with patch_start_tracing_child() as fake_start_child: - assert instance_or_class.static(1) == 1 - assert fake_start_child.call_count == 1 + (event,) = events + assert event["type"] == "transaction" + assert event["transaction"] == "test" + (span,) = event["spans"] + assert span["description"] == "tests.test_basics.TracingTestClass.static" -def test_classmethod_tracing(sentry_init): - test_classmethod_name = "tests.test_basics.TracingTestClass.class_" - assert ( - ".".join( - [ - TracingTestClass.class_.__module__, - TracingTestClass.class_.__qualname__, - ] - ) - == test_classmethod_name - ), "The test class method was moved or renamed. Please update the name accordingly" +# We need to fork here because the test modifies tests.test_basics.TracingTestClass +@pytest.mark.forked +def test_staticmethod_instance_tracing(sentry_init, capture_events): + sentry_init( + debug=True, + traces_sample_rate=1.0, + functions_to_trace=[ + {"qualified_name": "tests.test_basics.TracingTestClass.static"} + ], + ) + + events = capture_events() + + with sentry_sdk.start_transaction(name="test"): + assert TracingTestClass().static(1) == 1 + + (event,) = events + assert event["type"] == "transaction" + assert event["transaction"] == "test" - sentry_init(functions_to_trace=[{"qualified_name": test_classmethod_name}]) + (span,) = event["spans"] + assert span["description"] == "tests.test_basics.TracingTestClass.static" + + +# We need to fork here because the test modifies tests.test_basics.TracingTestClass +@pytest.mark.forked +def test_classmethod_class_tracing(sentry_init, capture_events): + sentry_init( + debug=True, + traces_sample_rate=1.0, + functions_to_trace=[ + {"qualified_name": "tests.test_basics.TracingTestClass.class_"} + ], + ) + + events = capture_events() + + with sentry_sdk.start_transaction(name="test"): + assert TracingTestClass.class_(1) == (TracingTestClass, 1) + + (event,) = events + assert event["type"] == "transaction" + assert event["transaction"] == "test" + + (span,) = event["spans"] + assert span["description"] == "tests.test_basics.TracingTestClass.class_" + + +# We need to fork here because the test modifies tests.test_basics.TracingTestClass +@pytest.mark.forked +def test_classmethod_instance_tracing(sentry_init, capture_events): + sentry_init( + debug=True, + traces_sample_rate=1.0, + functions_to_trace=[ + {"qualified_name": "tests.test_basics.TracingTestClass.class_"} + ], + ) + + events = capture_events() + + with sentry_sdk.start_transaction(name="test"): + assert TracingTestClass().class_(1) == (TracingTestClass, 1) + + (event,) = events + assert event["type"] == "transaction" + assert event["transaction"] == "test" - for instance_or_class in (TracingTestClass, TracingTestClass()): - with patch_start_tracing_child() as fake_start_child: - assert instance_or_class.class_(1) == (TracingTestClass, 1) - assert fake_start_child.call_count == 1 + (span,) = event["spans"] + assert span["description"] == "tests.test_basics.TracingTestClass.class_" def test_last_event_id(sentry_init): From 5689bc09fd223f80f65290e2ccb685b8acb9a5f2 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 14 Apr 2025 15:41:46 +0200 Subject: [PATCH 248/591] fix(debug): Do not consider parent loggers for debug logging (#4286) This reverts commit 37930840dcefba96e7708b19e461013a919e83a5, which made the SDK consider parent loggers when determining if the Sentry SDK should log debug messages. However, we should not consider parent loggers, since we only want the SDK to log debug messages when configured to do so via `debug=True` (in `sentry_sdk.init`), the `SENTRY_DEBUG` environment variable, or via a specific logger configuration for `sentry_sdk.errors`. With 37930840dcefba96e7708b19e461013a919e83a5, a custom root logger configuration would also cause SDK logs to be emitted. The issue 37930840dcefba96e7708b19e461013a919e83a5 was meant to fix (#3944) will require a different fix. Fixes #4266 --- sentry_sdk/debug.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/debug.py b/sentry_sdk/debug.py index f740d92dec..e4c686a3e8 100644 --- a/sentry_sdk/debug.py +++ b/sentry_sdk/debug.py @@ -19,7 +19,7 @@ def filter(self, record): def init_debug_support(): # type: () -> None - if not logger.hasHandlers(): + if not logger.handlers: configure_logger() From 54d2c7e37b0f31ffcbd43e1f904ee9e2d8f4b650 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 14 Apr 2025 13:45:15 +0000 Subject: [PATCH 249/591] release: 2.26.0 --- CHANGELOG.md | 21 +++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 24 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a9294eaec1..5327b323a2 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,26 @@ # Changelog +## 2.26.0 + +### Various fixes & improvements + +- fix(debug): Do not consider parent loggers for debug logging (#4286) by @szokeasaurusrex +- test(tracing): Simplify static/classmethod tracing tests (#4278) by @szokeasaurusrex +- feat(transport): Add a timeout (#4252) by @sentrivana +- meta: Change CODEOWNERS back to Python SDK owners (#4269) by @sentrivana +- feat(logs): Add sdk name and version as log attributes (#4262) by @AbhiPrasad +- feat(logs): Add server.address to logs (#4257) by @AbhiPrasad +- chore: Deprecate `same_process_as_parent` (#4244) by @sentrivana +- feat(logs): Add sentry.origin attribute for log handler (#4250) by @AbhiPrasad +- feat(tests): Add optional cutoff to toxgen (#4243) by @sentrivana +- toxgen: Retry & fail if we fail to fetch PyPI data (#4251) by @sentrivana +- build(deps): bump actions/create-github-app-token from 1.12.0 to 2.0.2 (#4248) by @dependabot +- Trying to prevent the grpc setup from being flaky (#4233) by @antonpirker +- feat(breadcrumbs): add `_meta` information for truncation of breadcrumbs (#4007) by @shellmayr +- tests: Move django under toxgen (#4238) by @sentrivana +- fix: Handle JSONDecodeError gracefully in StarletteRequestExtractor (#4226) by @moodix +- fix(asyncio): Remove shutdown handler (#4237) by @sentrivana + ## 2.25.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 2f575d3097..9c137d70a9 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.25.1" +release = "2.26.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index c0f6ff66c6..19d39acdc0 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -966,4 +966,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.25.1" +VERSION = "2.26.0" diff --git a/setup.py b/setup.py index 6de160dcfb..6c33887cf5 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.25.1", + version="2.26.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From e71ccbf19f644fe7928db37f6e4a09e1febbc4e2 Mon Sep 17 00:00:00 2001 From: Daniel Szoke Date: Mon, 14 Apr 2025 17:56:14 +0200 Subject: [PATCH 250/591] fix(logging): Send raw logging parameters This reverts commit 4c9731bbe68b6523cccec73fb764e04e61e441cb, adding tests to ensure the correct behavior going forward. That commit caused a regression when `record.args` contains a dictionary. Because we iterate over `record.args`, that change caused us to only send the dictionary's keys, not the values. A more robust fix for #3660 will be to send the formatted message in the [`formatted` field](https://develop.sentry.dev/sdk/data-model/event-payloads/message/) (which we have not been doing yet). I will open a follow-up PR to do this. Fixes #4267 --- sentry_sdk/integrations/logging.py | 6 +---- tests/integrations/logging/test_logging.py | 30 ++++++++++++++++++++++ 2 files changed, 31 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 1fbecb2e08..26ee957b27 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -265,11 +265,7 @@ def _emit(self, record): else: event["logentry"] = { "message": to_string(record.msg), - "params": ( - tuple(str(arg) if arg is None else arg for arg in record.args) - if record.args - else () - ), + "params": record.args, } event["extra"] = self._extra_from_record(record) diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 8c325bc86c..5b48540bb0 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -234,3 +234,33 @@ def test_ignore_logger_wildcard(sentry_init, capture_events): (event,) = events assert event["logentry"]["message"] == "hi" + + +def test_logging_dictionary_interpolation(sentry_init, capture_events): + """Here we test an entire dictionary being interpolated into the log message.""" + sentry_init(integrations=[LoggingIntegration()], default_integrations=False) + events = capture_events() + + logger.error("this is a log with a dictionary %s", {"foo": "bar"}) + + (event,) = events + assert event["logentry"]["message"] == "this is a log with a dictionary %s" + assert event["logentry"]["params"] == {"foo": "bar"} + + +def test_logging_dictionary_args(sentry_init, capture_events): + """Here we test items from a dictionary being interpolated into the log message.""" + sentry_init(integrations=[LoggingIntegration()], default_integrations=False) + events = capture_events() + + logger.error( + "the value of foo is %(foo)s, and the value of bar is %(bar)s", + {"foo": "bar", "bar": "baz"}, + ) + + (event,) = events + assert ( + event["logentry"]["message"] + == "the value of foo is %(foo)s, and the value of bar is %(bar)s" + ) + assert event["logentry"]["params"] == {"foo": "bar", "bar": "baz"} From 296e288e437b3e690bb7485f1d062f7f33ac373b Mon Sep 17 00:00:00 2001 From: Daniel Szoke Date: Mon, 14 Apr 2025 18:23:06 +0200 Subject: [PATCH 251/591] feat(logging): Add formatted message to log events Send the formatted log event to Sentry in the [`formatted` field](https://develop.sentry.dev/sdk/data-model/event-payloads/message/). This builds on #4291, providing a more robust fix for #3660. --- sentry_sdk/integrations/logging.py | 2 ++ tests/integrations/logging/test_logging.py | 19 +++++++++++++++++++ 2 files changed, 21 insertions(+) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 26ee957b27..ec13c86c6e 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -259,11 +259,13 @@ def _emit(self, record): event["logentry"] = { "message": msg, + "formatted": record.getMessage(), "params": (), } else: event["logentry"] = { + "formatted": record.getMessage(), "message": to_string(record.msg), "params": record.args, } diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 5b48540bb0..c08e960c00 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -26,6 +26,7 @@ def test_logging_works_with_many_loggers(sentry_init, capture_events, logger): assert event["level"] == "fatal" assert not event["logentry"]["params"] assert event["logentry"]["message"] == "LOL" + assert event["logentry"]["formatted"] == "LOL" assert any(crumb["message"] == "bread" for crumb in event["breadcrumbs"]["values"]) @@ -112,6 +113,7 @@ def test_logging_level(sentry_init, capture_events): (event,) = events assert event["level"] == "error" assert event["logentry"]["message"] == "hi" + assert event["logentry"]["formatted"] == "hi" del events[:] @@ -152,6 +154,7 @@ def test_custom_log_level_names(sentry_init, capture_events): assert events assert events[0]["level"] == sentry_level assert events[0]["logentry"]["message"] == "Trying level %s" + assert events[0]["logentry"]["formatted"] == f"Trying level {logging_level}" assert events[0]["logentry"]["params"] == [logging_level] del events[:] @@ -177,6 +180,7 @@ def filter(self, record): (event,) = events assert event["logentry"]["message"] == "hi" + assert event["logentry"]["formatted"] == "hi" def test_logging_captured_warnings(sentry_init, capture_events, recwarn): @@ -198,10 +202,16 @@ def test_logging_captured_warnings(sentry_init, capture_events, recwarn): assert events[0]["level"] == "warning" # Captured warnings start with the path where the warning was raised assert "UserWarning: first" in events[0]["logentry"]["message"] + assert "UserWarning: first" in events[0]["logentry"]["formatted"] + # For warnings, the message and formatted message are the same + assert events[0]["logentry"]["message"] == events[0]["logentry"]["formatted"] assert events[0]["logentry"]["params"] == [] assert events[1]["level"] == "warning" assert "UserWarning: second" in events[1]["logentry"]["message"] + assert "UserWarning: second" in events[1]["logentry"]["formatted"] + # For warnings, the message and formatted message are the same + assert events[1]["logentry"]["message"] == events[1]["logentry"]["formatted"] assert events[1]["logentry"]["params"] == [] # Using recwarn suppresses the "third" warning in the test output @@ -234,6 +244,7 @@ def test_ignore_logger_wildcard(sentry_init, capture_events): (event,) = events assert event["logentry"]["message"] == "hi" + assert event["logentry"]["formatted"] == "hi" def test_logging_dictionary_interpolation(sentry_init, capture_events): @@ -245,6 +256,10 @@ def test_logging_dictionary_interpolation(sentry_init, capture_events): (event,) = events assert event["logentry"]["message"] == "this is a log with a dictionary %s" + assert ( + event["logentry"]["formatted"] + == "this is a log with a dictionary {'foo': 'bar'}" + ) assert event["logentry"]["params"] == {"foo": "bar"} @@ -263,4 +278,8 @@ def test_logging_dictionary_args(sentry_init, capture_events): event["logentry"]["message"] == "the value of foo is %(foo)s, and the value of bar is %(bar)s" ) + assert ( + event["logentry"]["formatted"] + == "the value of foo is bar, and the value of bar is baz" + ) assert event["logentry"]["params"] == {"foo": "bar", "bar": "baz"} From 706d2d29e68848a3cb085f043287d908255344b5 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Apr 2025 12:14:49 +0200 Subject: [PATCH 252/591] Revert "chore: Deprecate `same_process_as_parent` (#4244)" (#4290) This reverts commit e05ed0aa62cfe2c992b26b07c64c3148f837a609. `same_process_as_parent` is `True` by default, so we actually don't have a way of detecting whether this was set explicitly by the user or not. Removing the deprecation altogether -- no one's using this. Closes https://github.com/getsentry/sentry-python/issues/4289 --- sentry_sdk/tracing.py | 7 ------- 1 file changed, 7 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index ab1a7a8fdf..13d9f63d5e 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -323,13 +323,6 @@ def __init__( self.scope = self.scope or hub.scope - if same_process_as_parent is not None: - warnings.warn( - "The `same_process_as_parent` parameter is deprecated.", - DeprecationWarning, - stacklevel=2, - ) - if start_timestamp is None: start_timestamp = datetime.now(timezone.utc) elif isinstance(start_timestamp, float): From 2d392af3ea6da91ddbdde55d18e15c24dce6b59b Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 15 Apr 2025 12:30:05 +0200 Subject: [PATCH 253/591] fix: Data leak in ThreadingIntegration between threads (#4281) It is possible to leak data from started threads into the main thread via the scopes. (Because the same scope object from the main thread could be changed in the started thread.) This change always makes a fork (copy) of the scopes of the main thread before it propagates those scopes into the started thread. --- sentry_sdk/integrations/threading.py | 33 +++++- tests/integrations/django/asgi/test_asgi.py | 22 +++- .../integrations/threading/test_threading.py | 101 ++++++++++++++++++ 3 files changed, 151 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/threading.py b/sentry_sdk/integrations/threading.py index 5de736e23b..9c99a8e896 100644 --- a/sentry_sdk/integrations/threading.py +++ b/sentry_sdk/integrations/threading.py @@ -1,4 +1,5 @@ import sys +import warnings from functools import wraps from threading import Thread, current_thread @@ -49,6 +50,15 @@ def setup_once(): # type: () -> None old_start = Thread.start + try: + from django import VERSION as django_version # noqa: N811 + import channels # type: ignore[import-not-found] + + channels_version = channels.__version__ + except ImportError: + django_version = None + channels_version = None + @wraps(old_start) def sentry_start(self, *a, **kw): # type: (Thread, *Any, **Any) -> Any @@ -57,8 +67,27 @@ def sentry_start(self, *a, **kw): return old_start(self, *a, **kw) if integration.propagate_scope: - isolation_scope = sentry_sdk.get_isolation_scope() - current_scope = sentry_sdk.get_current_scope() + if ( + sys.version_info < (3, 9) + and channels_version is not None + and channels_version < "4.0.0" + and django_version is not None + and django_version >= (3, 0) + and django_version < (4, 0) + ): + warnings.warn( + "There is a known issue with Django channels 2.x and 3.x when using Python 3.8 or older. " + "(Async support is emulated using threads and some Sentry data may be leaked between those threads.) " + "Please either upgrade to Django channels 4.0+, use Django's async features " + "available in Django 3.1+ instead of Django channels, or upgrade to Python 3.9+.", + stacklevel=2, + ) + isolation_scope = sentry_sdk.get_isolation_scope() + current_scope = sentry_sdk.get_current_scope() + + else: + isolation_scope = sentry_sdk.get_isolation_scope().fork() + current_scope = sentry_sdk.get_current_scope().fork() else: isolation_scope = None current_scope = None diff --git a/tests/integrations/django/asgi/test_asgi.py b/tests/integrations/django/asgi/test_asgi.py index 063aed63ad..82eae30b1d 100644 --- a/tests/integrations/django/asgi/test_asgi.py +++ b/tests/integrations/django/asgi/test_asgi.py @@ -38,9 +38,25 @@ async def test_basic(sentry_init, capture_events, application): events = capture_events() - comm = HttpCommunicator(application, "GET", "/view-exc?test=query") - response = await comm.get_response() - await comm.wait() + import channels # type: ignore[import-not-found] + + if ( + sys.version_info < (3, 9) + and channels.__version__ < "4.0.0" + and django.VERSION >= (3, 0) + and django.VERSION < (4, 0) + ): + # We emit a UserWarning for channels 2.x and 3.x on Python 3.8 and older + # because the async support was not really good back then and there is a known issue. + # See the TreadingIntegration for details. + with pytest.warns(UserWarning): + comm = HttpCommunicator(application, "GET", "/view-exc?test=query") + response = await comm.get_response() + await comm.wait() + else: + comm = HttpCommunicator(application, "GET", "/view-exc?test=query") + response = await comm.get_response() + await comm.wait() assert response["status"] == 500 diff --git a/tests/integrations/threading/test_threading.py b/tests/integrations/threading/test_threading.py index 0d14fae352..4395891d62 100644 --- a/tests/integrations/threading/test_threading.py +++ b/tests/integrations/threading/test_threading.py @@ -1,5 +1,6 @@ import gc from concurrent import futures +from textwrap import dedent from threading import Thread import pytest @@ -172,3 +173,103 @@ def target(): assert Thread.run.__qualname__ == original_run.__qualname__ assert t.run.__name__ == "run" assert t.run.__qualname__ == original_run.__qualname__ + + +@pytest.mark.parametrize( + "propagate_scope", + (True, False), + ids=["propagate_scope=True", "propagate_scope=False"], +) +def test_scope_data_not_leaked_in_threads(sentry_init, propagate_scope): + sentry_init( + integrations=[ThreadingIntegration(propagate_scope=propagate_scope)], + ) + + sentry_sdk.set_tag("initial_tag", "initial_value") + initial_iso_scope = sentry_sdk.get_isolation_scope() + + def do_some_work(): + # check if we have the initial scope data propagated into the thread + if propagate_scope: + assert sentry_sdk.get_isolation_scope()._tags == { + "initial_tag": "initial_value" + } + else: + assert sentry_sdk.get_isolation_scope()._tags == {} + + # change data in isolation scope in thread + sentry_sdk.set_tag("thread_tag", "thread_value") + + t = Thread(target=do_some_work) + t.start() + t.join() + + # check if the initial scope data is not modified by the started thread + assert initial_iso_scope._tags == { + "initial_tag": "initial_value" + }, "The isolation scope in the main thread should not be modified by the started thread." + + +@pytest.mark.parametrize( + "propagate_scope", + (True, False), + ids=["propagate_scope=True", "propagate_scope=False"], +) +def test_spans_from_multiple_threads( + sentry_init, capture_events, render_span_tree, propagate_scope +): + sentry_init( + traces_sample_rate=1.0, + integrations=[ThreadingIntegration(propagate_scope=propagate_scope)], + ) + events = capture_events() + + def do_some_work(number): + with sentry_sdk.start_span( + op=f"inner-run-{number}", name=f"Thread: child-{number}" + ): + pass + + threads = [] + + with sentry_sdk.start_transaction(op="outer-trx"): + for number in range(5): + with sentry_sdk.start_span( + op=f"outer-submit-{number}", name="Thread: main" + ): + t = Thread(target=do_some_work, args=(number,)) + t.start() + threads.append(t) + + for t in threads: + t.join() + + (event,) = events + if propagate_scope: + assert render_span_tree(event) == dedent( + """\ + - op="outer-trx": description=null + - op="outer-submit-0": description="Thread: main" + - op="inner-run-0": description="Thread: child-0" + - op="outer-submit-1": description="Thread: main" + - op="inner-run-1": description="Thread: child-1" + - op="outer-submit-2": description="Thread: main" + - op="inner-run-2": description="Thread: child-2" + - op="outer-submit-3": description="Thread: main" + - op="inner-run-3": description="Thread: child-3" + - op="outer-submit-4": description="Thread: main" + - op="inner-run-4": description="Thread: child-4"\ +""" + ) + + elif not propagate_scope: + assert render_span_tree(event) == dedent( + """\ + - op="outer-trx": description=null + - op="outer-submit-0": description="Thread: main" + - op="outer-submit-1": description="Thread: main" + - op="outer-submit-2": description="Thread: main" + - op="outer-submit-3": description="Thread: main" + - op="outer-submit-4": description="Thread: main"\ +""" + ) From b2693f4b3e1442330e991caaf5d0c1c08f634069 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 15 Apr 2025 12:42:58 +0200 Subject: [PATCH 254/591] ref(logging): Clarify separate warnings case is for Python <3.11 (#4296) The way the code was written before this change made it look like log records from the `warnings` module were always being handled by a separate code path. In fact, this separate path is only used for Python 3.10 and below. This change makes it clear that the branch is version specific. That way, when we eventually stop supporting 3.10, it is clear that we can delete this separate block. Depends on: - #4292 - #4291 --- sentry_sdk/integrations/logging.py | 39 +++++++++++++++--------------- 1 file changed, 19 insertions(+), 20 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index ec13c86c6e..bf538ac7c7 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -1,4 +1,5 @@ import logging +import sys from datetime import datetime, timezone from fnmatch import fnmatch @@ -248,27 +249,25 @@ def _emit(self, record): event["level"] = level # type: ignore[typeddict-item] event["logger"] = record.name - # Log records from `warnings` module as separate issues - record_captured_from_warnings_module = ( - record.name == "py.warnings" and record.msg == "%s" - ) - if record_captured_from_warnings_module: - # use the actual message and not "%s" as the message - # this prevents grouping all warnings under one "%s" issue - msg = record.args[0] # type: ignore - - event["logentry"] = { - "message": msg, - "formatted": record.getMessage(), - "params": (), - } - + if ( + sys.version_info < (3, 11) + and record.name == "py.warnings" + and record.msg == "%s" + ): + # warnings module on Python 3.10 and below sets record.msg to "%s" + # and record.args[0] to the actual warning message. + # This was fixed in https://github.com/python/cpython/pull/30975. + message = record.args[0] + params = () else: - event["logentry"] = { - "formatted": record.getMessage(), - "message": to_string(record.msg), - "params": record.args, - } + message = record.msg + params = record.args + + event["logentry"] = { + "message": to_string(message), + "formatted": record.getMessage(), + "params": params, + } event["extra"] = self._extra_from_record(record) From d552808330c873958b9d0803349a0e662e27d959 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 15 Apr 2025 11:13:44 +0000 Subject: [PATCH 255/591] release: 2.26.1 --- CHANGELOG.md | 10 ++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 13 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5327b323a2..97343dc0fc 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,15 @@ # Changelog +## 2.26.1 + +### Various fixes & improvements + +- ref(logging): Clarify separate warnings case is for Python <3.11 (#4296) by @szokeasaurusrex +- fix: Data leak in ThreadingIntegration between threads (#4281) by @antonpirker +- Revert "chore: Deprecate `same_process_as_parent` (#4244)" (#4290) by @sentrivana +- feat(logging): Add formatted message to log events (#4292) by @szokeasaurusrex +- fix(logging): Send raw logging parameters (#4291) by @szokeasaurusrex + ## 2.26.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 9c137d70a9..629b5b9eaa 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.26.0" +release = "2.26.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 19d39acdc0..3802980b82 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -966,4 +966,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.26.0" +VERSION = "2.26.1" diff --git a/setup.py b/setup.py index 6c33887cf5..62f4867b35 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.26.0", + version="2.26.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From ec050c0de436b9d4afb495df79f5d6ae72bec16f Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 15 Apr 2025 13:16:01 +0200 Subject: [PATCH 256/591] Updated changelog --- CHANGELOG.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 97343dc0fc..bb49ed54ca 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,11 +4,11 @@ ### Various fixes & improvements -- ref(logging): Clarify separate warnings case is for Python <3.11 (#4296) by @szokeasaurusrex -- fix: Data leak in ThreadingIntegration between threads (#4281) by @antonpirker -- Revert "chore: Deprecate `same_process_as_parent` (#4244)" (#4290) by @sentrivana -- feat(logging): Add formatted message to log events (#4292) by @szokeasaurusrex +- fix(threading): Data leak in ThreadingIntegration between threads (#4281) by @antonpirker +- fix(logging): Clarify separate warnings case is for Python <3.11 (#4296) by @szokeasaurusrex +- fix(logging): Add formatted message to log events (#4292) by @szokeasaurusrex - fix(logging): Send raw logging parameters (#4291) by @szokeasaurusrex +- fix: Revert "chore: Deprecate `same_process_as_parent` (#4244)" (#4290) by @sentrivana ## 2.26.0 From 12b3414894e1b3b7c3fa248d274fa5be9b6b939f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Apr 2025 13:45:43 +0200 Subject: [PATCH 257/591] tests: Update tox.ini (#4297) Regular update --- tox.ini | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/tox.ini b/tox.ini index c04691e2ac..e1e7c676f3 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-08T10:33:11.499210+00:00 +# Last generated: 2025-04-15T10:30:18.609730+00:00 [tox] requires = @@ -157,7 +157,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.11.3 + {py3.9,py3.12,py3.13}-pymongo-v4.12.0 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -175,11 +175,11 @@ envlist = {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 - {py3.9,py3.12,py3.13}-openfeature-v0.8.0 + {py3.9,py3.12,py3.13}-openfeature-v0.8.1 {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 - {py3.7,py3.12,py3.13}-statsig-v0.57.2 + {py3.7,py3.12,py3.13}-statsig-v0.57.3 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -202,7 +202,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.227.7 {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.263.2 + {py3.9,py3.12,py3.13}-strawberry-v0.264.0 # ~~~ Network ~~~ @@ -210,6 +210,7 @@ envlist = {py3.7,py3.9,py3.10}-grpc-v1.44.0 {py3.7,py3.10,py3.11}-grpc-v1.58.3 {py3.9,py3.12,py3.13}-grpc-v1.71.0 + {py3.9,py3.12,py3.13}-grpc-v1.72.0rc1 # ~~~ Tasks ~~~ @@ -245,7 +246,7 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.46.1 + {py3.9,py3.12,py3.13}-starlette-v0.46.2 # ~~~ Web 2 ~~~ @@ -519,7 +520,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.11.3: pymongo==4.11.3 + pymongo-v4.12.0: pymongo==4.12.0 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -538,11 +539,11 @@ deps = launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 - openfeature-v0.8.0: openfeature-sdk==0.8.0 + openfeature-v0.8.1: openfeature-sdk==0.8.1 statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 - statsig-v0.57.2: statsig==0.57.2 + statsig-v0.57.3: statsig==0.57.3 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -574,7 +575,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.263.2: strawberry-graphql[fastapi,flask]==0.263.2 + strawberry-v0.264.0: strawberry-graphql[fastapi,flask]==0.264.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.227.7: pydantic<2.11 @@ -586,6 +587,7 @@ deps = grpc-v1.44.0: grpcio==1.44.0 grpc-v1.58.3: grpcio==1.58.3 grpc-v1.71.0: grpcio==1.71.0 + grpc-v1.72.0rc1: grpcio==1.72.0rc1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -657,7 +659,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.46.1: starlette==0.46.1 + starlette-v0.46.2: starlette==0.46.2 starlette: pytest-asyncio starlette: python-multipart starlette: requests From fbf43bd9fdf748b0677bb82ddcdeaad0bc2776dc Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Apr 2025 13:56:54 +0200 Subject: [PATCH 258/591] toxgen: Add huey (#4298) --- scripts/populate_tox/populate_tox.py | 1 - tox.ini | 12 +++++++++++- 2 files changed, 11 insertions(+), 2 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 58dbed0308..8f588a1b26 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -78,7 +78,6 @@ "fastapi", "gcp", "httpx", - "huey", "huggingface_hub", "langchain", "langchain_notiktoken", diff --git a/tox.ini b/tox.ini index e1e7c676f3..0cc8a0cce2 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-15T10:30:18.609730+00:00 +# Last generated: 2025-04-15T11:48:52.985806+00:00 [tox] requires = @@ -223,6 +223,11 @@ envlist = {py3.7,py3.10,py3.11}-dramatiq-v1.15.0 {py3.8,py3.12,py3.13}-dramatiq-v1.17.1 + {py3.6,py3.7}-huey-v2.1.3 + {py3.6,py3.7}-huey-v2.2.0 + {py3.6,py3.7}-huey-v2.3.2 + {py3.6,py3.11,py3.12}-huey-v2.5.3 + {py3.8,py3.9}-spark-v3.0.3 {py3.8,py3.9}-spark-v3.2.4 {py3.8,py3.10,py3.11}-spark-v3.4.4 @@ -607,6 +612,11 @@ deps = dramatiq-v1.15.0: dramatiq==1.15.0 dramatiq-v1.17.1: dramatiq==1.17.1 + huey-v2.1.3: huey==2.1.3 + huey-v2.2.0: huey==2.2.0 + huey-v2.3.2: huey==2.3.2 + huey-v2.5.3: huey==2.5.3 + spark-v3.0.3: pyspark==3.0.3 spark-v3.2.4: pyspark==3.2.4 spark-v3.4.4: pyspark==3.4.4 From 08514584aa31d285a1eebefe3a5cc2a4a40ed5ff Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Apr 2025 15:00:13 +0200 Subject: [PATCH 259/591] toxgen: Add huggingface_hub (#4299) Also fixes ``` Repository Not Found for url: https://huggingface.co/api/models/some-model. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. For more details, see https://huggingface.co/docs/huggingface_hub/authentication Invalid username or password. FAILED tests/integrations/huggingface_hub/test_huggingface_hub.py::test_span_origin - huggingface_hub.errors.RepositoryNotFoundError: 401 Client Error. (Request ID: Root=1-67fe4547-10b0ce8f541a41c37ead3b2a;afe45d5d-3af1-45cd-a39a-c8ef4a5211c3) ``` which started popping up on huggingface_hub 0.30. --- .github/workflows/test-integrations-ai.yml | 2 +- scripts/populate_tox/populate_tox.py | 1 - .../huggingface_hub/test_huggingface_hub.py | 12 +++++------- tox.ini | 14 ++++++++++++++ 4 files changed, 20 insertions(+), 9 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 10171ce196..e497ba4280 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -104,7 +104,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.11","3.12"] + python-version: ["3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 8f588a1b26..912cc15bd5 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -78,7 +78,6 @@ "fastapi", "gcp", "httpx", - "huggingface_hub", "langchain", "langchain_notiktoken", "openai", diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index e017ce2449..090b0e4f3e 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -1,4 +1,5 @@ import itertools +from unittest import mock import pytest from huggingface_hub import ( @@ -9,8 +10,6 @@ from sentry_sdk import start_transaction from sentry_sdk.integrations.huggingface_hub import HuggingfaceHubIntegration -from unittest import mock # python 3.3 and above - def mock_client_post(client, post_mock): # huggingface-hub==0.28.0 deprecates the `post` method @@ -33,7 +32,7 @@ def test_nonstreaming_chat_completion( ) events = capture_events() - client = InferenceClient("some-model") + client = InferenceClient() if details_arg: post_mock = mock.Mock( return_value=b"""[{ @@ -92,7 +91,7 @@ def test_streaming_chat_completion( ) events = capture_events() - client = InferenceClient("some-model") + client = InferenceClient() post_mock = mock.Mock( return_value=[ @@ -116,7 +115,6 @@ def test_streaming_chat_completion( ) ) assert len(response) == 2 - print(response) if details_arg: assert response[0].token.text + response[1].token.text == "the model response" else: @@ -142,7 +140,7 @@ def test_bad_chat_completion(sentry_init, capture_events): sentry_init(integrations=[HuggingfaceHubIntegration()], traces_sample_rate=1.0) events = capture_events() - client = InferenceClient("some-model") + client = InferenceClient() post_mock = mock.Mock(side_effect=OverloadedError("The server is overloaded")) mock_client_post(client, post_mock) @@ -160,7 +158,7 @@ def test_span_origin(sentry_init, capture_events): ) events = capture_events() - client = InferenceClient("some-model") + client = InferenceClient() post_mock = mock.Mock( return_value=[ b"""data:{ diff --git a/tox.ini b/tox.ini index 0cc8a0cce2..50c4dcf4ac 100644 --- a/tox.ini +++ b/tox.ini @@ -151,6 +151,13 @@ envlist = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ AI ~~~ + {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 + {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 + + # ~~~ DBs ~~~ {py3.7,py3.11,py3.12}-clickhouse_driver-v0.2.9 @@ -519,6 +526,13 @@ deps = # These come from the populate_tox.py script. Eventually we should move all # integration tests there. + # ~~~ AI ~~~ + huggingface_hub-v0.22.2: huggingface_hub==0.22.2 + huggingface_hub-v0.25.2: huggingface_hub==0.25.2 + huggingface_hub-v0.28.1: huggingface_hub==0.28.1 + huggingface_hub-v0.30.2: huggingface_hub==0.30.2 + + # ~~~ DBs ~~~ clickhouse_driver-v0.2.9: clickhouse-driver==0.2.9 From e6c8798fd5d9246f60219349cdc4416a58285be9 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Apr 2025 16:51:37 +0200 Subject: [PATCH 260/591] toxgen: Migrate fastapi (#4302) With this we've migrated the whole Web 1 group, yay! So the whole `-latest` category is gone for Web 1, too. Also removed some `pytest.mark.asyncio`s on sync tests. --- .github/workflows/test-integrations-web-1.yml | 89 ------------------- scripts/populate_tox/config.py | 24 ++++- scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 14 --- tests/integrations/fastapi/test_fastapi.py | 3 - tox.ini | 35 ++++---- 6 files changed, 43 insertions(+), 123 deletions(-) diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 6d3e62a78a..ac364ccfc1 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -22,95 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_1-latest: - name: Web 1 (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.8","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - services: - postgres: - image: postgres - env: - POSTGRES_PASSWORD: sentry - # Set health checks to wait until postgres has started - options: >- - --health-cmd pg_isready - --health-interval 10s - --health-timeout 5s - --health-retries 5 - # Maps tcp port 5432 on service container to the host - ports: - - 5432:5432 - env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} - SENTRY_PYTHON_TEST_POSTGRES_USER: postgres - SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v4.2.2 - - uses: actions/setup-python@v5 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test django latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-django-latest" - - name: Test flask latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-flask-latest" - - name: Test starlette latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-starlette-latest" - - name: Test fastapi latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-fastapi-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugin: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-web_1-pinned: name: Web 1 (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 0bacfcaa7b..9496ef544a 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -55,6 +55,27 @@ "package": "falcon", "python": "<3.13", }, + "fastapi": { + "package": "fastapi", + "deps": { + "*": [ + "httpx", + "pytest-asyncio", + "python-multipart", + "requests", + "anyio<4", + ], + # There's an incompatibility between FastAPI's TestClient, which is + # actually Starlette's TestClient, which is actually httpx's Client. + # httpx dropped a deprecated Client argument in 0.28.0, Starlette + # dropped it from its TestClient in 0.37.2, and FastAPI only pinned + # Starlette>=0.37.2 from version 0.110.1 onwards -- so for older + # FastAPI versions we use older httpx which still supports the + # deprecated argument. + "<0.110.1": ["httpx<0.28.0"], + "py3.6": ["aiocontextvars"], + }, + }, "flask": { "package": "flask", "deps": { @@ -137,7 +158,8 @@ "jinja2", "httpx", ], - "<0.37": ["httpx<0.28.0"], + # See the comment on FastAPI's httpx bound for more info + "<0.37.2": ["httpx<0.28.0"], "<0.15": ["jinja2<3.1"], "py3.6": ["aiocontextvars"], }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 912cc15bd5..d51497c21e 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -75,7 +75,6 @@ "boto3", "chalice", "cohere", - "fastapi", "gcp", "httpx", "langchain", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index e599f45436..7b1d83f87a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -80,10 +80,6 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5 {py3.9,py3.11,py3.12}-cohere-latest - # FastAPI - {py3.7,py3.10}-fastapi-v{0.79} - {py3.8,py3.12,py3.13}-fastapi-latest - # GCP {py3.7}-gcp @@ -252,16 +248,6 @@ deps = cohere-v5: cohere~=5.3.3 cohere-latest: cohere - # FastAPI - fastapi: httpx - # (this is a dependency of httpx) - fastapi: anyio<4.0.0 - fastapi: pytest-asyncio - fastapi: python-multipart - fastapi: requests - fastapi-v{0.79}: fastapi~=0.79.0 - fastapi-latest: fastapi - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index 4cb9ea1716..95838b1009 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -247,7 +247,6 @@ async def _error(request: Request): assert event["request"]["headers"]["authorization"] == "[Filtered]" -@pytest.mark.asyncio def test_response_status_code_ok_in_transaction_context(sentry_init, capture_envelopes): """ Tests that the response status code is added to the transaction "response" context. @@ -276,7 +275,6 @@ def test_response_status_code_ok_in_transaction_context(sentry_init, capture_env assert transaction["contexts"]["response"]["status_code"] == 200 -@pytest.mark.asyncio def test_response_status_code_error_in_transaction_context( sentry_init, capture_envelopes, @@ -313,7 +311,6 @@ def test_response_status_code_error_in_transaction_context( assert transaction["contexts"]["response"]["status_code"] == 500 -@pytest.mark.asyncio def test_response_status_code_not_found_in_transaction_context( sentry_init, capture_envelopes, diff --git a/tox.ini b/tox.ini index 50c4dcf4ac..47bce49879 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-15T11:48:52.985806+00:00 +# Last generated: 2025-04-15T14:38:12.763407+00:00 [tox] requires = @@ -80,10 +80,6 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5 {py3.9,py3.11,py3.12}-cohere-latest - # FastAPI - {py3.7,py3.10}-fastapi-v{0.79} - {py3.8,py3.12,py3.13}-fastapi-latest - # GCP {py3.7}-gcp @@ -260,6 +256,11 @@ envlist = {py3.8,py3.11,py3.12}-starlette-v0.36.3 {py3.9,py3.12,py3.13}-starlette-v0.46.2 + {py3.6,py3.9,py3.10}-fastapi-v0.79.1 + {py3.7,py3.10,py3.11}-fastapi-v0.91.0 + {py3.7,py3.10,py3.11}-fastapi-v0.103.2 + {py3.8,py3.12,py3.13}-fastapi-v0.115.12 + # ~~~ Web 2 ~~~ {py3.6,py3.7}-bottle-v0.12.25 @@ -394,16 +395,6 @@ deps = cohere-v5: cohere~=5.3.3 cohere-latest: cohere - # FastAPI - fastapi: httpx - # (this is a dependency of httpx) - fastapi: anyio<4.0.0 - fastapi: pytest-asyncio - fastapi: python-multipart - fastapi: requests - fastapi-v{0.79}: fastapi~=0.79.0 - fastapi-latest: fastapi - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 @@ -695,6 +686,20 @@ deps = starlette-v0.36.3: httpx<0.28.0 py3.6-starlette: aiocontextvars + fastapi-v0.79.1: fastapi==0.79.1 + fastapi-v0.91.0: fastapi==0.91.0 + fastapi-v0.103.2: fastapi==0.103.2 + fastapi-v0.115.12: fastapi==0.115.12 + fastapi: httpx + fastapi: pytest-asyncio + fastapi: python-multipart + fastapi: requests + fastapi: anyio<4 + fastapi-v0.79.1: httpx<0.28.0 + fastapi-v0.91.0: httpx<0.28.0 + fastapi-v0.103.2: httpx<0.28.0 + py3.6-fastapi: aiocontextvars + # ~~~ Web 2 ~~~ bottle-v0.12.25: bottle==0.12.25 From 863228154f231338391cc228ba7f0f31fc20ac87 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 16 Apr 2025 09:40:58 +0200 Subject: [PATCH 261/591] toxgen: Add cohere (#4304) --- scripts/populate_tox/config.py | 4 ++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 8 -------- sentry_sdk/integrations/__init__.py | 1 + tox.ini | 20 +++++++++++--------- 5 files changed, 16 insertions(+), 18 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 9496ef544a..f3f1ba0092 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -29,6 +29,10 @@ "clickhouse_driver": { "package": "clickhouse-driver", }, + "cohere": { + "package": "cohere", + "python": ">=3.9", + }, "django": { "package": "django", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index d51497c21e..b274e8c077 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -74,7 +74,6 @@ "beam", "boto3", "chalice", - "cohere", "gcp", "httpx", "langchain", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 7b1d83f87a..380a80f690 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -76,10 +76,6 @@ envlist = # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context - # Cohere - {py3.9,py3.11,py3.12}-cohere-v5 - {py3.9,py3.11,py3.12}-cohere-latest - # GCP {py3.7}-gcp @@ -244,10 +240,6 @@ deps = chalice-v1.16: chalice~=1.16.0 chalice-latest: chalice - # Cohere - cohere-v5: cohere~=5.3.3 - cohere-latest: cohere - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 9bff264752..118289950c 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -131,6 +131,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "celery": (4, 4, 7), "chalice": (1, 16, 0), "clickhouse_driver": (0, 2, 0), + "cohere": (5, 4, 0), "django": (1, 8), "dramatiq": (1, 9), "falcon": (1, 4), diff --git a/tox.ini b/tox.ini index 47bce49879..45627b83ec 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-15T14:38:12.763407+00:00 +# Last generated: 2025-04-15T15:09:46.980440+00:00 [tox] requires = @@ -76,10 +76,6 @@ envlist = # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context - # Cohere - {py3.9,py3.11,py3.12}-cohere-v5 - {py3.9,py3.11,py3.12}-cohere-latest - # GCP {py3.7}-gcp @@ -148,6 +144,11 @@ envlist = # integration tests there. # ~~~ AI ~~~ + {py3.9,py3.10,py3.11}-cohere-v5.4.0 + {py3.9,py3.11,py3.12}-cohere-v5.9.4 + {py3.9,py3.11,py3.12}-cohere-v5.13.9 + {py3.9,py3.11,py3.12}-cohere-v5.15.0 + {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 @@ -391,10 +392,6 @@ deps = chalice-v1.16: chalice~=1.16.0 chalice-latest: chalice - # Cohere - cohere-v5: cohere~=5.3.3 - cohere-latest: cohere - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 @@ -518,6 +515,11 @@ deps = # integration tests there. # ~~~ AI ~~~ + cohere-v5.4.0: cohere==5.4.0 + cohere-v5.9.4: cohere==5.9.4 + cohere-v5.13.9: cohere==5.13.9 + cohere-v5.15.0: cohere==5.15.0 + huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.25.2: huggingface_hub==0.25.2 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 From 815de9f9175317c2d1d31bc6ccba9fee47273d79 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 17 Apr 2025 15:13:18 +0200 Subject: [PATCH 262/591] toxgen: Remove unused code and rerun (#4313) Noticed some unused code in toxgen, probably the result of a bad merge? --- scripts/populate_tox/populate_tox.py | 7 ------- tox.ini | 20 +++++++++++--------- 2 files changed, 11 insertions(+), 16 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index b274e8c077..11ea94c0f4 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -234,13 +234,6 @@ def get_supported_releases( integration, pypi_data["releases"], older_than ) - # Determine Python support - expected_python_versions = TEST_SUITE_CONFIG[integration].get("python") - if expected_python_versions: - expected_python_versions = SpecifierSet(expected_python_versions) - else: - expected_python_versions = SpecifierSet(f">={MIN_PYTHON_VERSION}") - def _supports_lowest(release: Version) -> bool: time.sleep(PYPI_COOLDOWN) # don't DoS PYPI diff --git a/tox.ini b/tox.ini index 45627b83ec..9497708ff8 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-15T15:09:46.980440+00:00 +# Last generated: 2025-04-17T11:01:25.976599+00:00 [tox] requires = @@ -177,6 +177,7 @@ envlist = {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 + {py3.8,py3.12,py3.13}-launchdarkly-v9.11.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.1 @@ -204,9 +205,9 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.227.7 - {py3.8,py3.11,py3.12}-strawberry-v0.245.0 - {py3.9,py3.12,py3.13}-strawberry-v0.264.0 + {py3.8,py3.11,py3.12}-strawberry-v0.228.0 + {py3.8,py3.12,py3.13}-strawberry-v0.247.2 + {py3.9,py3.12,py3.13}-strawberry-v0.265.1 # ~~~ Network ~~~ @@ -549,6 +550,7 @@ deps = launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 + launchdarkly-v9.11.0: launchdarkly-server-sdk==9.11.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.1: openfeature-sdk==0.8.1 @@ -585,13 +587,13 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.227.7: strawberry-graphql[fastapi,flask]==0.227.7 - strawberry-v0.245.0: strawberry-graphql[fastapi,flask]==0.245.0 - strawberry-v0.264.0: strawberry-graphql[fastapi,flask]==0.264.0 + strawberry-v0.228.0: strawberry-graphql[fastapi,flask]==0.228.0 + strawberry-v0.247.2: strawberry-graphql[fastapi,flask]==0.247.2 + strawberry-v0.265.1: strawberry-graphql[fastapi,flask]==0.265.1 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.227.7: pydantic<2.11 - strawberry-v0.245.0: pydantic<2.11 + strawberry-v0.228.0: pydantic<2.11 + strawberry-v0.247.2: pydantic<2.11 # ~~~ Network ~~~ From f3687fcbd367187c395a802a98ce7eb275239ca1 Mon Sep 17 00:00:00 2001 From: Colton Allen Date: Thu, 17 Apr 2025 08:24:49 -0500 Subject: [PATCH 263/591] feat(spans): Record flag evaluations as span attributes (#4280) Flags evaluated within a span are appended to the span as attributes. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/feature_flags.py | 4 ++ sentry_sdk/integrations/launchdarkly.py | 6 +-- sentry_sdk/integrations/openfeature.py | 8 ++-- sentry_sdk/integrations/unleash.py | 5 +-- sentry_sdk/tracing.py | 13 +++++- .../launchdarkly/test_launchdarkly.py | 41 +++++++++++++++++++ .../openfeature/test_openfeature.py | 26 ++++++++++++ tests/integrations/statsig/test_statsig.py | 20 +++++++++ tests/integrations/unleash/test_unleash.py | 20 +++++++++ tests/test_feature_flags.py | 39 ++++++++++++++++++ 10 files changed, 170 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/feature_flags.py b/sentry_sdk/feature_flags.py index a0b1338356..dd8d41c32e 100644 --- a/sentry_sdk/feature_flags.py +++ b/sentry_sdk/feature_flags.py @@ -66,3 +66,7 @@ def add_feature_flag(flag, result): """ flags = sentry_sdk.get_current_scope().flags flags.set(flag, result) + + span = sentry_sdk.get_current_span() + if span: + span.set_flag(f"flag.evaluation.{flag}", result) diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index cb9e911463..d3c423e7be 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -1,6 +1,6 @@ from typing import TYPE_CHECKING -import sentry_sdk +from sentry_sdk.feature_flags import add_feature_flag from sentry_sdk.integrations import DidNotEnable, Integration try: @@ -53,8 +53,8 @@ def metadata(self): def after_evaluation(self, series_context, data, detail): # type: (EvaluationSeriesContext, dict[Any, Any], EvaluationDetail) -> dict[Any, Any] if isinstance(detail.value, bool): - flags = sentry_sdk.get_current_scope().flags - flags.set(series_context.key, detail.value) + add_feature_flag(series_context.key, detail.value) + return data def before_evaluation(self, series_context, data): diff --git a/sentry_sdk/integrations/openfeature.py b/sentry_sdk/integrations/openfeature.py index bf66b94e8b..e2b33d83f2 100644 --- a/sentry_sdk/integrations/openfeature.py +++ b/sentry_sdk/integrations/openfeature.py @@ -1,6 +1,6 @@ from typing import TYPE_CHECKING -import sentry_sdk +from sentry_sdk.feature_flags import add_feature_flag from sentry_sdk.integrations import DidNotEnable, Integration try: @@ -29,11 +29,9 @@ class OpenFeatureHook(Hook): def after(self, hook_context, details, hints): # type: (HookContext, FlagEvaluationDetails[bool], HookHints) -> None if isinstance(details.value, bool): - flags = sentry_sdk.get_current_scope().flags - flags.set(details.flag_key, details.value) + add_feature_flag(details.flag_key, details.value) def error(self, hook_context, exception, hints): # type: (HookContext, Exception, HookHints) -> None if isinstance(hook_context.default_value, bool): - flags = sentry_sdk.get_current_scope().flags - flags.set(hook_context.flag_key, hook_context.default_value) + add_feature_flag(hook_context.flag_key, hook_context.default_value) diff --git a/sentry_sdk/integrations/unleash.py b/sentry_sdk/integrations/unleash.py index 873f36c68b..6daa0a411f 100644 --- a/sentry_sdk/integrations/unleash.py +++ b/sentry_sdk/integrations/unleash.py @@ -1,7 +1,7 @@ from functools import wraps from typing import Any -import sentry_sdk +from sentry_sdk.feature_flags import add_feature_flag from sentry_sdk.integrations import Integration, DidNotEnable try: @@ -26,8 +26,7 @@ def sentry_is_enabled(self, feature, *args, **kwargs): # We have no way of knowing what type of unleash feature this is, so we have to treat # it as a boolean / toggle feature. - flags = sentry_sdk.get_current_scope().flags - flags.set(feature, enabled) + add_feature_flag(feature, enabled) return enabled diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 13d9f63d5e..ae0b90253e 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -278,6 +278,8 @@ class Span: "scope", "origin", "name", + "_flags", + "_flags_capacity", ) def __init__( @@ -313,6 +315,8 @@ def __init__( self._tags = {} # type: MutableMapping[str, str] self._data = {} # type: Dict[str, Any] self._containing_transaction = containing_transaction + self._flags = {} # type: Dict[str, bool] + self._flags_capacity = 10 if hub is not None: warnings.warn( @@ -597,6 +601,11 @@ def set_data(self, key, value): # type: (str, Any) -> None self._data[key] = value + def set_flag(self, flag, result): + # type: (str, bool) -> None + if len(self._flags) < self._flags_capacity: + self._flags[flag] = result + def set_status(self, value): # type: (str) -> None self.status = value @@ -700,7 +709,9 @@ def to_json(self): if tags: rv["tags"] = tags - data = self._data + data = {} + data.update(self._flags) + data.update(self._data) if data: rv["data"] = data diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index 20566ce09a..20bb4d031f 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -12,6 +12,8 @@ import sentry_sdk from sentry_sdk.integrations import DidNotEnable from sentry_sdk.integrations.launchdarkly import LaunchDarklyIntegration +from sentry_sdk import start_span, start_transaction +from tests.conftest import ApproxDict @pytest.mark.parametrize( @@ -202,3 +204,42 @@ def test_launchdarkly_integration_did_not_enable(monkeypatch): monkeypatch.setattr(client, "is_initialized", lambda: False) with pytest.raises(DidNotEnable): LaunchDarklyIntegration(ld_client=client) + + +@pytest.mark.parametrize( + "use_global_client", + (False, True), +) +def test_launchdarkly_span_integration( + sentry_init, use_global_client, capture_events, uninstall_integration +): + td = TestData.data_source() + td.update(td.flag("hello").variation_for_all(True)) + # Disable background requests as we aren't using a server. + config = Config( + "sdk-key", update_processor_class=td, diagnostic_opt_out=True, send_events=False + ) + + uninstall_integration(LaunchDarklyIntegration.identifier) + if use_global_client: + ldclient.set_config(config) + sentry_init(traces_sample_rate=1.0, integrations=[LaunchDarklyIntegration()]) + client = ldclient.get() + else: + client = LDClient(config=config) + sentry_init( + traces_sample_rate=1.0, + integrations=[LaunchDarklyIntegration(ld_client=client)], + ) + + events = capture_events() + + with start_transaction(name="hi"): + with start_span(op="foo", name="bar"): + client.variation("hello", Context.create("my-org", "organization"), False) + client.variation("other", Context.create("my-org", "organization"), False) + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + {"flag.evaluation.hello": True, "flag.evaluation.other": False} + ) diff --git a/tests/integrations/openfeature/test_openfeature.py b/tests/integrations/openfeature/test_openfeature.py index c180211c3f..46acc61ae7 100644 --- a/tests/integrations/openfeature/test_openfeature.py +++ b/tests/integrations/openfeature/test_openfeature.py @@ -7,7 +7,9 @@ from openfeature.provider.in_memory_provider import InMemoryFlag, InMemoryProvider import sentry_sdk +from sentry_sdk import start_span, start_transaction from sentry_sdk.integrations.openfeature import OpenFeatureIntegration +from tests.conftest import ApproxDict def test_openfeature_integration(sentry_init, capture_events, uninstall_integration): @@ -151,3 +153,27 @@ async def runner(): {"flag": "world", "result": False}, ] } + + +def test_openfeature_span_integration( + sentry_init, capture_events, uninstall_integration +): + uninstall_integration(OpenFeatureIntegration.identifier) + sentry_init(traces_sample_rate=1.0, integrations=[OpenFeatureIntegration()]) + + api.set_provider( + InMemoryProvider({"hello": InMemoryFlag("on", {"on": True, "off": False})}) + ) + client = api.get_client() + + events = capture_events() + + with start_transaction(name="hi"): + with start_span(op="foo", name="bar"): + client.get_boolean_value("hello", default_value=False) + client.get_boolean_value("world", default_value=False) + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + {"flag.evaluation.hello": True, "flag.evaluation.world": False} + ) diff --git a/tests/integrations/statsig/test_statsig.py b/tests/integrations/statsig/test_statsig.py index c1666bde4d..5eb2cf39f3 100644 --- a/tests/integrations/statsig/test_statsig.py +++ b/tests/integrations/statsig/test_statsig.py @@ -5,6 +5,8 @@ from statsig.statsig_user import StatsigUser from random import random from unittest.mock import Mock +from sentry_sdk import start_span, start_transaction +from tests.conftest import ApproxDict import pytest @@ -181,3 +183,21 @@ def test_wrapper_attributes(sentry_init, uninstall_integration): # Clean up statsig.check_gate = original_check_gate + + +def test_statsig_span_integration(sentry_init, capture_events, uninstall_integration): + uninstall_integration(StatsigIntegration.identifier) + + with mock_statsig({"hello": True}): + sentry_init(traces_sample_rate=1.0, integrations=[StatsigIntegration()]) + events = capture_events() + user = StatsigUser(user_id="user-id") + with start_transaction(name="hi"): + with start_span(op="foo", name="bar"): + statsig.check_gate(user, "hello") + statsig.check_gate(user, "world") + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + {"flag.evaluation.hello": True, "flag.evaluation.world": False} + ) diff --git a/tests/integrations/unleash/test_unleash.py b/tests/integrations/unleash/test_unleash.py index 379abba8f6..98a6188181 100644 --- a/tests/integrations/unleash/test_unleash.py +++ b/tests/integrations/unleash/test_unleash.py @@ -8,7 +8,9 @@ import sentry_sdk from sentry_sdk.integrations.unleash import UnleashIntegration +from sentry_sdk import start_span, start_transaction from tests.integrations.unleash.testutils import mock_unleash_client +from tests.conftest import ApproxDict def test_is_enabled(sentry_init, capture_events, uninstall_integration): @@ -164,3 +166,21 @@ def test_wrapper_attributes(sentry_init, uninstall_integration): # Mock clients methods have not lost their qualified names after decoration. assert client.is_enabled.__name__ == "is_enabled" assert client.is_enabled.__qualname__ == original_is_enabled.__qualname__ + + +def test_unleash_span_integration(sentry_init, capture_events, uninstall_integration): + uninstall_integration(UnleashIntegration.identifier) + + with mock_unleash_client(): + sentry_init(traces_sample_rate=1.0, integrations=[UnleashIntegration()]) + events = capture_events() + client = UnleashClient() # type: ignore[arg-type] + with start_transaction(name="hi"): + with start_span(op="foo", name="bar"): + client.is_enabled("hello") + client.is_enabled("other") + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + {"flag.evaluation.hello": True, "flag.evaluation.other": False} + ) diff --git a/tests/test_feature_flags.py b/tests/test_feature_flags.py index 0df30bd0ea..1b0ed13d49 100644 --- a/tests/test_feature_flags.py +++ b/tests/test_feature_flags.py @@ -7,6 +7,8 @@ import sentry_sdk from sentry_sdk.feature_flags import add_feature_flag, FlagBuffer +from sentry_sdk import start_span, start_transaction +from tests.conftest import ApproxDict def test_featureflags_integration(sentry_init, capture_events, uninstall_integration): @@ -220,3 +222,40 @@ def reader(): # shared resource. When deepcopying we should have exclusive access to the underlying # memory. assert error_occurred is False + + +def test_flag_limit(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + + events = capture_events() + + with start_transaction(name="hi"): + with start_span(op="foo", name="bar"): + add_feature_flag("0", True) + add_feature_flag("1", True) + add_feature_flag("2", True) + add_feature_flag("3", True) + add_feature_flag("4", True) + add_feature_flag("5", True) + add_feature_flag("6", True) + add_feature_flag("7", True) + add_feature_flag("8", True) + add_feature_flag("9", True) + add_feature_flag("10", True) + + (event,) = events + assert event["spans"][0]["data"] == ApproxDict( + { + "flag.evaluation.0": True, + "flag.evaluation.1": True, + "flag.evaluation.2": True, + "flag.evaluation.3": True, + "flag.evaluation.4": True, + "flag.evaluation.5": True, + "flag.evaluation.6": True, + "flag.evaluation.7": True, + "flag.evaluation.8": True, + "flag.evaluation.9": True, + } + ) + assert "flag.evaluation.10" not in event["spans"][0]["data"] From c3613370f638086bbd4ff235e500e508b1ca877d Mon Sep 17 00:00:00 2001 From: Roman Inflianskas Date: Tue, 22 Apr 2025 12:09:32 +0300 Subject: [PATCH 264/591] test(logs): Avoid failure when running with integrations enabled (#4316) When (at least) one of integrations is enabled (because some dependencies are installed in the environment), `sentry.sdk.name` is changed from `sentry.python` to `sentry.python.[FIRST_ENABLED_INTEGRATION]` which makes `test_logs_attributes` fail. Prevent failure by relaxing the check. This change is beneficial not only for packaging (this patch was required for packaging for Fedora), but also for running tests with `pytest` directly. --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- tests/test_logs.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/test_logs.py b/tests/test_logs.py index 1c34d52b20..5ede277e3b 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -186,7 +186,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): assert "sentry.release" in logs[0]["attributes"] assert logs[0]["attributes"]["sentry.message.parameters.my_var"] == "some value" assert logs[0]["attributes"][SPANDATA.SERVER_ADDRESS] == "test-server" - assert logs[0]["attributes"]["sentry.sdk.name"] == "sentry.python" + assert logs[0]["attributes"]["sentry.sdk.name"].startswith("sentry.python") assert logs[0]["attributes"]["sentry.sdk.version"] == VERSION From 11e26483d5eeb3f9b35f51e49c69622cd85c88bd Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 22 Apr 2025 09:14:37 +0000 Subject: [PATCH 265/591] build(deps): bump codecov/codecov-action from 5.4.0 to 5.4.2 (#4318) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.4.0 to 5.4.2.
Release notes

Sourced from codecov/codecov-action's releases.

v5.4.2

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.1...v5.4.2

v5.4.1

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.0...v5.4.1

v5.4.1-beta

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.0...v5.4.1-beta

Changelog

Sourced from codecov/codecov-action's changelog.

v5.4.2

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.1..v5.4.2

v5.4.1

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.0..v5.4.1

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=codecov/codecov-action&package-manager=github_actions&previous-version=5.4.0&new-version=5.4.2)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 13 files changed, 19 insertions(+), 19 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index e497ba4280..f392f57f46 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -83,7 +83,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -158,7 +158,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 1d728f3486..7763aa509d 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 4fa12607eb..864583532d 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 435ec9d7bb..815b550027 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -107,7 +107,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -206,7 +206,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index f2fdfd5473..e28067841b 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index eb6aa1297f..41a77ffe34 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 9713f80c25..b741302de6 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 607835ee94..7da9929435 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index b51c7bfb07..43b5e4a6a5 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -75,7 +75,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index a27c13278f..a6850256b2 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -186,7 +186,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index ac364ccfc1..b40027ddc7 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 3d3d6e7c84..1fbff47b65 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 91849beff4..901e4808e4 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -91,7 +91,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.4.0 + uses: codecov/codecov-action@v5.4.2 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From d1819c7786de40bfc322aeab1681715c9dbf05bc Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 22 Apr 2025 11:17:55 +0200 Subject: [PATCH 266/591] Make all relevant types public (#4315) Make types that users can use when configuring the SDK public. Accompaniyng docs update: https://github.com/getsentry/sentry-docs/pull/13437 Fixes #4127 --- sentry_sdk/_types.py | 6 ++++++ sentry_sdk/types.py | 28 ++++++++++++++++++++++++++-- 2 files changed, 32 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 9bcb5a61f9..7da76e63dc 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -220,7 +220,9 @@ class SDKInfo(TypedDict): tuple[None, None, None], ] + # TODO: Make a proper type definition for this (PRs welcome!) Hint = Dict[str, Any] + Log = TypedDict( "Log", { @@ -233,9 +235,13 @@ class SDKInfo(TypedDict): }, ) + # TODO: Make a proper type definition for this (PRs welcome!) Breadcrumb = Dict[str, Any] + + # TODO: Make a proper type definition for this (PRs welcome!) BreadcrumbHint = Dict[str, Any] + # TODO: Make a proper type definition for this (PRs welcome!) SamplingContext = Dict[str, Any] EventProcessor = Callable[[Event, Hint], Optional[Event]] diff --git a/sentry_sdk/types.py b/sentry_sdk/types.py index 2b9f04c097..1a65247584 100644 --- a/sentry_sdk/types.py +++ b/sentry_sdk/types.py @@ -11,15 +11,39 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from sentry_sdk._types import Event, EventDataCategory, Hint, Log + # Re-export types to make them available in the public API + from sentry_sdk._types import ( + Breadcrumb, + BreadcrumbHint, + Event, + EventDataCategory, + Hint, + Log, + MonitorConfig, + SamplingContext, + ) else: from typing import Any # The lines below allow the types to be imported from outside `if TYPE_CHECKING` # guards. The types in this module are only intended to be used for type hints. + Breadcrumb = Any + BreadcrumbHint = Any Event = Any EventDataCategory = Any Hint = Any Log = Any + MonitorConfig = Any + SamplingContext = Any -__all__ = ("Event", "EventDataCategory", "Hint", "Log") + +__all__ = ( + "Breadcrumb", + "BreadcrumbHint", + "Event", + "EventDataCategory", + "Hint", + "Log", + "MonitorConfig", + "SamplingContext", +) From b96e2b64a8fd29d5b55bf419be5c299fc28956e4 Mon Sep 17 00:00:00 2001 From: Dong Guo Date: Tue, 22 Apr 2025 17:27:09 +0800 Subject: [PATCH 267/591] fix(integrations): ASGI integration not capture transactions in Websocket (#4293) In [ASGI Specs](https://github.com/django/asgiref/blob/main/specs/www.rst#websocket-connection-scope), `method` is not in Websocket Connection Scope. --- sentry_sdk/integrations/asgi.py | 25 +++++++++++++------------ tests/integrations/asgi/test_asgi.py | 25 +++++++++++-------------- 2 files changed, 24 insertions(+), 26 deletions(-) diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index 3569336aae..fc8ee29b1a 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -192,8 +192,8 @@ async def _run_app(self, scope, receive, send, asgi_version): method = scope.get("method", "").upper() transaction = None - if method in self.http_methods_to_capture: - if ty in ("http", "websocket"): + if ty in ("http", "websocket"): + if ty == "websocket" or method in self.http_methods_to_capture: transaction = continue_trace( _get_headers(scope), op="{}.server".format(ty), @@ -205,17 +205,18 @@ async def _run_app(self, scope, receive, send, asgi_version): "[ASGI] Created transaction (continuing trace): %s", transaction, ) - else: - transaction = Transaction( - op=OP.HTTP_SERVER, - name=transaction_name, - source=transaction_source, - origin=self.span_origin, - ) - logger.debug( - "[ASGI] Created transaction (new): %s", transaction - ) + else: + transaction = Transaction( + op=OP.HTTP_SERVER, + name=transaction_name, + source=transaction_source, + origin=self.span_origin, + ) + logger.debug( + "[ASGI] Created transaction (new): %s", transaction + ) + if transaction: transaction.set_tag("asgi.type", ty) logger.debug( "[ASGI] Set transaction name and source on transaction: '%s' / '%s'", diff --git a/tests/integrations/asgi/test_asgi.py b/tests/integrations/asgi/test_asgi.py index f95ea14d01..ec2796c140 100644 --- a/tests/integrations/asgi/test_asgi.py +++ b/tests/integrations/asgi/test_asgi.py @@ -349,35 +349,32 @@ async def test_trace_from_headers_if_performance_disabled( @pytest.mark.asyncio async def test_websocket(sentry_init, asgi3_ws_app, capture_events, request): - sentry_init(send_default_pii=True) + sentry_init(send_default_pii=True, traces_sample_rate=1.0) events = capture_events() asgi3_ws_app = SentryAsgiMiddleware(asgi3_ws_app) - scope = { - "type": "websocket", - "endpoint": asgi3_app, - "client": ("127.0.0.1", 60457), - "route": "some_url", - "headers": [ - ("accept", "*/*"), - ], - } + request_url = "/ws" with pytest.raises(ValueError): - async with TestClient(asgi3_ws_app, scope=scope) as client: - async with client.websocket_connect("/ws") as ws: - await ws.receive_text() + client = TestClient(asgi3_ws_app) + async with client.websocket_connect(request_url) as ws: + await ws.receive_text() - msg_event, error_event = events + msg_event, error_event, transaction_event = events + assert msg_event["transaction"] == request_url + assert msg_event["transaction_info"] == {"source": "url"} assert msg_event["message"] == "Some message to the world!" (exc,) = error_event["exception"]["values"] assert exc["type"] == "ValueError" assert exc["value"] == "Oh no" + assert transaction_event["transaction"] == request_url + assert transaction_event["transaction_info"] == {"source": "url"} + @pytest.mark.asyncio async def test_auto_session_tracking_with_aggregates( From 434e8afb9762e6eab22165937069271729958d3d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 23 Apr 2025 10:54:54 +0200 Subject: [PATCH 268/591] tests: Fix version picking in toxgen (#4323) Toxgen should only consider the highest patch release of each `major.minor` version. For the most part this was working fine as long as the releases were ordered as expected in PyPI, but in cases where a lower patch version succeeded a higher patch version in the release list from PyPI, we would incorrectly consider the lower patch version as well, instead of ignoring it in favor of the higher patch. Example: - we pull releases `[1.2.3, 1.2.4, 1.2.5, 1.2.2]` from PyPI (in that order) - we consolidate `1.2.3, 1.2.4, 1.2.5` into one version, `1.2.5`, as expected - `1.2.2` will not disappear into `1.2.5` because of a faulty check in toxgen and will instead be considered as a new version - our resulting list of releases eligible for testing will be `[1.2.5, 1.2.2]` instead of just `[1.2.5]`, which then results in picking versions that are not nicely spaced apart --- scripts/populate_tox/populate_tox.py | 4 +-- tox.ini | 51 ++++++++++++---------------- 2 files changed, 24 insertions(+), 31 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 11ea94c0f4..f741496f93 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -190,10 +190,10 @@ def _prefilter_releases( if ( version.major == saved_version.major and version.minor == saved_version.minor - and version.micro > saved_version.micro ): # Don't save all patch versions of a release, just the newest one - filtered_releases[i] = version + if version.micro > saved_version.micro: + filtered_releases[i] = version break else: filtered_releases.append(version) diff --git a/tox.ini b/tox.ini index 9497708ff8..49411b3189 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-17T11:01:25.976599+00:00 +# Last generated: 2025-04-23T07:46:44.042662+00:00 [tox] requires = @@ -145,8 +145,8 @@ envlist = # ~~~ AI ~~~ {py3.9,py3.10,py3.11}-cohere-v5.4.0 - {py3.9,py3.11,py3.12}-cohere-v5.9.4 - {py3.9,py3.11,py3.12}-cohere-v5.13.9 + {py3.9,py3.11,py3.12}-cohere-v5.8.1 + {py3.9,py3.11,py3.12}-cohere-v5.11.4 {py3.9,py3.11,py3.12}-cohere-v5.15.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 @@ -167,9 +167,8 @@ envlist = {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 - {py3.6,py3.7}-sqlalchemy-v1.3.9 + {py3.6,py3.8,py3.9}-sqlalchemy-v1.3.24 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 - {py3.7,py3.10,py3.11}-sqlalchemy-v2.0.9 {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.40 @@ -195,7 +194,7 @@ envlist = {py3.8,py3.10,py3.11}-ariadne-v0.20.1 {py3.8,py3.11,py3.12}-ariadne-v0.22 {py3.8,py3.11,py3.12}-ariadne-v0.24.0 - {py3.9,py3.12,py3.13}-ariadne-v0.26.1 + {py3.9,py3.12,py3.13}-ariadne-v0.26.2 {py3.6,py3.9,py3.10}-gql-v3.4.1 {py3.7,py3.11,py3.12}-gql-v3.5.2 @@ -207,7 +206,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.228.0 {py3.8,py3.12,py3.13}-strawberry-v0.247.2 - {py3.9,py3.12,py3.13}-strawberry-v0.265.1 + {py3.9,py3.12,py3.13}-strawberry-v0.266.0 # ~~~ Network ~~~ @@ -240,12 +239,11 @@ envlist = # ~~~ Web 1 ~~~ - {py3.6}-django-v1.11.9 {py3.6,py3.7}-django-v1.11.29 {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 {py3.8,py3.11,py3.12}-django-v4.2.20 - {py3.10,py3.11,py3.12}-django-v5.0.9 + {py3.10,py3.11,py3.12}-django-v5.0.14 {py3.10,py3.12,py3.13}-django-v5.2 {py3.6,py3.7,py3.8}-flask-v1.1.4 @@ -266,7 +264,7 @@ envlist = # ~~~ Web 2 ~~~ {py3.6,py3.7}-bottle-v0.12.25 - {py3.6,py3.8,py3.9}-bottle-v0.13.2 + {py3.8,py3.12,py3.13}-bottle-v0.13.3 {py3.6}-falcon-v1.4.1 {py3.6,py3.7}-falcon-v2.0.0 @@ -296,11 +294,11 @@ envlist = # ~~~ Misc ~~~ {py3.6,py3.12,py3.13}-loguru-v0.7.3 - {py3.6}-trytond-v4.6.9 + {py3.6}-trytond-v4.6.22 {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.9 + {py3.8,py3.11,py3.12}-trytond-v7.0.29 {py3.8,py3.11,py3.12}-trytond-v7.4.9 {py3.7,py3.12,py3.13}-typer-v0.15.2 @@ -517,8 +515,8 @@ deps = # ~~~ AI ~~~ cohere-v5.4.0: cohere==5.4.0 - cohere-v5.9.4: cohere==5.9.4 - cohere-v5.13.9: cohere==5.13.9 + cohere-v5.8.1: cohere==5.8.1 + cohere-v5.11.4: cohere==5.11.4 cohere-v5.15.0: cohere==5.15.0 huggingface_hub-v0.22.2: huggingface_hub==0.22.2 @@ -540,9 +538,8 @@ deps = redis_py_cluster_legacy-v2.0.0: redis-py-cluster==2.0.0 redis_py_cluster_legacy-v2.1.3: redis-py-cluster==2.1.3 - sqlalchemy-v1.3.9: sqlalchemy==1.3.9 + sqlalchemy-v1.3.24: sqlalchemy==1.3.24 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 - sqlalchemy-v2.0.9: sqlalchemy==2.0.9 sqlalchemy-v2.0.40: sqlalchemy==2.0.40 @@ -569,7 +566,7 @@ deps = ariadne-v0.20.1: ariadne==0.20.1 ariadne-v0.22: ariadne==0.22 ariadne-v0.24.0: ariadne==0.24.0 - ariadne-v0.26.1: ariadne==0.26.1 + ariadne-v0.26.2: ariadne==0.26.2 ariadne: fastapi ariadne: flask ariadne: httpx @@ -589,7 +586,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.228.0: strawberry-graphql[fastapi,flask]==0.228.0 strawberry-v0.247.2: strawberry-graphql[fastapi,flask]==0.247.2 - strawberry-v0.265.1: strawberry-graphql[fastapi,flask]==0.265.1 + strawberry-v0.266.0: strawberry-graphql[fastapi,flask]==0.266.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.228.0: pydantic<2.11 @@ -633,12 +630,11 @@ deps = # ~~~ Web 1 ~~~ - django-v1.11.9: django==1.11.9 django-v1.11.29: django==1.11.29 django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 django-v4.2.20: django==4.2.20 - django-v5.0.9: django==5.0.9 + django-v5.0.14: django==5.0.14 django-v5.2: django==5.2 django: psycopg2-binary django: djangorestframework @@ -646,24 +642,21 @@ deps = django: Werkzeug django-v3.2.25: pytest-asyncio django-v4.2.20: pytest-asyncio - django-v5.0.9: pytest-asyncio + django-v5.0.14: pytest-asyncio django-v5.2: pytest-asyncio django-v2.2.28: six - django-v1.11.9: djangorestframework>=3.0,<4.0 - django-v1.11.9: Werkzeug<2.1.0 django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 django-v2.2.28: Werkzeug<2.1.0 django-v3.2.25: djangorestframework>=3.0,<4.0 django-v3.2.25: Werkzeug<2.1.0 - django-v1.11.9: pytest-django<4.0 django-v1.11.29: pytest-django<4.0 django-v2.2.28: pytest-django<4.0 django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] django-v4.2.20: channels[daphne] - django-v5.0.9: channels[daphne] + django-v5.0.14: channels[daphne] django-v5.2: channels[daphne] flask-v1.1.4: flask==1.1.4 @@ -707,7 +700,7 @@ deps = # ~~~ Web 2 ~~~ bottle-v0.12.25: bottle==0.12.25 - bottle-v0.13.2: bottle==0.13.2 + bottle-v0.13.3: bottle==0.13.3 bottle: werkzeug<2.1.0 falcon-v1.4.1: falcon==1.4.1 @@ -756,14 +749,14 @@ deps = # ~~~ Misc ~~~ loguru-v0.7.3: loguru==0.7.3 - trytond-v4.6.9: trytond==4.6.9 + trytond-v4.6.22: trytond==4.6.22 trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.9: trytond==7.0.9 + trytond-v7.0.29: trytond==7.0.29 trytond-v7.4.9: trytond==7.4.9 trytond: werkzeug - trytond-v4.6.9: werkzeug<1.0 + trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 typer-v0.15.2: typer==0.15.2 From 2c3776c582a23b6936c76ef53008bf63f861b6fd Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 23 Apr 2025 11:03:10 +0200 Subject: [PATCH 269/591] tests: Move aiohttp under toxgen (#4319) Depends on https://github.com/getsentry/sentry-python/pull/4323 --- scripts/populate_tox/config.py | 8 +++++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 13 ---------- tests/integrations/aiohttp/test_aiohttp.py | 24 ++++++++++++------- tox.ini | 28 +++++++++++----------- 5 files changed, 37 insertions(+), 37 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f3f1ba0092..f874ff8a9c 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -6,6 +6,14 @@ # See scripts/populate_tox/README.md for more info on the format and examples. TEST_SUITE_CONFIG = { + "aiohttp": { + "package": "aiohttp", + "deps": { + "*": ["pytest-aiohttp"], + ">=3.8": ["pytest-asyncio"], + }, + "python": ">=3.7", + }, "ariadne": { "package": "ariadne", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index f741496f93..c04ab1b209 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "aiohttp", "anthropic", "arq", "asyncpg", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 380a80f690..3cfb5e1252 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -36,11 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # AIOHTTP - {py3.7}-aiohttp-v{3.4} - {py3.7,py3.9,py3.11}-aiohttp-v{3.8} - {py3.8,py3.12,py3.13}-aiohttp-latest - # Anthropic {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} {py3.7,py3.11,py3.12}-anthropic-latest @@ -184,14 +179,6 @@ deps = # === Integrations === - # AIOHTTP - aiohttp-v3.4: aiohttp~=3.4.0 - aiohttp-v3.8: aiohttp~=3.8.0 - aiohttp-latest: aiohttp - aiohttp: pytest-aiohttp - aiohttp-v3.8: pytest-asyncio - aiohttp-latest: pytest-asyncio - # Anthropic anthropic: pytest-asyncio anthropic-v{0.16,0.28}: httpx<0.28.0 diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index ef7c04e90a..06859b127f 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -1,10 +1,16 @@ import asyncio import json -import sys + from contextlib import suppress from unittest import mock import pytest + +try: + import pytest_asyncio +except ImportError: + pytest_asyncio = None + from aiohttp import web, ClientSession from aiohttp.client import ServerDisconnectedError from aiohttp.web_request import Request @@ -21,6 +27,14 @@ from tests.conftest import ApproxDict +if pytest_asyncio is None: + # `loop` was deprecated in `pytest-aiohttp` + # in favor of `event_loop` from `pytest-asyncio` + @pytest.fixture + def event_loop(loop): + yield loop + + @pytest.mark.asyncio async def test_basic(sentry_init, aiohttp_client, capture_events): sentry_init(integrations=[AioHttpIntegration()]) @@ -474,14 +488,6 @@ async def hello(request): assert error_event["contexts"]["trace"]["trace_id"] == trace_id -if sys.version_info < (3, 12): - # `loop` was deprecated in `pytest-aiohttp` - # in favor of `event_loop` from `pytest-asyncio` - @pytest.fixture - def event_loop(loop): - yield loop - - @pytest.mark.asyncio async def test_crumb_capture( sentry_init, aiohttp_raw_server, aiohttp_client, event_loop, capture_events diff --git a/tox.ini b/tox.ini index 49411b3189..6f3b9863e8 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-23T07:46:44.042662+00:00 +# Last generated: 2025-04-23T08:07:00.653648+00:00 [tox] requires = @@ -36,11 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # AIOHTTP - {py3.7}-aiohttp-v{3.4} - {py3.7,py3.9,py3.11}-aiohttp-v{3.8} - {py3.8,py3.12,py3.13}-aiohttp-latest - # Anthropic {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} {py3.7,py3.11,py3.12}-anthropic-latest @@ -263,6 +258,11 @@ envlist = # ~~~ Web 2 ~~~ + {py3.7}-aiohttp-v3.4.4 + {py3.7}-aiohttp-v3.6.3 + {py3.7,py3.9,py3.10}-aiohttp-v3.8.6 + {py3.9,py3.12,py3.13}-aiohttp-v3.11.18 + {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.3 @@ -335,14 +335,6 @@ deps = # === Integrations === - # AIOHTTP - aiohttp-v3.4: aiohttp~=3.4.0 - aiohttp-v3.8: aiohttp~=3.8.0 - aiohttp-latest: aiohttp - aiohttp: pytest-aiohttp - aiohttp-v3.8: pytest-asyncio - aiohttp-latest: pytest-asyncio - # Anthropic anthropic: pytest-asyncio anthropic-v{0.16,0.28}: httpx<0.28.0 @@ -699,6 +691,14 @@ deps = # ~~~ Web 2 ~~~ + aiohttp-v3.4.4: aiohttp==3.4.4 + aiohttp-v3.6.3: aiohttp==3.6.3 + aiohttp-v3.8.6: aiohttp==3.8.6 + aiohttp-v3.11.18: aiohttp==3.11.18 + aiohttp: pytest-aiohttp + aiohttp-v3.8.6: pytest-asyncio + aiohttp-v3.11.18: pytest-asyncio + bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.3: bottle==0.13.3 bottle: werkzeug<2.1.0 From bbb41a31a71e90b3a72ded603ca0cd9173e23522 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 23 Apr 2025 15:06:32 +0200 Subject: [PATCH 270/591] Make sure to use the default decimal context in our code (#4231) Fixes #4213 --- sentry_sdk/tracing.py | 7 +++---- sentry_sdk/tracing_utils.py | 13 ++++++++----- tests/tracing/test_sample_rand.py | 10 +++++++++- 3 files changed, 20 insertions(+), 10 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index ae0b90253e..ca249fe8fe 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1,3 +1,4 @@ +from decimal import Decimal import uuid import warnings from datetime import datetime, timedelta, timezone @@ -1198,10 +1199,8 @@ def _set_initial_sampling_decision(self, sampling_context): self.sampled = False return - # Now we roll the dice. self._sample_rand is inclusive of 0, but not of 1, - # so strict < is safe here. In case sample_rate is a boolean, cast it - # to a float (True becomes 1.0 and False becomes 0.0) - self.sampled = self._sample_rand < self.sample_rate + # Now we roll the dice. + self.sampled = self._sample_rand < Decimal.from_float(self.sample_rate) if self.sampled: logger.debug( diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index ba56695740..552f4fd59a 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -5,7 +5,7 @@ import sys from collections.abc import Mapping from datetime import timedelta -from decimal import ROUND_DOWN, Context, Decimal +from decimal import ROUND_DOWN, Decimal, DefaultContext, localcontext from functools import wraps from random import Random from urllib.parse import quote, unquote @@ -872,10 +872,13 @@ def _generate_sample_rand( # Round down to exactly six decimal-digit precision. # Setting the context is needed to avoid an InvalidOperation exception - # in case the user has changed the default precision. - return Decimal(sample_rand).quantize( - Decimal("0.000001"), rounding=ROUND_DOWN, context=Context(prec=6) - ) + # in case the user has changed the default precision or set traps. + with localcontext(DefaultContext) as ctx: + ctx.prec = 6 + return Decimal(sample_rand).quantize( + Decimal("0.000001"), + rounding=ROUND_DOWN, + ) def _sample_rand_range(parent_sampled, sample_rate): diff --git a/tests/tracing/test_sample_rand.py b/tests/tracing/test_sample_rand.py index ef277a3dec..f9c10aa04e 100644 --- a/tests/tracing/test_sample_rand.py +++ b/tests/tracing/test_sample_rand.py @@ -1,4 +1,5 @@ import decimal +from decimal import Inexact, FloatOperation from unittest import mock import pytest @@ -58,14 +59,19 @@ def test_transaction_uses_incoming_sample_rand( def test_decimal_context(sentry_init, capture_events): """ - Ensure that having a decimal context with a precision below 6 + Ensure that having a user altered decimal context with a precision below 6 does not cause an InvalidOperation exception. """ sentry_init(traces_sample_rate=1.0) events = capture_events() old_prec = decimal.getcontext().prec + old_inexact = decimal.getcontext().traps[Inexact] + old_float_operation = decimal.getcontext().traps[FloatOperation] + decimal.getcontext().prec = 2 + decimal.getcontext().traps[Inexact] = True + decimal.getcontext().traps[FloatOperation] = True try: with mock.patch( @@ -77,5 +83,7 @@ def test_decimal_context(sentry_init, capture_events): ) finally: decimal.getcontext().prec = old_prec + decimal.getcontext().traps[Inexact] = old_inexact + decimal.getcontext().traps[FloatOperation] = old_float_operation assert len(events) == 1 From 049f2a0b18e22be7b5e77eb31b11122f2a38c92a Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 24 Apr 2025 08:02:13 +0000 Subject: [PATCH 271/591] release: 2.27.0 --- CHANGELOG.md | 19 +++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 22 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index bb49ed54ca..70915e75c5 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,24 @@ # Changelog +## 2.27.0 + +### Various fixes & improvements + +- Make sure to use the default decimal context in our code (#4231) by @antonpirker +- tests: Move aiohttp under toxgen (#4319) by @sentrivana +- tests: Fix version picking in toxgen (#4323) by @sentrivana +- fix(integrations): ASGI integration not capture transactions in Websocket (#4293) by @guodong000 +- Make all relevant types public (#4315) by @antonpirker +- build(deps): bump codecov/codecov-action from 5.4.0 to 5.4.2 (#4318) by @dependabot +- test(logs): Avoid failure when running with integrations enabled (#4316) by @rominf +- feat(spans): Record flag evaluations as span attributes (#4280) by @cmanallen +- toxgen: Remove unused code and rerun (#4313) by @sentrivana +- toxgen: Add cohere (#4304) by @sentrivana +- toxgen: Migrate fastapi (#4302) by @sentrivana +- toxgen: Add huggingface_hub (#4299) by @sentrivana +- toxgen: Add huey (#4298) by @sentrivana +- tests: Update tox.ini (#4297) by @sentrivana + ## 2.26.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 629b5b9eaa..709f557d16 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.26.1" +release = "2.27.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 3802980b82..e1f18fe4ae 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -966,4 +966,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.26.1" +VERSION = "2.27.0" diff --git a/setup.py b/setup.py index 62f4867b35..877585472b 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.26.1", + version="2.27.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 919bdeab17dff035131b0f70848d5675efd96808 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 24 Apr 2025 10:04:12 +0200 Subject: [PATCH 272/591] Update CHANGELOG.md --- CHANGELOG.md | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 70915e75c5..786a9a34e5 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,20 +4,20 @@ ### Various fixes & improvements -- Make sure to use the default decimal context in our code (#4231) by @antonpirker -- tests: Move aiohttp under toxgen (#4319) by @sentrivana -- tests: Fix version picking in toxgen (#4323) by @sentrivana +- fix: Make sure to use the default decimal context in our code (#4231) by @antonpirker - fix(integrations): ASGI integration not capture transactions in Websocket (#4293) by @guodong000 -- Make all relevant types public (#4315) by @antonpirker -- build(deps): bump codecov/codecov-action from 5.4.0 to 5.4.2 (#4318) by @dependabot -- test(logs): Avoid failure when running with integrations enabled (#4316) by @rominf +- feat(typing): Make all relevant types public (#4315) by @antonpirker - feat(spans): Record flag evaluations as span attributes (#4280) by @cmanallen -- toxgen: Remove unused code and rerun (#4313) by @sentrivana -- toxgen: Add cohere (#4304) by @sentrivana -- toxgen: Migrate fastapi (#4302) by @sentrivana -- toxgen: Add huggingface_hub (#4299) by @sentrivana -- toxgen: Add huey (#4298) by @sentrivana +- test(logs): Avoid failure when running with integrations enabled (#4316) by @rominf +- tests: Remove unused code and rerun (#4313) by @sentrivana +- tests: Add cohere to toxgen (#4304) by @sentrivana +- tests: Migrate fastapi to toxgen (#4302) by @sentrivana +- tests: Add huggingface_hub to toxgen (#4299) by @sentrivana +- tests: Add huey to toxgen (#4298) by @sentrivana - tests: Update tox.ini (#4297) by @sentrivana +- tests: Move aiohttp under toxgen (#4319) by @sentrivana +- tests: Fix version picking in toxgen (#4323) by @sentrivana +- build(deps): bump codecov/codecov-action from 5.4.0 to 5.4.2 (#4318) by @dependabot ## 2.26.1 From bbdf789902e3d8ee7940d7b7442934b0d6b8b30d Mon Sep 17 00:00:00 2001 From: Stephanie Anderson Date: Fri, 25 Apr 2025 13:36:32 +0200 Subject: [PATCH 273/591] Update GH issue templates for Linear compatibility (#4328) --- .github/ISSUE_TEMPLATE/bug.yml | 1 + .github/ISSUE_TEMPLATE/feature.yml | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/.github/ISSUE_TEMPLATE/bug.yml b/.github/ISSUE_TEMPLATE/bug.yml index 78f1e03d21..c13d6c4bb0 100644 --- a/.github/ISSUE_TEMPLATE/bug.yml +++ b/.github/ISSUE_TEMPLATE/bug.yml @@ -1,5 +1,6 @@ name: 🐞 Bug Report description: Tell us about something that's not working the way we (probably) intend. +labels: ["Python", "Bug"] body: - type: dropdown id: type diff --git a/.github/ISSUE_TEMPLATE/feature.yml b/.github/ISSUE_TEMPLATE/feature.yml index e462e3bae7..64b31873d8 100644 --- a/.github/ISSUE_TEMPLATE/feature.yml +++ b/.github/ISSUE_TEMPLATE/feature.yml @@ -1,6 +1,6 @@ name: 💡 Feature Request description: Create a feature request for sentry-python SDK. -labels: 'enhancement' +labels: ["Python", "Feature"] body: - type: markdown attributes: From c6db4204c12c677839a5fd7b8536ca57866cb5e1 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Apr 2025 10:40:28 +0200 Subject: [PATCH 274/591] tests: Update tox.ini (#4347) Regular tox.ini update --- tox.ini | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-) diff --git a/tox.ini b/tox.ini index 6f3b9863e8..0632a4e8e3 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-23T08:07:00.653648+00:00 +# Last generated: 2025-04-29T08:15:04.584844+00:00 [tox] requires = @@ -215,7 +215,7 @@ envlist = # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 - {py3.8,py3.12,py3.13}-celery-v5.5.1 + {py3.8,py3.12,py3.13}-celery-v5.5.2 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -298,10 +298,10 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.29 - {py3.8,py3.11,py3.12}-trytond-v7.4.9 + {py3.8,py3.11,py3.12}-trytond-v7.0.30 + {py3.9,py3.12,py3.13}-trytond-v7.6.0 - {py3.7,py3.12,py3.13}-typer-v0.15.2 + {py3.7,py3.12,py3.13}-typer-v0.15.3 @@ -600,7 +600,7 @@ deps = # ~~~ Tasks ~~~ celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 - celery-v5.5.1: celery==5.5.1 + celery-v5.5.2: celery==5.5.2 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 @@ -753,13 +753,13 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.29: trytond==7.0.29 - trytond-v7.4.9: trytond==7.4.9 + trytond-v7.0.30: trytond==7.0.30 + trytond-v7.6.0: trytond==7.6.0 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 - typer-v0.15.2: typer==0.15.2 + typer-v0.15.3: typer==0.15.3 From 28a87dfdca0ae6aeb87a3079d799afe2f89d6de5 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 29 Apr 2025 11:43:37 +0200 Subject: [PATCH 275/591] Deprecate `set_measurement()` API. (#3934) Deprecate `set_measurement()`. This will be replaced by `set_data()` which internally is using the Otel `set_attribute()`. Fixes #3074 --- sentry_sdk/api.py | 4 ++++ sentry_sdk/tracing.py | 20 ++++++++++++++++++ tests/tracing/test_misc.py | 42 ++++++++++++++++++++++++++++++++++++++ 3 files changed, 66 insertions(+) diff --git a/sentry_sdk/api.py b/sentry_sdk/api.py index d60434079c..a6b3c293dc 100644 --- a/sentry_sdk/api.py +++ b/sentry_sdk/api.py @@ -388,6 +388,10 @@ def start_transaction( def set_measurement(name, value, unit=""): # type: (str, float, MeasurementUnit) -> None + """ + .. deprecated:: 2.28.0 + This function is deprecated and will be removed in the next major release. + """ transaction = get_current_scope().transaction if transaction is not None: transaction.set_measurement(name, value, unit) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index ca249fe8fe..fc40221b9f 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -613,6 +613,16 @@ def set_status(self, value): def set_measurement(self, name, value, unit=""): # type: (str, float, MeasurementUnit) -> None + """ + .. deprecated:: 2.28.0 + This function is deprecated and will be removed in the next major release. + """ + + warnings.warn( + "`set_measurement()` is deprecated and will be removed in the next major version. Please use `set_data()` instead.", + DeprecationWarning, + stacklevel=2, + ) self._measurements[name] = {"value": value, "unit": unit} def set_thread(self, thread_id, thread_name): @@ -1061,6 +1071,16 @@ def finish( def set_measurement(self, name, value, unit=""): # type: (str, float, MeasurementUnit) -> None + """ + .. deprecated:: 2.28.0 + This function is deprecated and will be removed in the next major release. + """ + + warnings.warn( + "`set_measurement()` is deprecated and will be removed in the next major version. Please use `set_data()` instead.", + DeprecationWarning, + stacklevel=2, + ) self._measurements[name] = {"value": value, "unit": unit} def set_context(self, key, value): diff --git a/tests/tracing/test_misc.py b/tests/tracing/test_misc.py index 040fb24213..b954d36e1a 100644 --- a/tests/tracing/test_misc.py +++ b/tests/tracing/test_misc.py @@ -323,6 +323,48 @@ def test_set_meaurement_public_api(sentry_init, capture_events): assert event["measurements"]["metric.bar"] == {"value": 456, "unit": "second"} +def test_set_measurement_deprecated(sentry_init): + sentry_init(traces_sample_rate=1.0) + + with start_transaction(name="measuring stuff") as trx: + with pytest.warns(DeprecationWarning): + set_measurement("metric.foo", 123) + + with pytest.warns(DeprecationWarning): + trx.set_measurement("metric.bar", 456) + + with start_span(op="measuring span") as span: + with pytest.warns(DeprecationWarning): + span.set_measurement("metric.baz", 420.69, unit="custom") + + +def test_set_meaurement_compared_to_set_data(sentry_init, capture_events): + """ + This is just a test to see the difference + between measurements and data in the resulting event payload. + """ + sentry_init(traces_sample_rate=1.0) + + events = capture_events() + + with start_transaction(name="measuring stuff") as transaction: + transaction.set_measurement("metric.foo", 123) + transaction.set_data("metric.bar", 456) + + with start_span(op="measuring span") as span: + span.set_measurement("metric.baz", 420.69, unit="custom") + span.set_data("metric.qux", 789) + + (event,) = events + assert event["measurements"]["metric.foo"] == {"value": 123, "unit": ""} + assert event["contexts"]["trace"]["data"]["metric.bar"] == 456 + assert event["spans"][0]["measurements"]["metric.baz"] == { + "value": 420.69, + "unit": "custom", + } + assert event["spans"][0]["data"]["metric.qux"] == 789 + + @pytest.mark.parametrize( "trace_propagation_targets,url,expected_propagation_decision", [ From 1041dbb6b2aec9d75b323e57a65ef2c02bed750e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Apr 2025 11:58:28 +0200 Subject: [PATCH 276/591] tests: Move anthropic under toxgen (#4348) --- .github/workflows/test-integrations-ai.yml | 2 +- scripts/populate_tox/config.py | 8 +++++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 12 ---------- tox.ini | 28 ++++++++++++---------- 5 files changed, 24 insertions(+), 27 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index f392f57f46..bc89cb9afe 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.9","3.11","3.12"] + python-version: ["3.9","3.11","3.12"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f874ff8a9c..4d5d5b14ce 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -14,6 +14,14 @@ }, "python": ">=3.7", }, + "anthropic": { + "package": "anthropic", + "deps": { + "*": ["pytest-asyncio"], + "<0.50": ["httpx<0.28.0"], + }, + "python": ">=3.8", + }, "ariadne": { "package": "ariadne", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index c04ab1b209..0aeb0f02ef 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "anthropic", "arq", "asyncpg", "beam", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 3cfb5e1252..2869da275b 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -36,10 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # Anthropic - {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} - {py3.7,py3.11,py3.12}-anthropic-latest - # Arq {py3.7,py3.11}-arq-v{0.23} {py3.7,py3.12,py3.13}-arq-latest @@ -179,14 +175,6 @@ deps = # === Integrations === - # Anthropic - anthropic: pytest-asyncio - anthropic-v{0.16,0.28}: httpx<0.28.0 - anthropic-v0.16: anthropic~=0.16.0 - anthropic-v0.28: anthropic~=0.28.0 - anthropic-v0.40: anthropic~=0.40.0 - anthropic-latest: anthropic - # Arq arq-v0.23: arq~=0.23.0 arq-v0.23: pydantic<2 diff --git a/tox.ini b/tox.ini index 0632a4e8e3..4c05bcaa75 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-29T08:15:04.584844+00:00 +# Last generated: 2025-04-29T08:35:44.624881+00:00 [tox] requires = @@ -36,10 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # Anthropic - {py3.8,py3.11,py3.12}-anthropic-v{0.16,0.28,0.40} - {py3.7,py3.11,py3.12}-anthropic-latest - # Arq {py3.7,py3.11}-arq-v{0.23} {py3.7,py3.12,py3.13}-arq-latest @@ -139,6 +135,11 @@ envlist = # integration tests there. # ~~~ AI ~~~ + {py3.8,py3.11,py3.12}-anthropic-v0.16.0 + {py3.8,py3.11,py3.12}-anthropic-v0.27.0 + {py3.8,py3.11,py3.12}-anthropic-v0.38.0 + {py3.8,py3.11,py3.12}-anthropic-v0.50.0 + {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 {py3.9,py3.11,py3.12}-cohere-v5.11.4 @@ -335,14 +336,6 @@ deps = # === Integrations === - # Anthropic - anthropic: pytest-asyncio - anthropic-v{0.16,0.28}: httpx<0.28.0 - anthropic-v0.16: anthropic~=0.16.0 - anthropic-v0.28: anthropic~=0.28.0 - anthropic-v0.40: anthropic~=0.40.0 - anthropic-latest: anthropic - # Arq arq-v0.23: arq~=0.23.0 arq-v0.23: pydantic<2 @@ -506,6 +499,15 @@ deps = # integration tests there. # ~~~ AI ~~~ + anthropic-v0.16.0: anthropic==0.16.0 + anthropic-v0.27.0: anthropic==0.27.0 + anthropic-v0.38.0: anthropic==0.38.0 + anthropic-v0.50.0: anthropic==0.50.0 + anthropic: pytest-asyncio + anthropic-v0.16.0: httpx<0.28.0 + anthropic-v0.27.0: httpx<0.28.0 + anthropic-v0.38.0: httpx<0.28.0 + cohere-v5.4.0: cohere==5.4.0 cohere-v5.8.1: cohere==5.8.1 cohere-v5.11.4: cohere==5.11.4 From 970a3503dcf700a8f07b8730ae0c44265238388b Mon Sep 17 00:00:00 2001 From: Ihar Hrachyshka Date: Tue, 29 Apr 2025 10:03:19 -0400 Subject: [PATCH 277/591] tests: fix test_stacktrace_big_recursion failure due to argv (#4346) Sometimes I see the test failing because the event contains `extras` with `sys.argv` key in addition to `exception`. There's probably some state leaking between tests, but regardless this patch should make the test case slightly more robust. Signed-off-by: Ihar Hrachyshka --- tests/test_basics.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/tests/test_basics.py b/tests/test_basics.py index 94ced5013a..7aa2f0f0d5 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -1151,10 +1151,8 @@ def recurse(): (event,) = events assert event["exception"]["values"][0]["stacktrace"] is None - assert event["_meta"] == { - "exception": { - "values": {"0": {"stacktrace": {"": {"rem": [["!config", "x"]]}}}} - } + assert event["_meta"]["exception"] == { + "values": {"0": {"stacktrace": {"": {"rem": [["!config", "x"]]}}}} } # On my machine, it takes about 100-200ms to capture the exception, From 7f013720c08048943595d48bdc46237deb6809aa Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Tue, 29 Apr 2025 11:34:23 -0400 Subject: [PATCH 278/591] chore(ourlogs): Use new transport (#4317) We've added a more efficient transport for logs handling, use that. Solves LOGS-60 --- sentry_sdk/_log_batcher.py | 75 ++++++++++++++++++++++++-------------- sentry_sdk/envelope.py | 8 +--- tests/test_logs.py | 48 ++++++++++++------------ 3 files changed, 73 insertions(+), 58 deletions(-) diff --git a/sentry_sdk/_log_batcher.py b/sentry_sdk/_log_batcher.py index 77efe29a2c..87bebdb226 100644 --- a/sentry_sdk/_log_batcher.py +++ b/sentry_sdk/_log_batcher.py @@ -5,7 +5,7 @@ from typing import Optional, List, Callable, TYPE_CHECKING, Any from sentry_sdk.utils import format_timestamp, safe_repr -from sentry_sdk.envelope import Envelope +from sentry_sdk.envelope import Envelope, Item, PayloadRef if TYPE_CHECKING: from sentry_sdk._types import Log @@ -97,34 +97,36 @@ def flush(self): self._flush() @staticmethod - def _log_to_otel(log): + def _log_to_transport_format(log): # type: (Log) -> Any - def format_attribute(key, val): - # type: (str, int | float | str | bool) -> Any + def format_attribute(val): + # type: (int | float | str | bool) -> Any if isinstance(val, bool): - return {"key": key, "value": {"boolValue": val}} + return {"value": val, "type": "boolean"} if isinstance(val, int): - return {"key": key, "value": {"intValue": str(val)}} + return {"value": val, "type": "integer"} if isinstance(val, float): - return {"key": key, "value": {"doubleValue": val}} + return {"value": val, "type": "double"} if isinstance(val, str): - return {"key": key, "value": {"stringValue": val}} - return {"key": key, "value": {"stringValue": safe_repr(val)}} - - otel_log = { - "severityText": log["severity_text"], - "severityNumber": log["severity_number"], - "body": {"stringValue": log["body"]}, - "timeUnixNano": str(log["time_unix_nano"]), - "attributes": [ - format_attribute(k, v) for (k, v) in log["attributes"].items() - ], + return {"value": val, "type": "string"} + return {"value": safe_repr(val), "type": "string"} + + if "sentry.severity_number" not in log["attributes"]: + log["attributes"]["sentry.severity_number"] = log["severity_number"] + if "sentry.severity_text" not in log["attributes"]: + log["attributes"]["sentry.severity_text"] = log["severity_text"] + + res = { + "timestamp": int(log["time_unix_nano"]) / 1.0e9, + "trace_id": log.get("trace_id", "00000000-0000-0000-0000-000000000000"), + "level": str(log["severity_text"]), + "body": str(log["body"]), + "attributes": { + k: format_attribute(v) for (k, v) in log["attributes"].items() + }, } - if "trace_id" in log: - otel_log["traceId"] = log["trace_id"] - - return otel_log + return res def _flush(self): # type: (...) -> Optional[Envelope] @@ -133,10 +135,27 @@ def _flush(self): headers={"sent_at": format_timestamp(datetime.now(timezone.utc))} ) with self._lock: - for log in self._log_buffer: - envelope.add_log(self._log_to_otel(log)) + if len(self._log_buffer) == 0: + return None + + envelope.add_item( + Item( + type="log", + content_type="application/vnd.sentry.items.log+json", + headers={ + "item_count": len(self._log_buffer), + }, + payload=PayloadRef( + json={ + "items": [ + self._log_to_transport_format(log) + for log in self._log_buffer + ] + } + ), + ) + ) self._log_buffer.clear() - if envelope.items: - self._capture_func(envelope) - return envelope - return None + + self._capture_func(envelope) + return envelope diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index 044d282005..5f7220bf21 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -106,12 +106,6 @@ def add_sessions( # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=sessions), type="sessions")) - def add_log( - self, log # type: Any - ): - # type: (...) -> None - self.add_item(Item(payload=PayloadRef(json=log), type="otel_log")) - def add_item( self, item # type: Item ): @@ -278,7 +272,7 @@ def data_category(self): return "transaction" elif ty == "event": return "error" - elif ty == "otel_log": + elif ty == "log": return "log" elif ty == "client_report": return "internal" diff --git a/tests/test_logs.py b/tests/test_logs.py index 5ede277e3b..c6ef8bcc9d 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -19,42 +19,44 @@ def otel_attributes_to_dict(otel_attrs): - # type: (List[Mapping[str, Any]]) -> Mapping[str, Any] + # type: (Mapping[str, Any]) -> Mapping[str, Any] def _convert_attr(attr): # type: (Mapping[str, Union[str, float, bool]]) -> Any - if "boolValue" in attr: - return bool(attr["boolValue"]) - if "doubleValue" in attr: - return float(attr["doubleValue"]) - if "intValue" in attr: - return int(attr["intValue"]) - if attr["stringValue"].startswith("{"): + if attr["type"] == "boolean": + return attr["value"] + if attr["type"] == "double": + return attr["value"] + if attr["type"] == "integer": + return attr["value"] + if attr["value"].startswith("{"): try: return json.loads(attr["stringValue"]) except ValueError: pass - return str(attr["stringValue"]) + return str(attr["value"]) - return {item["key"]: _convert_attr(item["value"]) for item in otel_attrs} + return {k: _convert_attr(v) for (k, v) in otel_attrs.items()} def envelopes_to_logs(envelopes: List[Envelope]) -> List[Log]: res = [] # type: List[Log] for envelope in envelopes: for item in envelope.items: - if item.type == "otel_log": - log_json = item.payload.json - log = { - "severity_text": log_json["severityText"], - "severity_number": log_json["severityNumber"], - "body": log_json["body"]["stringValue"], - "attributes": otel_attributes_to_dict(log_json["attributes"]), - "time_unix_nano": int(log_json["timeUnixNano"]), - "trace_id": None, - } # type: Log - if "traceId" in log_json: - log["trace_id"] = log_json["traceId"] - res.append(log) + if item.type == "log": + for log_json in item.payload.json["items"]: + log = { + "severity_text": log_json["attributes"]["sentry.severity_text"][ + "value" + ], + "severity_number": int( + log_json["attributes"]["sentry.severity_number"]["value"] + ), + "body": log_json["body"], + "attributes": otel_attributes_to_dict(log_json["attributes"]), + "time_unix_nano": int(float(log_json["timestamp"]) * 1e9), + "trace_id": log_json["trace_id"], + } # type: Log + res.append(log) return res From 2f54dbda2f6356eca20a507c75fdab42c27cc73d Mon Sep 17 00:00:00 2001 From: Colin <161344340+colin-sentry@users.noreply.github.com> Date: Tue, 29 Apr 2025 13:56:00 -0400 Subject: [PATCH 279/591] feat(ourlogs): canonicalize paths from the logger integration (#4336) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit We'd like to allow linking to the 'source code' line in the logs page - this canonicalizes the path relative to the project root (if one is defined) ![Screenshot 2025-04-28 at 12 03 45 PM](https://github.com/user-attachments/assets/89dde691-d9c3-45b2-b289-c42996496bf3) Solves LOGS-58 --- sentry_sdk/integrations/logging.py | 6 +++++- tests/test_logs.py | 31 +++++++++++++++++++++++++++++- 2 files changed, 35 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index bf538ac7c7..46628bb04b 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -355,6 +355,7 @@ def _capture_log_from_record(client, record): # type: (BaseClient, LogRecord) -> None scope = sentry_sdk.get_current_scope() otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) + project_root = client.options["project_root"] attrs = { "sentry.origin": "auto.logger.log", } # type: dict[str, str | bool | float | int] @@ -374,7 +375,10 @@ def _capture_log_from_record(client, record): if record.lineno: attrs["code.line.number"] = record.lineno if record.pathname: - attrs["code.file.path"] = record.pathname + if project_root is not None and record.pathname.startswith(project_root): + attrs["code.file.path"] = record.pathname[len(project_root) + 1 :] + else: + attrs["code.file.path"] = record.pathname if record.funcName: attrs["code.function.name"] = record.funcName diff --git a/tests/test_logs.py b/tests/test_logs.py index c6ef8bcc9d..49ffd31ec7 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -346,7 +346,6 @@ def test_logging_errors(sentry_init, capture_envelopes): error_event_2 = envelopes[1].items[0].payload.json assert error_event_2["level"] == "error" - print(envelopes) logs = envelopes_to_logs(envelopes) assert logs[0]["severity_text"] == "error" assert "sentry.message.template" not in logs[0]["attributes"] @@ -364,6 +363,36 @@ def test_logging_errors(sentry_init, capture_envelopes): assert len(logs) == 2 +def test_log_strips_project_root(sentry_init, capture_envelopes): + """ + The python logger should strip project roots from the log record path + """ + sentry_init( + _experiments={"enable_logs": True}, + project_root="/custom/test", + ) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.handle( + logging.LogRecord( + name="test-logger", + level=logging.WARN, + pathname="/custom/test/blah/path.py", + lineno=123, + msg="This is a test log with a custom pathname", + args=(), + exc_info=None, + ) + ) + get_client().flush() + + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + attrs = logs[0]["attributes"] + assert attrs["code.file.path"] == "blah/path.py" + + def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): """ If you log >100 logs, it should automatically trigger a flush. From 18a110433668d26fd341b3c87eecea7ff212b7f3 Mon Sep 17 00:00:00 2001 From: Ihar Hrachyshka Date: Wed, 30 Apr 2025 03:15:54 -0400 Subject: [PATCH 280/591] tests: bump test timeout for recursion stacktrace extract to 2s (#4351) In some loaded environments, the test may take slightly longer than 1s to extract the stacktrace. This was noticed in nixpkgs build system where the load is generally high due to high build parallelism and resource constraints. I was sometimes getting failures because the time it took was e.g. ~1.2s (less than current timeout of 1s). Disclosure: we'll probably end up disabling the test in nixpkgs anyway because we try to avoid time sensitive tests. Regardless, this bump may help someone else in a similar situation or environment. Signed-off-by: Ihar Hrachyshka --- tests/test_basics.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/test_basics.py b/tests/test_basics.py index 7aa2f0f0d5..0fdf9f811f 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -1158,5 +1158,5 @@ def recurse(): # On my machine, it takes about 100-200ms to capture the exception, # so this limit should be generous enough. assert ( - capture_end_time - capture_start_time < 10**9 + capture_end_time - capture_start_time < 10**9 * 2 ), "stacktrace capture took too long, check that frame limit is set correctly" From ebde4760e2403d3f5296bd464485afc7dee4ca4d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 5 May 2025 16:54:15 +0200 Subject: [PATCH 281/591] Put feature flags on isolation scope (#4363) Feature flags should life on the isolation Scope. This has been first [implemented in SDK 3.0](https://github.com/getsentry/sentry-python/pull/4353) and is now back ported to 2.x. --- docs/api.rst | 2 +- sentry_sdk/__init__.py | 1 + sentry_sdk/api.py | 15 ++++++ sentry_sdk/feature_flags.py | 2 +- tests/integrations/fastapi/test_fastapi.py | 40 +++++++++++++++ tests/test_feature_flags.py | 57 ++++++++++++++++++++++ 6 files changed, 115 insertions(+), 2 deletions(-) diff --git a/docs/api.rst b/docs/api.rst index 87c2535abd..a6fb49346d 100644 --- a/docs/api.rst +++ b/docs/api.rst @@ -25,6 +25,7 @@ Capturing Data Enriching Events ================ +.. autofunction:: sentry_sdk.api.add_attachment .. autofunction:: sentry_sdk.api.add_breadcrumb .. autofunction:: sentry_sdk.api.set_context .. autofunction:: sentry_sdk.api.set_extra @@ -63,4 +64,3 @@ Managing Scope (advanced) .. autofunction:: sentry_sdk.api.push_scope .. autofunction:: sentry_sdk.api.new_scope - diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index b4859cc5d2..9fd7253fc2 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -15,6 +15,7 @@ "integrations", # From sentry_sdk.api "init", + "add_attachment", "add_breadcrumb", "capture_event", "capture_exception", diff --git a/sentry_sdk/api.py b/sentry_sdk/api.py index a6b3c293dc..e56109cbd0 100644 --- a/sentry_sdk/api.py +++ b/sentry_sdk/api.py @@ -51,6 +51,7 @@ def overload(x): # When changing this, update __all__ in __init__.py too __all__ = [ "init", + "add_attachment", "add_breadcrumb", "capture_event", "capture_exception", @@ -184,6 +185,20 @@ def capture_exception( return get_current_scope().capture_exception(error, scope=scope, **scope_kwargs) +@scopemethod +def add_attachment( + bytes=None, # type: Union[None, bytes, Callable[[], bytes]] + filename=None, # type: Optional[str] + path=None, # type: Optional[str] + content_type=None, # type: Optional[str] + add_to_transactions=False, # type: bool +): + # type: (...) -> None + return get_isolation_scope().add_attachment( + bytes, filename, path, content_type, add_to_transactions + ) + + @scopemethod def add_breadcrumb( crumb=None, # type: Optional[Breadcrumb] diff --git a/sentry_sdk/feature_flags.py b/sentry_sdk/feature_flags.py index dd8d41c32e..eb53acae5d 100644 --- a/sentry_sdk/feature_flags.py +++ b/sentry_sdk/feature_flags.py @@ -64,7 +64,7 @@ def add_feature_flag(flag, result): Records a flag and its value to be sent on subsequent error events. We recommend you do this on flag evaluations. Flags are buffered per Sentry scope. """ - flags = sentry_sdk.get_current_scope().flags + flags = sentry_sdk.get_isolation_scope().flags flags.set(flag, result) span = sentry_sdk.get_current_span() diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index 95838b1009..3d79da92cc 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -10,7 +10,9 @@ from fastapi.testclient import TestClient from fastapi.middleware.trustedhost import TrustedHostMiddleware +import sentry_sdk from sentry_sdk import capture_message +from sentry_sdk.feature_flags import add_feature_flag from sentry_sdk.integrations.asgi import SentryAsgiMiddleware from sentry_sdk.integrations.fastapi import FastApiIntegration from sentry_sdk.integrations.starlette import StarletteIntegration @@ -714,3 +716,41 @@ async def subapp_route(): assert event["transaction"] == "/subapp" else: assert event["transaction"].endswith("subapp_route") + + +@pytest.mark.asyncio +async def test_feature_flags(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + integrations=[StarletteIntegration(), FastApiIntegration()], + ) + + events = capture_events() + + app = FastAPI() + + @app.get("/error") + async def _error(): + add_feature_flag("hello", False) + + with sentry_sdk.start_span(name="test-span"): + with sentry_sdk.start_span(name="test-span-2"): + raise ValueError("something is wrong!") + + try: + client = TestClient(app) + client.get("/error") + except ValueError: + pass + + found = False + for event in events: + if "exception" in event.keys(): + assert event["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + found = True + + assert found, "No event with exception found" diff --git a/tests/test_feature_flags.py b/tests/test_feature_flags.py index 1b0ed13d49..e0ab1e254e 100644 --- a/tests/test_feature_flags.py +++ b/tests/test_feature_flags.py @@ -31,6 +31,63 @@ def test_featureflags_integration(sentry_init, capture_events, uninstall_integra } +@pytest.mark.asyncio +async def test_featureflags_integration_spans_async(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + ) + events = capture_events() + + add_feature_flag("hello", False) + + try: + with sentry_sdk.start_span(name="test-span"): + with sentry_sdk.start_span(name="test-span-2"): + raise ValueError("something wrong!") + except ValueError as e: + sentry_sdk.capture_exception(e) + + found = False + for event in events: + if "exception" in event.keys(): + assert event["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + found = True + + assert found, "No event with exception found" + + +def test_featureflags_integration_spans_sync(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + ) + events = capture_events() + + add_feature_flag("hello", False) + + try: + with sentry_sdk.start_span(name="test-span"): + with sentry_sdk.start_span(name="test-span-2"): + raise ValueError("something wrong!") + except ValueError as e: + sentry_sdk.capture_exception(e) + + found = False + for event in events: + if "exception" in event.keys(): + assert event["contexts"]["flags"] == { + "values": [ + {"flag": "hello", "result": False}, + ] + } + found = True + + assert found, "No event with exception found" + + def test_featureflags_integration_threaded( sentry_init, capture_events, uninstall_integration ): From c25d4ff4e3ed93dc0e30bd87c91448d5398be1a2 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 6 May 2025 12:10:33 +0200 Subject: [PATCH 282/591] build(deps): bump actions/create-github-app-token from 2.0.2 to 2.0.6 (#4358) Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index a0e39a5784..34815da549 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@3ff1caaa28b64c9cc276ce0a02e2ff584f3900c5 # v2.0.2 + uses: actions/create-github-app-token@df432ceedc7162793a195dd1713ff69aefc7379e # v2.0.6 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From b16fa5ffbad39843ebd2e9bc4ea6e91c0c9aa192 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 6 May 2025 13:04:09 +0200 Subject: [PATCH 283/591] tests: Regular tox update (#4367) Regular tox.ini update. Note: the DB (latest) CI being red has nothing to do with the changes in this PR (redis) --- tox.ini | 16 +++++++++------- 1 file changed, 9 insertions(+), 7 deletions(-) diff --git a/tox.ini b/tox.ini index 4c05bcaa75..332f541793 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-04-29T08:35:44.624881+00:00 +# Last generated: 2025-05-06T10:23:50.156629+00:00 [tox] requires = @@ -157,7 +157,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.12.0 + {py3.9,py3.12,py3.13}-pymongo-v4.12.1 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -275,7 +275,7 @@ envlist = {py3.8,py3.10,py3.11}-litestar-v2.0.1 {py3.8,py3.11,py3.12}-litestar-v2.5.5 {py3.8,py3.11,py3.12}-litestar-v2.10.0 - {py3.8,py3.12,py3.13}-litestar-v2.15.2 + {py3.8,py3.12,py3.13}-litestar-v2.16.0 {py3.6}-pyramid-v1.8.6 {py3.6,py3.8,py3.9}-pyramid-v1.10.8 @@ -290,6 +290,7 @@ envlist = {py3.6,py3.8,py3.9}-tornado-v6.1 {py3.7,py3.9,py3.10}-tornado-v6.2 {py3.8,py3.10,py3.11}-tornado-v6.4.2 + {py3.9,py3.12,py3.13}-tornado-v6.5b1 # ~~~ Misc ~~~ @@ -299,7 +300,7 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.30 + {py3.8,py3.11,py3.12}-trytond-v7.0.31 {py3.9,py3.12,py3.13}-trytond-v7.6.0 {py3.7,py3.12,py3.13}-typer-v0.15.3 @@ -525,7 +526,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.12.0: pymongo==4.12.0 + pymongo-v4.12.1: pymongo==4.12.1 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -713,7 +714,7 @@ deps = litestar-v2.0.1: litestar==2.0.1 litestar-v2.5.5: litestar==2.5.5 litestar-v2.10.0: litestar==2.10.0 - litestar-v2.15.2: litestar==2.15.2 + litestar-v2.16.0: litestar==2.16.0 litestar: pytest-asyncio litestar: python-multipart litestar: requests @@ -741,6 +742,7 @@ deps = tornado-v6.1: tornado==6.1 tornado-v6.2: tornado==6.2 tornado-v6.4.2: tornado==6.4.2 + tornado-v6.5b1: tornado==6.5b1 tornado: pytest tornado-v6.0.4: pytest<8.2 tornado-v6.1: pytest<8.2 @@ -755,7 +757,7 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.30: trytond==7.0.30 + trytond-v7.0.31: trytond==7.0.31 trytond-v7.6.0: trytond==7.6.0 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 From 2df4dc7589da9c9f6a253fb07e02c2a757ec63c2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 8 May 2025 12:57:06 +0200 Subject: [PATCH 284/591] Pin snowballstemmer for now (#4372) Make apidocs buildable again --- requirements-docs.txt | 1 + 1 file changed, 1 insertion(+) diff --git a/requirements-docs.txt b/requirements-docs.txt index 81e04ba3ef..a662a0d83f 100644 --- a/requirements-docs.txt +++ b/requirements-docs.txt @@ -3,3 +3,4 @@ shibuya sphinx<8.2 sphinx-autodoc-typehints[type_comments]>=1.8.0 typing-extensions +snowballstemmer<3.0 From ca5ba8957101e5b1b8ac76d1c94a99e5db95bd9c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 8 May 2025 13:14:14 +0200 Subject: [PATCH 285/591] Fix Discord link (#4371) --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 10bc8eb2ed..a3afdc6e72 100644 --- a/README.md +++ b/README.md @@ -6,7 +6,7 @@ _Bad software is everywhere, and we're tired of it. Sentry is on a mission to help developers write better software faster, so we can get back to enjoying technology. If you want to join us [**Check out our open positions**](https://sentry.io/careers/)_. -[![Discord](https://img.shields.io/discord/621778831602221064?logo=discord&labelColor=%20%235462eb&logoColor=%20%23f5f5f5&color=%20%235462eb)](https://discord.gg/wdNEHETs87) +[![Discord](https://img.shields.io/discord/621778831602221064?logo=discord&labelColor=%20%235462eb&logoColor=%20%23f5f5f5&color=%20%235462eb)](https://discord.com/invite/Ww9hbqr) [![Twitter Follow](https://img.shields.io/twitter/follow/getsentry?label=@getsentry&style=social)](https://twitter.com/intent/follow?screen_name=getsentry) [![PyPi page link -- version](https://img.shields.io/pypi/v/sentry-sdk.svg)](https://pypi.python.org/pypi/sentry-sdk) python @@ -106,7 +106,7 @@ If you encounter issues or need help setting up or configuring the SDK, don't he Here are all resources to help you make the most of Sentry: - [Documentation](https://docs.sentry.io/platforms/python/) - Official documentation to get started. -- [Discord](https://img.shields.io/discord/621778831602221064) - Join our Discord community. +- [Discord](https://discord.com/invite/Ww9hbqr) - Join our Discord community. - [X/Twitter](https://twitter.com/intent/follow?screen_name=getsentry) - Follow us on X (Twitter) for updates. - [Stack Overflow](https://stackoverflow.com/questions/tagged/sentry) - Questions and answers related to Sentry. From cb824834e40921e9d488f81afc18495d811883a8 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 9 May 2025 10:34:09 +0200 Subject: [PATCH 286/591] Make use of `SPANDATA` consistent (#4373) The AI integrations sometimes used plain strings for setting `SPANDATA` attributes. Changed to always use `SPANDATA`. --- sentry_sdk/ai/monitoring.py | 7 ++- sentry_sdk/consts.py | 63 ++++++++++++++++++- sentry_sdk/integrations/cohere.py | 20 +++--- sentry_sdk/integrations/huggingface_hub.py | 4 +- sentry_sdk/integrations/openai.py | 8 +-- .../integrations/anthropic/test_anthropic.py | 14 ++--- tests/integrations/cohere/test_cohere.py | 29 ++++----- .../huggingface_hub/test_huggingface_hub.py | 17 ++--- .../integrations/langchain/test_langchain.py | 26 ++++---- tests/integrations/openai/test_openai.py | 41 ++++++------ 10 files changed, 147 insertions(+), 82 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index 860833b8f5..ed33acd0f1 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -1,6 +1,7 @@ import inspect from functools import wraps +from sentry_sdk.consts import SPANDATA import sentry_sdk.utils from sentry_sdk import start_span from sentry_sdk.tracing import Span @@ -39,7 +40,7 @@ def sync_wrapped(*args, **kwargs): for k, v in kwargs.pop("sentry_data", {}).items(): span.set_data(k, v) if curr_pipeline: - span.set_data("ai.pipeline.name", curr_pipeline) + span.set_data(SPANDATA.AI_PIPELINE_NAME, curr_pipeline) return f(*args, **kwargs) else: _ai_pipeline_name.set(description) @@ -68,7 +69,7 @@ async def async_wrapped(*args, **kwargs): for k, v in kwargs.pop("sentry_data", {}).items(): span.set_data(k, v) if curr_pipeline: - span.set_data("ai.pipeline.name", curr_pipeline) + span.set_data(SPANDATA.AI_PIPELINE_NAME, curr_pipeline) return await f(*args, **kwargs) else: _ai_pipeline_name.set(description) @@ -100,7 +101,7 @@ def record_token_usage( # type: (Span, Optional[int], Optional[int], Optional[int]) -> None ai_pipeline_name = get_ai_pipeline_name() if ai_pipeline_name: - span.set_data("ai.pipeline.name", ai_pipeline_name) + span.set_data(SPANDATA.AI_PIPELINE_NAME, ai_pipeline_name) if prompt_tokens is not None: span.set_measurement("ai_prompt_tokens_used", value=prompt_tokens) if completion_tokens is not None: diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index e1f18fe4ae..e3c29fc2d4 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -187,7 +187,7 @@ class SPANDATA: For an AI model call, the format of the response """ - AI_LOGIT_BIAS = "ai.response_format" + AI_LOGIT_BIAS = "ai.logit_bias" """ For an AI model call, the logit bias """ @@ -204,7 +204,6 @@ class SPANDATA: Minimize pre-processing done to the prompt sent to the LLM. Example: true """ - AI_RESPONSES = "ai.responses" """ The responses to an AI model call. Always as a list. @@ -217,6 +216,66 @@ class SPANDATA: Example: 123.45 """ + AI_CITATIONS = "ai.citations" + """ + References or sources cited by the AI model in its response. + Example: ["Smith et al. 2020", "Jones 2019"] + """ + + AI_DOCUMENTS = "ai.documents" + """ + Documents or content chunks used as context for the AI model. + Example: ["doc1.txt", "doc2.pdf"] + """ + + AI_SEARCH_QUERIES = "ai.search_queries" + """ + Queries used to search for relevant context or documents. + Example: ["climate change effects", "renewable energy"] + """ + + AI_SEARCH_RESULTS = "ai.search_results" + """ + Results returned from search queries for context. + Example: ["Result 1", "Result 2"] + """ + + AI_GENERATION_ID = "ai.generation_id" + """ + Unique identifier for the completion. + Example: "gen_123abc" + """ + + AI_SEARCH_REQUIRED = "ai.is_search_required" + """ + Boolean indicating if the model needs to perform a search. + Example: true + """ + + AI_FINISH_REASON = "ai.finish_reason" + """ + The reason why the model stopped generating. + Example: "length" + """ + + AI_PIPELINE_NAME = "ai.pipeline.name" + """ + Name of the AI pipeline or chain being executed. + Example: "qa-pipeline" + """ + + AI_TEXTS = "ai.texts" + """ + Raw text inputs provided to the model. + Example: ["What is machine learning?"] + """ + + AI_WARNINGS = "ai.warnings" + """ + Warning messages generated during model execution. + Example: ["Token limit exceeded"] + """ + DB_NAME = "db.name" """ The name of the database being accessed. For commands that switch the database, this should be set to the target database (even if the command fails). diff --git a/sentry_sdk/integrations/cohere.py b/sentry_sdk/integrations/cohere.py index b4c2af91da..433b285bf0 100644 --- a/sentry_sdk/integrations/cohere.py +++ b/sentry_sdk/integrations/cohere.py @@ -52,17 +52,17 @@ } COLLECTED_CHAT_RESP_ATTRS = { - "generation_id": "ai.generation_id", - "is_search_required": "ai.is_search_required", - "finish_reason": "ai.finish_reason", + "generation_id": SPANDATA.AI_GENERATION_ID, + "is_search_required": SPANDATA.AI_SEARCH_REQUIRED, + "finish_reason": SPANDATA.AI_FINISH_REASON, } COLLECTED_PII_CHAT_RESP_ATTRS = { - "citations": "ai.citations", - "documents": "ai.documents", - "search_queries": "ai.search_queries", - "search_results": "ai.search_results", - "tool_calls": "ai.tool_calls", + "citations": SPANDATA.AI_CITATIONS, + "documents": SPANDATA.AI_DOCUMENTS, + "search_queries": SPANDATA.AI_SEARCH_QUERIES, + "search_results": SPANDATA.AI_SEARCH_RESULTS, + "tool_calls": SPANDATA.AI_TOOL_CALLS, } @@ -127,7 +127,7 @@ def collect_chat_response_fields(span, res, include_pii): ) if hasattr(res.meta, "warnings"): - set_data_normalized(span, "ai.warnings", res.meta.warnings) + set_data_normalized(span, SPANDATA.AI_WARNINGS, res.meta.warnings) @wraps(f) def new_chat(*args, **kwargs): @@ -238,7 +238,7 @@ def new_embed(*args, **kwargs): should_send_default_pii() and integration.include_prompts ): if isinstance(kwargs["texts"], str): - set_data_normalized(span, "ai.texts", [kwargs["texts"]]) + set_data_normalized(span, SPANDATA.AI_TEXTS, [kwargs["texts"]]) elif ( isinstance(kwargs["texts"], list) and len(kwargs["texts"]) > 0 diff --git a/sentry_sdk/integrations/huggingface_hub.py b/sentry_sdk/integrations/huggingface_hub.py index d09f6e2163..dfac77e996 100644 --- a/sentry_sdk/integrations/huggingface_hub.py +++ b/sentry_sdk/integrations/huggingface_hub.py @@ -97,7 +97,7 @@ def new_text_generation(*args, **kwargs): if should_send_default_pii() and integration.include_prompts: set_data_normalized( span, - "ai.responses", + SPANDATA.AI_RESPONSES, [res], ) span.__exit__(None, None, None) @@ -107,7 +107,7 @@ def new_text_generation(*args, **kwargs): if should_send_default_pii() and integration.include_prompts: set_data_normalized( span, - "ai.responses", + SPANDATA.AI_RESPONSES, [res.generated_text], ) if res.details is not None and res.details.generated_tokens > 0: diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 61d335b170..e95753f6e1 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -155,7 +155,7 @@ def _new_chat_completion_common(f, *args, **kwargs): if should_send_default_pii() and integration.include_prompts: set_data_normalized( span, - "ai.responses", + SPANDATA.AI_RESPONSES, list(map(lambda x: x.message, res.choices)), ) _calculate_chat_completion_usage( @@ -329,15 +329,15 @@ def _new_embeddings_create_common(f, *args, **kwargs): should_send_default_pii() and integration.include_prompts ): if isinstance(kwargs["input"], str): - set_data_normalized(span, "ai.input_messages", [kwargs["input"]]) + set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, [kwargs["input"]]) elif ( isinstance(kwargs["input"], list) and len(kwargs["input"]) > 0 and isinstance(kwargs["input"][0], str) ): - set_data_normalized(span, "ai.input_messages", kwargs["input"]) + set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, kwargs["input"]) if "model" in kwargs: - set_data_normalized(span, "ai.model_id", kwargs["model"]) + set_data_normalized(span, SPANDATA.AI_MODEL_ID, kwargs["model"]) response = yield f, args, kwargs diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index 7f6622a1ba..9ab0f879d1 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -128,7 +128,7 @@ def test_nonstreaming_create_message( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 20 assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 - assert span["data"]["ai.streaming"] is False + assert span["data"][SPANDATA.AI_STREAMING] is False @pytest.mark.asyncio @@ -196,7 +196,7 @@ async def test_nonstreaming_create_message_async( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 20 assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 - assert span["data"]["ai.streaming"] is False + assert span["data"][SPANDATA.AI_STREAMING] is False @pytest.mark.parametrize( @@ -296,7 +296,7 @@ def test_streaming_create_message( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 30 assert span["measurements"]["ai_total_tokens_used"]["value"] == 40 - assert span["data"]["ai.streaming"] is True + assert span["data"][SPANDATA.AI_STREAMING] is True @pytest.mark.asyncio @@ -399,7 +399,7 @@ async def test_streaming_create_message_async( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 30 assert span["measurements"]["ai_total_tokens_used"]["value"] == 40 - assert span["data"]["ai.streaming"] is True + assert span["data"][SPANDATA.AI_STREAMING] is True @pytest.mark.skipif( @@ -528,7 +528,7 @@ def test_streaming_create_message_with_input_json_delta( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 366 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 51 assert span["measurements"]["ai_total_tokens_used"]["value"] == 417 - assert span["data"]["ai.streaming"] is True + assert span["data"][SPANDATA.AI_STREAMING] is True @pytest.mark.asyncio @@ -665,7 +665,7 @@ async def test_streaming_create_message_with_input_json_delta_async( assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 366 assert span["measurements"]["ai_completion_tokens_used"]["value"] == 51 assert span["measurements"]["ai_total_tokens_used"]["value"] == 417 - assert span["data"]["ai.streaming"] is True + assert span["data"][SPANDATA.AI_STREAMING] is True def test_exception_message_create(sentry_init, capture_events): @@ -810,7 +810,7 @@ def test_add_ai_data_to_span_with_input_json_delta(sentry_init): assert span._data.get(SPANDATA.AI_RESPONSES) == [ {"type": "text", "text": "{'test': 'data','more': 'json'}"} ] - assert span._data.get("ai.streaming") is True + assert span._data.get(SPANDATA.AI_STREAMING) is True assert span._measurements.get("ai_prompt_tokens_used")["value"] == 10 assert span._measurements.get("ai_completion_tokens_used")["value"] == 20 assert span._measurements.get("ai_total_tokens_used")["value"] == 30 diff --git a/tests/integrations/cohere/test_cohere.py b/tests/integrations/cohere/test_cohere.py index c0dff2214e..6c1185a28e 100644 --- a/tests/integrations/cohere/test_cohere.py +++ b/tests/integrations/cohere/test_cohere.py @@ -5,6 +5,7 @@ from cohere import Client, ChatMessage from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.cohere import CohereIntegration from unittest import mock # python 3.3 and above @@ -53,15 +54,15 @@ def test_nonstreaming_chat( assert tx["type"] == "transaction" span = tx["spans"][0] assert span["op"] == "ai.chat_completions.create.cohere" - assert span["data"]["ai.model_id"] == "some-model" + assert span["data"][SPANDATA.AI_MODEL_ID] == "some-model" if send_default_pii and include_prompts: - assert "some context" in span["data"]["ai.input_messages"][0]["content"] - assert "hello" in span["data"]["ai.input_messages"][1]["content"] - assert "the model response" in span["data"]["ai.responses"] + assert "some context" in span["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES][1]["content"] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 @@ -124,15 +125,15 @@ def test_streaming_chat(sentry_init, capture_events, send_default_pii, include_p assert tx["type"] == "transaction" span = tx["spans"][0] assert span["op"] == "ai.chat_completions.create.cohere" - assert span["data"]["ai.model_id"] == "some-model" + assert span["data"][SPANDATA.AI_MODEL_ID] == "some-model" if send_default_pii and include_prompts: - assert "some context" in span["data"]["ai.input_messages"][0]["content"] - assert "hello" in span["data"]["ai.input_messages"][1]["content"] - assert "the model response" in span["data"]["ai.responses"] + assert "some context" in span["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES][1]["content"] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 @@ -194,9 +195,9 @@ def test_embed(sentry_init, capture_events, send_default_pii, include_prompts): span = tx["spans"][0] assert span["op"] == "ai.embeddings.create.cohere" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] else: - assert "ai.input_messages" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index 090b0e4f3e..ee47cc7e56 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -8,6 +8,7 @@ from huggingface_hub.errors import OverloadedError from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.huggingface_hub import HuggingfaceHubIntegration @@ -67,11 +68,11 @@ def test_nonstreaming_chat_completion( assert span["op"] == "ai.chat_completions.create.huggingface_hub" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"] - assert "the model response" in span["data"]["ai.responses"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] if details_arg: assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 @@ -126,11 +127,11 @@ def test_streaming_chat_completion( assert span["op"] == "ai.chat_completions.create.huggingface_hub" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"] - assert "the model response" in span["data"]["ai.responses"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] if details_arg: assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index b9e5705b88..3f1b3b1da5 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -3,6 +3,8 @@ import pytest +from sentry_sdk.consts import SPANDATA + try: # Langchain >= 0.2 from langchain_openai import ChatOpenAI @@ -189,23 +191,23 @@ def test_langchain_agent( if send_default_pii and include_prompts: assert ( "You are very powerful" - in chat_spans[0]["data"]["ai.input_messages"][0]["content"] + in chat_spans[0]["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] ) - assert "5" in chat_spans[0]["data"]["ai.responses"] - assert "word" in tool_exec_span["data"]["ai.input_messages"] - assert 5 == int(tool_exec_span["data"]["ai.responses"]) + assert "5" in chat_spans[0]["data"][SPANDATA.AI_RESPONSES] + assert "word" in tool_exec_span["data"][SPANDATA.AI_INPUT_MESSAGES] + assert 5 == int(tool_exec_span["data"][SPANDATA.AI_RESPONSES]) assert ( "You are very powerful" - in chat_spans[1]["data"]["ai.input_messages"][0]["content"] + in chat_spans[1]["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] ) - assert "5" in chat_spans[1]["data"]["ai.responses"] + assert "5" in chat_spans[1]["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in chat_spans[0].get("data", {}) - assert "ai.responses" not in chat_spans[0].get("data", {}) - assert "ai.input_messages" not in chat_spans[1].get("data", {}) - assert "ai.responses" not in chat_spans[1].get("data", {}) - assert "ai.input_messages" not in tool_exec_span.get("data", {}) - assert "ai.responses" not in tool_exec_span.get("data", {}) + assert SPANDATA.AI_INPUT_MESSAGES not in chat_spans[0].get("data", {}) + assert SPANDATA.AI_RESPONSES not in chat_spans[0].get("data", {}) + assert SPANDATA.AI_INPUT_MESSAGES not in chat_spans[1].get("data", {}) + assert SPANDATA.AI_RESPONSES not in chat_spans[1].get("data", {}) + assert SPANDATA.AI_INPUT_MESSAGES not in tool_exec_span.get("data", {}) + assert SPANDATA.AI_RESPONSES not in tool_exec_span.get("data", {}) def test_langchain_error(sentry_init, capture_events): diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 011192e49f..3fdc138f39 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -7,6 +7,7 @@ from openai.types.create_embedding_response import Usage as EmbeddingTokenUsage from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.openai import ( OpenAIIntegration, _calculate_chat_completion_usage, @@ -83,11 +84,11 @@ def test_nonstreaming_chat_completion( assert span["op"] == "ai.chat_completions.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"]["content"] - assert "the model response" in span["data"]["ai.responses"]["content"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES]["content"] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 @@ -125,11 +126,11 @@ async def test_nonstreaming_chat_completion_async( assert span["op"] == "ai.chat_completions.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"]["content"] - assert "the model response" in span["data"]["ai.responses"]["content"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] + assert "the model response" in span["data"][SPANDATA.AI_RESPONSES]["content"] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 @@ -218,11 +219,11 @@ def test_streaming_chat_completion( assert span["op"] == "ai.chat_completions.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"]["content"] - assert "hello world" in span["data"]["ai.responses"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] + assert "hello world" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import @@ -314,11 +315,11 @@ async def test_streaming_chat_completion_async( assert span["op"] == "ai.chat_completions.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"]["content"] - assert "hello world" in span["data"]["ai.responses"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] + assert "hello world" in span["data"][SPANDATA.AI_RESPONSES] else: - assert "ai.input_messages" not in span["data"] - assert "ai.responses" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.AI_RESPONSES not in span["data"] try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import @@ -404,9 +405,9 @@ def test_embeddings_create( span = tx["spans"][0] assert span["op"] == "ai.embeddings.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] else: - assert "ai.input_messages" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 @@ -452,9 +453,9 @@ async def test_embeddings_create_async( span = tx["spans"][0] assert span["op"] == "ai.embeddings.create.openai" if send_default_pii and include_prompts: - assert "hello" in span["data"]["ai.input_messages"] + assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] else: - assert "ai.input_messages" not in span["data"] + assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 From de6856f5b06d5d516fac5655b052f252e0b62cb3 Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Fri, 9 May 2025 08:35:44 -0400 Subject: [PATCH 287/591] feat(logs): Forward extra from logger as attributes (#4374) resolves https://linear.app/getsentry/issue/LOGS-101 --- sentry_sdk/integrations/logging.py | 10 ++-- tests/test_logs.py | 74 +++++++++++++++++++++++++++++- 2 files changed, 77 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 46628bb04b..74baf3d33a 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -348,17 +348,15 @@ def emit(self, record): if not client.options["_experiments"].get("enable_logs", False): return - SentryLogsHandler._capture_log_from_record(client, record) + self._capture_log_from_record(client, record) - @staticmethod - def _capture_log_from_record(client, record): + def _capture_log_from_record(self, client, record): # type: (BaseClient, LogRecord) -> None scope = sentry_sdk.get_current_scope() otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) project_root = client.options["project_root"] - attrs = { - "sentry.origin": "auto.logger.log", - } # type: dict[str, str | bool | float | int] + attrs = self._extra_from_record(record) # type: Any + attrs["sentry.origin"] = "auto.logger.log" if isinstance(record.msg, str): attrs["sentry.message.template"] = record.msg if record.args is not None: diff --git a/tests/test_logs.py b/tests/test_logs.py index 49ffd31ec7..1f6b07e762 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -30,7 +30,7 @@ def _convert_attr(attr): return attr["value"] if attr["value"].startswith("{"): try: - return json.loads(attr["stringValue"]) + return json.loads(attr["value"]) except ValueError: pass return str(attr["value"]) @@ -393,6 +393,78 @@ def test_log_strips_project_root(sentry_init, capture_envelopes): assert attrs["code.file.path"] == "blah/path.py" +def test_logger_with_all_attributes(sentry_init, capture_envelopes): + """ + The python logger should be able to log all attributes, including extra data. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning( + "log #%d", + 1, + extra={"foo": "bar", "numeric": 42, "more_complex": {"nested": "data"}}, + ) + get_client().flush() + + logs = envelopes_to_logs(envelopes) + + attributes = logs[0]["attributes"] + + assert "process.pid" in attributes + assert isinstance(attributes["process.pid"], int) + del attributes["process.pid"] + + assert "sentry.release" in attributes + assert isinstance(attributes["sentry.release"], str) + del attributes["sentry.release"] + + assert "server.address" in attributes + assert isinstance(attributes["server.address"], str) + del attributes["server.address"] + + assert "thread.id" in attributes + assert isinstance(attributes["thread.id"], int) + del attributes["thread.id"] + + assert "code.file.path" in attributes + assert isinstance(attributes["code.file.path"], str) + del attributes["code.file.path"] + + assert "code.function.name" in attributes + assert isinstance(attributes["code.function.name"], str) + del attributes["code.function.name"] + + assert "code.line.number" in attributes + assert isinstance(attributes["code.line.number"], int) + del attributes["code.line.number"] + + assert "process.executable.name" in attributes + assert isinstance(attributes["process.executable.name"], str) + del attributes["process.executable.name"] + + assert "thread.name" in attributes + assert isinstance(attributes["thread.name"], str) + del attributes["thread.name"] + + # Assert on the remaining non-dynamic attributes. + assert attributes == { + "foo": "bar", + "numeric": 42, + "more_complex": "{'nested': 'data'}", + "logger.name": "test-logger", + "sentry.origin": "auto.logger.log", + "sentry.message.template": "log #%d", + "sentry.message.parameters.0": 1, + "sentry.environment": "production", + "sentry.sdk.name": "sentry.python", + "sentry.sdk.version": VERSION, + "sentry.severity_number": 13, + "sentry.severity_text": "warn", + } + + def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): """ If you log >100 logs, it should automatically trigger a flush. From 53f413e4cb0ec700a4dfa693557afa63aed47c10 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 12 May 2025 07:39:55 +0000 Subject: [PATCH 288/591] release: 2.28.0 --- CHANGELOG.md | 20 ++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 23 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 786a9a34e5..17e1086b80 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,25 @@ # Changelog +## 2.28.0 + +### Various fixes & improvements + +- feat(logs): Forward extra from logger as attributes (#4374) by @AbhiPrasad +- Make use of `SPANDATA` consistent (#4373) by @antonpirker +- Fix Discord link (#4371) by @sentrivana +- Pin snowballstemmer for now (#4372) by @sentrivana +- tests: Regular tox update (#4367) by @sentrivana +- build(deps): bump actions/create-github-app-token from 2.0.2 to 2.0.6 (#4358) by @dependabot +- Put feature flags on isolation scope (#4363) by @antonpirker +- tests: bump test timeout for recursion stacktrace extract to 2s (#4351) by @booxter +- feat(ourlogs): canonicalize paths from the logger integration (#4336) by @colin-sentry +- chore(ourlogs): Use new transport (#4317) by @colin-sentry +- tests: fix test_stacktrace_big_recursion failure due to argv (#4346) by @booxter +- tests: Move anthropic under toxgen (#4348) by @sentrivana +- Deprecate `set_measurement()` API. (#3934) by @antonpirker +- tests: Update tox.ini (#4347) by @sentrivana +- Update GH issue templates for Linear compatibility (#4328) by @stephanie-anderson + ## 2.27.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 709f557d16..34c88ae6cd 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.27.0" +release = "2.28.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index e3c29fc2d4..2ceec2738b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1025,4 +1025,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.27.0" +VERSION = "2.28.0" diff --git a/setup.py b/setup.py index 877585472b..8fd1ae6293 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.27.0", + version="2.28.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From c7a17a06836587a89766bb934ebf11d4b87d6353 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 12 May 2025 09:44:22 +0200 Subject: [PATCH 289/591] Update changelog.md --- CHANGELOG.md | 24 ++++++++++++------------ 1 file changed, 12 insertions(+), 12 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 17e1086b80..81f749d83a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,21 +4,21 @@ ### Various fixes & improvements -- feat(logs): Forward extra from logger as attributes (#4374) by @AbhiPrasad -- Make use of `SPANDATA` consistent (#4373) by @antonpirker -- Fix Discord link (#4371) by @sentrivana -- Pin snowballstemmer for now (#4372) by @sentrivana +- fix(logs): Forward `extra` from logger as attributes (#4374) by @AbhiPrasad +- fix(logs): Canonicalize paths from the logger integration (#4336) by @colin-sentry +- fix(logs): Use new transport (#4317) by @colin-sentry +- fix: Deprecate `set_measurement()` API. (#3934) by @antonpirker +- fix: Put feature flags on isolation scope (#4363) by @antonpirker +- fix: Make use of `SPANDATA` consistent (#4373) by @antonpirker +- fix: Discord link (#4371) by @sentrivana +- tests: Pin snowballstemmer for now (#4372) by @sentrivana - tests: Regular tox update (#4367) by @sentrivana -- build(deps): bump actions/create-github-app-token from 2.0.2 to 2.0.6 (#4358) by @dependabot -- Put feature flags on isolation scope (#4363) by @antonpirker -- tests: bump test timeout for recursion stacktrace extract to 2s (#4351) by @booxter -- feat(ourlogs): canonicalize paths from the logger integration (#4336) by @colin-sentry -- chore(ourlogs): Use new transport (#4317) by @colin-sentry -- tests: fix test_stacktrace_big_recursion failure due to argv (#4346) by @booxter +- tests: Bump test timeout for recursion stacktrace extract to 2s (#4351) by @booxter +- tests: Fix test_stacktrace_big_recursion failure due to argv (#4346) by @booxter - tests: Move anthropic under toxgen (#4348) by @sentrivana -- Deprecate `set_measurement()` API. (#3934) by @antonpirker - tests: Update tox.ini (#4347) by @sentrivana -- Update GH issue templates for Linear compatibility (#4328) by @stephanie-anderson +- chore: Update GH issue templates for Linear compatibility (#4328) by @stephanie-anderson +- chore: Bump actions/create-github-app-token from 2.0.2 to 2.0.6 (#4358) by @dependabot ## 2.27.0 From 1090eeed8cd18f41d358b0dc0508d55fe614a05b Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 12 May 2025 11:31:07 +0200 Subject: [PATCH 290/591] apidocs: Remove snowballstemmer pin (#4379) Fixed upstream --- requirements-docs.txt | 1 - 1 file changed, 1 deletion(-) diff --git a/requirements-docs.txt b/requirements-docs.txt index a662a0d83f..81e04ba3ef 100644 --- a/requirements-docs.txt +++ b/requirements-docs.txt @@ -3,4 +3,3 @@ shibuya sphinx<8.2 sphinx-autodoc-typehints[type_comments]>=1.8.0 typing-extensions -snowballstemmer<3.0 From eee4cac8f8ae44e5bee9364a482597680b81b52f Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Tue, 13 May 2025 10:23:19 -0400 Subject: [PATCH 291/591] fix(logs): Make `sentry.message.parameters` singular as per spec (#4387) As per https://develop.sentry.dev/sdk/telemetry/logs/#default-attributes we should be using `sentry.message.parameter` (singular) instead of `sentry.message.parameters`. This matches how we approach the other attributes in our conventions. --- sentry_sdk/integrations/logging.py | 2 +- sentry_sdk/logger.py | 2 +- tests/test_logs.py | 23 +++++++++++------------ 3 files changed, 13 insertions(+), 14 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 74baf3d33a..449a05fbf7 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -362,7 +362,7 @@ def _capture_log_from_record(self, client, record): if record.args is not None: if isinstance(record.args, tuple): for i, arg in enumerate(record.args): - attrs[f"sentry.message.parameters.{i}"] = ( + attrs[f"sentry.message.parameter.{i}"] = ( arg if isinstance(arg, str) or isinstance(arg, float) diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index 1fa31b786b..704d7d4f9d 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -18,7 +18,7 @@ def _capture_log(severity_text, severity_number, template, **kwargs): if "attributes" in kwargs: attrs.update(kwargs.pop("attributes")) for k, v in kwargs.items(): - attrs[f"sentry.message.parameters.{k}"] = v + attrs[f"sentry.message.parameter.{k}"] = v attrs = { k: ( diff --git a/tests/test_logs.py b/tests/test_logs.py index 1f6b07e762..ec8ce917c2 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -186,7 +186,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): assert logs[0]["attributes"][k] == v assert logs[0]["attributes"]["sentry.environment"] == "production" assert "sentry.release" in logs[0]["attributes"] - assert logs[0]["attributes"]["sentry.message.parameters.my_var"] == "some value" + assert logs[0]["attributes"]["sentry.message.parameter.my_var"] == "some value" assert logs[0]["attributes"][SPANDATA.SERVER_ADDRESS] == "test-server" assert logs[0]["attributes"]["sentry.sdk.name"].startswith("sentry.python") assert logs[0]["attributes"]["sentry.sdk.version"] == VERSION @@ -214,23 +214,23 @@ def test_logs_message_params(sentry_init, capture_envelopes): logs = envelopes_to_logs(envelopes) assert logs[0]["body"] == "The recorded value was '1'" - assert logs[0]["attributes"]["sentry.message.parameters.int_var"] == 1 + assert logs[0]["attributes"]["sentry.message.parameter.int_var"] == 1 assert logs[1]["body"] == "The recorded value was '2.0'" - assert logs[1]["attributes"]["sentry.message.parameters.float_var"] == 2.0 + assert logs[1]["attributes"]["sentry.message.parameter.float_var"] == 2.0 assert logs[2]["body"] == "The recorded value was 'False'" - assert logs[2]["attributes"]["sentry.message.parameters.bool_var"] is False + assert logs[2]["attributes"]["sentry.message.parameter.bool_var"] is False assert logs[3]["body"] == "The recorded value was 'some string value'" assert ( - logs[3]["attributes"]["sentry.message.parameters.string_var"] + logs[3]["attributes"]["sentry.message.parameter.string_var"] == "some string value" ) assert logs[4]["body"] == "The recorded error was 'some error'" assert ( - logs[4]["attributes"]["sentry.message.parameters.error"] + logs[4]["attributes"]["sentry.message.parameter.error"] == "Exception('some error')" ) @@ -287,8 +287,8 @@ def test_logger_integration_warning(sentry_init, capture_envelopes): assert "code.line.number" in attrs assert attrs["logger.name"] == "test-logger" assert attrs["sentry.environment"] == "production" - assert attrs["sentry.message.parameters.0"] == "1" - assert attrs["sentry.message.parameters.1"] == "2" + assert attrs["sentry.message.parameter.0"] == "1" + assert attrs["sentry.message.parameter.1"] == "2" assert attrs["sentry.origin"] == "auto.logger.log" assert logs[0]["severity_number"] == 13 assert logs[0]["severity_text"] == "warn" @@ -349,14 +349,13 @@ def test_logging_errors(sentry_init, capture_envelopes): logs = envelopes_to_logs(envelopes) assert logs[0]["severity_text"] == "error" assert "sentry.message.template" not in logs[0]["attributes"] - assert "sentry.message.parameters.0" not in logs[0]["attributes"] + assert "sentry.message.parameter.0" not in logs[0]["attributes"] assert "code.line.number" in logs[0]["attributes"] assert logs[1]["severity_text"] == "error" assert logs[1]["attributes"]["sentry.message.template"] == "error is %s" assert ( - logs[1]["attributes"]["sentry.message.parameters.0"] - == "Exception('test exc 2')" + logs[1]["attributes"]["sentry.message.parameter.0"] == "Exception('test exc 2')" ) assert "code.line.number" in logs[1]["attributes"] @@ -456,7 +455,7 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): "logger.name": "test-logger", "sentry.origin": "auto.logger.log", "sentry.message.template": "log #%d", - "sentry.message.parameters.0": 1, + "sentry.message.parameter.0": 1, "sentry.environment": "production", "sentry.sdk.name": "sentry.python", "sentry.sdk.version": VERSION, From c639f1c083eefb6b4cc67164d9d8dfa9ac568b98 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 15 May 2025 11:40:28 +0200 Subject: [PATCH 292/591] ci: Fix pyspark test suite (#4382) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - pyspark was failing to bind to an address - aws_lambda fixed itself 🪄 --- tests/integrations/spark/test_spark.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/tests/integrations/spark/test_spark.py b/tests/integrations/spark/test_spark.py index 7eeab15dc4..91882a0b8f 100644 --- a/tests/integrations/spark/test_spark.py +++ b/tests/integrations/spark/test_spark.py @@ -10,7 +10,7 @@ ) from sentry_sdk.integrations.spark.spark_worker import SparkWorkerIntegration -from pyspark import SparkContext +from pyspark import SparkConf, SparkContext from py4j.protocol import Py4JJavaError @@ -25,12 +25,13 @@ def sentry_init_with_reset(sentry_init): from sentry_sdk.integrations import _processed_integrations yield lambda: sentry_init(integrations=[SparkIntegration()]) - _processed_integrations.remove("spark") + _processed_integrations.discard("spark") @pytest.fixture(scope="function") def create_spark_context(): - yield lambda: SparkContext(appName="Testing123") + conf = SparkConf().set("spark.driver.bindAddress", "127.0.0.1") + yield lambda: SparkContext(conf=conf, appName="Testing123") SparkContext._active_spark_context.stop() From aa0eaab19a9867ea5b80f8aab3ddc6c2e9293835 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 15 May 2025 11:40:48 +0200 Subject: [PATCH 293/591] fix(typing): Add before_send_log to Experiments (#4383) --- sentry_sdk/consts.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 2ceec2738b..30461b4524 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1,5 +1,4 @@ import itertools - from enum import Enum from typing import TYPE_CHECKING @@ -47,6 +46,7 @@ class CompressionAlgo(Enum): Event, EventProcessor, Hint, + Log, MeasurementUnit, ProfilerMode, TracesSampler, @@ -79,6 +79,7 @@ class CompressionAlgo(Enum): ], "metric_code_locations": Optional[bool], "enable_logs": Optional[bool], + "before_send_log": Optional[Callable[[Log, Hint], Optional[Log]]], }, total=False, ) From 2f97cdaf53eee2737b6dc4aab1d99e91d33b0b76 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 19 May 2025 10:53:31 +0200 Subject: [PATCH 294/591] fix(celery): Do not send extra check-in (#4395) When using the RedBeatScheduler, we're sending an extra in-progress check-in at scheduler start. Since this is never followed by an ok or error check-in, the check-in is marked as timed out in Sentry. We're patching the scheduler's `maybe_due`, which (as the name implies) [might not end up executing the task](https://github.com/sibson/redbeat/blob/bdefad23b47f75e2e85d45f46f9f16dd00a93d40/redbeat/schedulers.py#L506-L508). This is indeed what seems to be happening -- `maybe_due` is run when the scheduler starts, but without scheduling the task. We don't check whether `maybe_due` actually ended up scheduling anything and always fire an in-progress check-in. Patching the [scheduler's `apply_async`](https://github.com/sibson/redbeat/blob/bdefad23b47f75e2e85d45f46f9f16dd00a93d40/redbeat/schedulers.py#L511) instead. Closes https://github.com/getsentry/sentry-python/issues/4392 --- sentry_sdk/integrations/celery/__init__.py | 4 ++-- sentry_sdk/integrations/celery/beat.py | 4 ++-- .../celery/test_celery_beat_crons.py | 18 ++++++++++-------- 3 files changed, 14 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index e8811d767e..d8d89217ca 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -9,7 +9,7 @@ from sentry_sdk.integrations import _check_minimum_version, Integration, DidNotEnable from sentry_sdk.integrations.celery.beat import ( _patch_beat_apply_entry, - _patch_redbeat_maybe_due, + _patch_redbeat_apply_async, _setup_celery_beat_signals, ) from sentry_sdk.integrations.celery.utils import _now_seconds_since_epoch @@ -73,7 +73,7 @@ def __init__( self.exclude_beat_tasks = exclude_beat_tasks _patch_beat_apply_entry() - _patch_redbeat_maybe_due() + _patch_redbeat_apply_async() _setup_celery_beat_signals(monitor_beat_tasks) @staticmethod diff --git a/sentry_sdk/integrations/celery/beat.py b/sentry_sdk/integrations/celery/beat.py index ddbc8561a4..4b7e45e6f0 100644 --- a/sentry_sdk/integrations/celery/beat.py +++ b/sentry_sdk/integrations/celery/beat.py @@ -202,12 +202,12 @@ def _patch_beat_apply_entry(): Scheduler.apply_entry = _wrap_beat_scheduler(Scheduler.apply_entry) -def _patch_redbeat_maybe_due(): +def _patch_redbeat_apply_async(): # type: () -> None if RedBeatScheduler is None: return - RedBeatScheduler.maybe_due = _wrap_beat_scheduler(RedBeatScheduler.maybe_due) + RedBeatScheduler.apply_async = _wrap_beat_scheduler(RedBeatScheduler.apply_async) def _setup_celery_beat_signals(monitor_beat_tasks): diff --git a/tests/integrations/celery/test_celery_beat_crons.py b/tests/integrations/celery/test_celery_beat_crons.py index 58c4c6208d..17b4a5e73d 100644 --- a/tests/integrations/celery/test_celery_beat_crons.py +++ b/tests/integrations/celery/test_celery_beat_crons.py @@ -10,7 +10,7 @@ _get_headers, _get_monitor_config, _patch_beat_apply_entry, - _patch_redbeat_maybe_due, + _patch_redbeat_apply_async, crons_task_failure, crons_task_retry, crons_task_success, @@ -454,10 +454,10 @@ def test_exclude_redbeat_tasks_option( """ Test excluding Celery RedBeat tasks from automatic instrumentation. """ - fake_maybe_due = MagicMock() + fake_apply_async = MagicMock() fake_redbeat_scheduler = MagicMock() - fake_redbeat_scheduler.maybe_due = fake_maybe_due + fake_redbeat_scheduler.apply_async = fake_apply_async fake_integration = MagicMock() fake_integration.exclude_beat_tasks = exclude_beat_tasks @@ -481,17 +481,19 @@ def test_exclude_redbeat_tasks_option( "sentry_sdk.integrations.celery.beat._get_monitor_config", fake_get_monitor_config, ) as _get_monitor_config: - # Mimic CeleryIntegration patching of RedBeatScheduler.maybe_due() - _patch_redbeat_maybe_due() + # Mimic CeleryIntegration patching of RedBeatScheduler.apply_async() + _patch_redbeat_apply_async() # Mimic Celery RedBeat calling a task from the RedBeat schedule - RedBeatScheduler.maybe_due(fake_redbeat_scheduler, fake_schedule_entry) + RedBeatScheduler.apply_async( + fake_redbeat_scheduler, fake_schedule_entry + ) if task_in_excluded_beat_tasks: # Only the original RedBeatScheduler.maybe_due() is called, _get_monitor_config is NOT called. - assert fake_maybe_due.call_count == 1 + assert fake_apply_async.call_count == 1 _get_monitor_config.assert_not_called() else: # The original RedBeatScheduler.maybe_due() is called, AND _get_monitor_config is called. - assert fake_maybe_due.call_count == 1 + assert fake_apply_async.call_count == 1 assert _get_monitor_config.call_count == 1 From f5727572281578a73b21cd757288b820932eded6 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 19 May 2025 08:42:22 -0400 Subject: [PATCH 295/591] feat: Allow configuring `keep_alive` via environment variable (#4366) This commit enables the `keep_alive` option to be set via the `SENTRY_KEEP_ALIVE` environment variable. When both the environment variable and the argument are provided, the argument takes precedence. Closes #4354 --- sentry_sdk/client.py | 5 ++++ sentry_sdk/consts.py | 2 +- tests/test_client.py | 64 ++++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 70 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index f06166bcc8..9747c792d9 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -142,6 +142,11 @@ def _get_options(*args, **kwargs): ) rv["socket_options"] = None + if rv["keep_alive"] is None: + rv["keep_alive"] = ( + env_to_bool(os.environ.get("SENTRY_KEEP_ALIVE"), strict=True) or False + ) + if rv["enable_tracing"] is not None: warnings.warn( "The `enable_tracing` parameter is deprecated. Please use `traces_sample_rate` instead.", diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 30461b4524..ed81a512ba 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -578,7 +578,7 @@ def __init__( ignore_errors=[], # type: Sequence[Union[type, str]] # noqa: B006 max_request_body_size="medium", # type: str socket_options=None, # type: Optional[List[Tuple[int, int, int | bytes]]] - keep_alive=False, # type: bool + keep_alive=None, # type: Optional[bool] before_send=None, # type: Optional[EventProcessor] before_breadcrumb=None, # type: Optional[BreadcrumbProcessor] debug=None, # type: Optional[bool] diff --git a/tests/test_client.py b/tests/test_client.py index 67f53d989a..2986920452 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -1,3 +1,4 @@ +import contextlib import os import json import subprocess @@ -1496,3 +1497,66 @@ def test_dropped_transaction(sentry_init, capture_record_lost_event_calls, test_ def test_enable_tracing_deprecated(sentry_init, enable_tracing): with pytest.warns(DeprecationWarning): sentry_init(enable_tracing=enable_tracing) + + +def make_options_transport_cls(): + """Make an options transport class that captures the options passed to it.""" + # We need a unique class for each test so that the options are not + # shared between tests. + + class OptionsTransport(Transport): + """Transport that captures the options passed to it.""" + + def __init__(self, options): + super().__init__(options) + type(self).options = options + + def capture_envelope(self, _): + pass + + return OptionsTransport + + +@contextlib.contextmanager +def clear_env_var(name): + """Helper to clear the a given environment variable, + and restore it to its original value on exit.""" + old_value = os.environ.pop(name, None) + + try: + yield + finally: + if old_value is not None: + os.environ[name] = old_value + elif name in os.environ: + del os.environ[name] + + +@pytest.mark.parametrize( + ("env_value", "arg_value", "expected_value"), + [ + (None, None, False), # default + ("0", None, False), # env var false + ("1", None, True), # env var true + (None, False, False), # arg false + (None, True, True), # arg true + # Argument overrides environment variable + ("0", True, True), # env false, arg true + ("1", False, False), # env true, arg false + ], +) +def test_keep_alive(env_value, arg_value, expected_value): + transport_cls = make_options_transport_cls() + keep_alive_kwarg = {} if arg_value is None else {"keep_alive": arg_value} + + with clear_env_var("SENTRY_KEEP_ALIVE"): + if env_value is not None: + os.environ["SENTRY_KEEP_ALIVE"] = env_value + + sentry_sdk.init( + dsn="http://foo@sentry.io/123", + transport=transport_cls, + **keep_alive_kwarg, + ) + + assert transport_cls.options["keep_alive"] is expected_value From 830f270791270bd31462acd879ee7a0e8a2926c3 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 19 May 2025 15:05:16 +0200 Subject: [PATCH 296/591] fix(loguru): Move integration setup from `__init__` to `setup_once` (#4399) We shouldn't have setup code in an integration's `__init__`. This can be called an arbitrary amount of times. `setup_once` exists so that the code is only guaranteed to run once. Having integration setup in `__init__` also means that the integration currently can't be disabled via `disabled_integrations`. Fixes https://github.com/getsentry/sentry-python/issues/4398 --- sentry_sdk/integrations/loguru.py | 55 ++++++++++-------------- tests/integrations/loguru/test_loguru.py | 17 ++++++-- 2 files changed, 37 insertions(+), 35 deletions(-) diff --git a/sentry_sdk/integrations/loguru.py b/sentry_sdk/integrations/loguru.py index 5b76ea812a..a71c4ac87f 100644 --- a/sentry_sdk/integrations/loguru.py +++ b/sentry_sdk/integrations/loguru.py @@ -11,7 +11,7 @@ if TYPE_CHECKING: from logging import LogRecord - from typing import Optional, Tuple, Any + from typing import Optional, Any try: import loguru @@ -43,16 +43,16 @@ class LoggingLevels(enum.IntEnum): DEFAULT_LEVEL = LoggingLevels.INFO.value DEFAULT_EVENT_LEVEL = LoggingLevels.ERROR.value -# We need to save the handlers to be able to remove them later -# in tests (they call `LoguruIntegration.__init__` multiple times, -# and we can't use `setup_once` because it's called before -# than we get configuration). -_ADDED_HANDLERS = (None, None) # type: Tuple[Optional[int], Optional[int]] class LoguruIntegration(Integration): identifier = "loguru" + level = DEFAULT_LEVEL # type: Optional[int] + event_level = DEFAULT_EVENT_LEVEL # type: Optional[int] + breadcrumb_format = DEFAULT_FORMAT + event_format = DEFAULT_FORMAT + def __init__( self, level=DEFAULT_LEVEL, @@ -61,36 +61,27 @@ def __init__( event_format=DEFAULT_FORMAT, ): # type: (Optional[int], Optional[int], str | loguru.FormatFunction, str | loguru.FormatFunction) -> None - global _ADDED_HANDLERS - breadcrumb_handler, event_handler = _ADDED_HANDLERS - - if breadcrumb_handler is not None: - logger.remove(breadcrumb_handler) - breadcrumb_handler = None - if event_handler is not None: - logger.remove(event_handler) - event_handler = None - - if level is not None: - breadcrumb_handler = logger.add( - LoguruBreadcrumbHandler(level=level), - level=level, - format=breadcrumb_format, - ) - - if event_level is not None: - event_handler = logger.add( - LoguruEventHandler(level=event_level), - level=event_level, - format=event_format, - ) - - _ADDED_HANDLERS = (breadcrumb_handler, event_handler) + LoguruIntegration.level = level + LoguruIntegration.event_level = event_level + LoguruIntegration.breadcrumb_format = breadcrumb_format + LoguruIntegration.event_format = event_format @staticmethod def setup_once(): # type: () -> None - pass # we do everything in __init__ + if LoguruIntegration.level is not None: + logger.add( + LoguruBreadcrumbHandler(level=LoguruIntegration.level), + level=LoguruIntegration.level, + format=LoguruIntegration.breadcrumb_format, + ) + + if LoguruIntegration.event_level is not None: + logger.add( + LoguruEventHandler(level=LoguruIntegration.event_level), + level=LoguruIntegration.event_level, + format=LoguruIntegration.event_format, + ) class _LoguruBaseHandler(_BaseHandler): diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 64e9f22ba5..6be09b86dc 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -32,7 +32,12 @@ def test_just_log( expected_sentry_level, disable_breadcrumbs, disable_events, + uninstall_integration, + request, ): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + sentry_init( integrations=[ LoguruIntegration( @@ -49,7 +54,7 @@ def test_just_log( formatted_message = ( " | " + "{:9}".format(level.name.upper()) - + "| tests.integrations.loguru.test_loguru:test_just_log:47 - test" + + "| tests.integrations.loguru.test_loguru:test_just_log:52 - test" ) if not created_event: @@ -78,7 +83,10 @@ def test_just_log( assert event["logentry"]["message"][23:] == formatted_message -def test_breadcrumb_format(sentry_init, capture_events): +def test_breadcrumb_format(sentry_init, capture_events, uninstall_integration, request): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + sentry_init( integrations=[ LoguruIntegration( @@ -98,7 +106,10 @@ def test_breadcrumb_format(sentry_init, capture_events): assert breadcrumb["message"] == formatted_message -def test_event_format(sentry_init, capture_events): +def test_event_format(sentry_init, capture_events, uninstall_integration, request): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + sentry_init( integrations=[ LoguruIntegration( From 0c9333ed8bc2c4da15106fd37b4fa9fd3e9aed93 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 19 May 2025 13:30:09 +0000 Subject: [PATCH 297/591] release: 2.29.0 --- CHANGELOG.md | 12 ++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 15 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 81f749d83a..35ee9056dd 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,17 @@ # Changelog +## 2.29.0 + +### Various fixes & improvements + +- fix(loguru): Move integration setup from `__init__` to `setup_once` (#4399) by @sentrivana +- feat: Allow configuring `keep_alive` via environment variable (#4366) by @szokeasaurusrex +- fix(celery): Do not send extra check-in (#4395) by @sentrivana +- fix(typing): Add before_send_log to Experiments (#4383) by @sentrivana +- ci: Fix pyspark test suite (#4382) by @sentrivana +- fix(logs): Make `sentry.message.parameters` singular as per spec (#4387) by @AbhiPrasad +- apidocs: Remove snowballstemmer pin (#4379) by @sentrivana + ## 2.28.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 34c88ae6cd..7735eab04f 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.28.0" +release = "2.29.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index ed81a512ba..a5a6343b0b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1026,4 +1026,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.28.0" +VERSION = "2.29.0" diff --git a/setup.py b/setup.py index 8fd1ae6293..543882dcd6 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.28.0", + version="2.29.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 3ebd122604835bdf5cf440df1d2794490ce8b5f0 Mon Sep 17 00:00:00 2001 From: Lorenzo Cian Date: Mon, 19 May 2025 16:10:31 +0200 Subject: [PATCH 298/591] fix(logs): send `severity_text`: `warn` instead of `warning` (#4396) We need to send `warn` as the `severity_text` for warning level logs, not `warning`. If we send `warning` it will not be recognized by the backend and the severity will show up as `UNKNOWN` in the frontend. See screenshot for comparison of after and before. ![Screenshot 2025-05-16 at 18 24 50](https://github.com/user-attachments/assets/240e3c6d-0f68-49e8-b79b-a67ac0e8a5d4) --- sentry_sdk/logger.py | 2 +- tests/test_logs.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index 704d7d4f9d..c675c4d95d 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -51,6 +51,6 @@ def _capture_log(severity_text, severity_number, template, **kwargs): trace = functools.partial(_capture_log, "trace", 1) debug = functools.partial(_capture_log, "debug", 5) info = functools.partial(_capture_log, "info", 9) -warning = functools.partial(_capture_log, "warning", 13) +warning = functools.partial(_capture_log, "warn", 13) error = functools.partial(_capture_log, "error", 17) fatal = functools.partial(_capture_log, "fatal", 21) diff --git a/tests/test_logs.py b/tests/test_logs.py index ec8ce917c2..7b886b7a36 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -102,7 +102,7 @@ def test_logs_basics(sentry_init, capture_envelopes): assert logs[2].get("severity_text") == "info" assert logs[2].get("severity_number") == 9 - assert logs[3].get("severity_text") == "warning" + assert logs[3].get("severity_text") == "warn" assert logs[3].get("severity_number") == 13 assert logs[4].get("severity_text") == "error" @@ -155,7 +155,7 @@ def _before_log(record, hint): assert logs[0]["severity_text"] == "trace" assert logs[1]["severity_text"] == "debug" assert logs[2]["severity_text"] == "info" - assert logs[3]["severity_text"] == "warning" + assert logs[3]["severity_text"] == "warn" assert before_log_called[0] From 7973ac095c0ecedba4514f11042e1536bb1234ee Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 19 May 2025 14:20:49 +0000 Subject: [PATCH 299/591] release: 2.29.1 --- CHANGELOG.md | 6 ++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 9 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 35ee9056dd..8d73e1bdd0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,11 @@ # Changelog +## 2.29.1 + +### Various fixes & improvements + +- fix(logs): send `severity_text`: `warn` instead of `warning` (#4396) by @lcian + ## 2.29.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 7735eab04f..e639185528 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.29.0" +release = "2.29.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index a5a6343b0b..241d197699 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1026,4 +1026,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.29.0" +VERSION = "2.29.1" diff --git a/setup.py b/setup.py index 543882dcd6..92634bd2ef 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.29.0", + version="2.29.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 83e0386b4dfc3a40d0301fa90f5750dade4bb024 Mon Sep 17 00:00:00 2001 From: Roman Inflianskas Date: Tue, 20 May 2025 10:20:06 +0300 Subject: [PATCH 300/591] tests(logs): avoid failures when running with integrations enabled (#4388) When (at least) one of integrations is enabled (because some dependencies are installed in the environment), `sentry.sdk.name` is changed from `sentry.python` to `sentry.python.[FIRST_ENABLED_INTEGRATION]` which makes `test_logs_attributes` fail. Prevent failure by relaxing the check. This change is beneficial not only for packaging (this patch was required for packaging for Fedora), but also for running tests with `pytest` directly. See also: https://github.com/getsentry/sentry-python/pull/4316 --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- tests/test_logs.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/tests/test_logs.py b/tests/test_logs.py index 7b886b7a36..7e8a72d30a 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -447,6 +447,8 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): assert isinstance(attributes["thread.name"], str) del attributes["thread.name"] + assert attributes.pop("sentry.sdk.name").startswith("sentry.python") + # Assert on the remaining non-dynamic attributes. assert attributes == { "foo": "bar", @@ -457,7 +459,6 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): "sentry.message.template": "log #%d", "sentry.message.parameter.0": 1, "sentry.environment": "production", - "sentry.sdk.name": "sentry.python", "sentry.sdk.version": VERSION, "sentry.severity_number": 13, "sentry.severity_text": "warn", From 94414cdcdce2a311355967ee7b43e7035cf954ce Mon Sep 17 00:00:00 2001 From: Marcelo Galigniana Date: Tue, 20 May 2025 04:21:49 -0300 Subject: [PATCH 301/591] Increase test coverage (#4393) Fixes GH-3515 Increase test coverage for `attachments.py` and `utils.py`. --------- Co-authored-by: Anton Pirker --- tests/test_basics.py | 60 ++++++++++ tests/test_utils.py | 188 +++++++++++++++++++++++++------- tests/utils/test_contextvars.py | 15 ++- tests/utils/test_general.py | 39 +++++++ 4 files changed, 258 insertions(+), 44 deletions(-) diff --git a/tests/test_basics.py b/tests/test_basics.py index 0fdf9f811f..2eeba78216 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -515,6 +515,66 @@ def test_attachments_graceful_failure( assert envelope.items[1].payload.get_bytes() == b"" +def test_attachments_exceptions(sentry_init): + sentry_init() + + scope = sentry_sdk.get_isolation_scope() + + # bytes and path are None + with pytest.raises(TypeError) as e: + scope.add_attachment() + + assert str(e.value) == "path or raw bytes required for attachment" + + # filename is None + with pytest.raises(TypeError) as e: + scope.add_attachment(bytes=b"Hello World!") + + assert str(e.value) == "filename is required for attachment" + + +def test_attachments_content_type_is_none(sentry_init, capture_envelopes): + sentry_init() + envelopes = capture_envelopes() + + scope = sentry_sdk.get_isolation_scope() + + scope.add_attachment( + bytes=b"Hello World!", filename="message.txt", content_type="foo/bar" + ) + capture_exception(ValueError()) + + (envelope,) = envelopes + attachments = [x for x in envelope.items if x.type == "attachment"] + (message,) = attachments + + assert message.headers["filename"] == "message.txt" + assert message.headers["content_type"] == "foo/bar" + + +def test_attachments_repr(sentry_init): + sentry_init() + + scope = sentry_sdk.get_isolation_scope() + + scope.add_attachment(bytes=b"Hello World!", filename="message.txt") + + assert repr(scope._attachments[0]) == "" + + +def test_attachments_bytes_callable_payload(sentry_init): + sentry_init() + + scope = sentry_sdk.get_isolation_scope() + + scope.add_attachment(bytes=bytes, filename="message.txt") + + attachment = scope._attachments[0] + item = attachment.to_envelope_item() + + assert item.payload.bytes == b"" + + def test_integration_scoping(sentry_init, capture_events): logger = logging.getLogger("test_basics") diff --git a/tests/test_utils.py b/tests/test_utils.py index b731c3e3ab..efa2e7c068 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -32,6 +32,10 @@ _get_installed_modules, _generate_installed_modules, ensure_integration_enabled, + to_string, + exc_info_from_error, + get_lines_from_file, + package_version, ) @@ -61,55 +65,72 @@ def _normalize_distribution_name(name): return re.sub(r"[-_.]+", "-", name).lower() -@pytest.mark.parametrize( - ("input_str", "expected_output"), +isoformat_inputs_and_datetime_outputs = ( ( - ( - "2021-01-01T00:00:00.000000Z", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), # UTC time - ( - "2021-01-01T00:00:00.000000", - datetime(2021, 1, 1).astimezone(timezone.utc), - ), # No TZ -- assume local but convert to UTC - ( - "2021-01-01T00:00:00Z", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), # UTC - No milliseconds - ( - "2021-01-01T00:00:00.000000+00:00", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), - ( - "2021-01-01T00:00:00.000000-00:00", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), - ( - "2021-01-01T00:00:00.000000+0000", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), - ( - "2021-01-01T00:00:00.000000-0000", - datetime(2021, 1, 1, tzinfo=timezone.utc), - ), - ( - "2020-12-31T00:00:00.000000+02:00", - datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=2))), - ), # UTC+2 time - ( - "2020-12-31T00:00:00.000000-0200", - datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=-2))), - ), # UTC-2 time - ( - "2020-12-31T00:00:00-0200", - datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=-2))), - ), # UTC-2 time - no milliseconds + "2021-01-01T00:00:00.000000Z", + datetime(2021, 1, 1, tzinfo=timezone.utc), + ), # UTC time + ( + "2021-01-01T00:00:00.000000", + datetime(2021, 1, 1).astimezone(timezone.utc), + ), # No TZ -- assume local but convert to UTC + ( + "2021-01-01T00:00:00Z", + datetime(2021, 1, 1, tzinfo=timezone.utc), + ), # UTC - No milliseconds + ( + "2021-01-01T00:00:00.000000+00:00", + datetime(2021, 1, 1, tzinfo=timezone.utc), + ), + ( + "2021-01-01T00:00:00.000000-00:00", + datetime(2021, 1, 1, tzinfo=timezone.utc), + ), + ( + "2021-01-01T00:00:00.000000+0000", + datetime(2021, 1, 1, tzinfo=timezone.utc), + ), + ( + "2021-01-01T00:00:00.000000-0000", + datetime(2021, 1, 1, tzinfo=timezone.utc), ), + ( + "2020-12-31T00:00:00.000000+02:00", + datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=2))), + ), # UTC+2 time + ( + "2020-12-31T00:00:00.000000-0200", + datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=-2))), + ), # UTC-2 time + ( + "2020-12-31T00:00:00-0200", + datetime(2020, 12, 31, tzinfo=timezone(timedelta(hours=-2))), + ), # UTC-2 time - no milliseconds +) + + +@pytest.mark.parametrize( + ("input_str", "expected_output"), + isoformat_inputs_and_datetime_outputs, ) def test_datetime_from_isoformat(input_str, expected_output): assert datetime_from_isoformat(input_str) == expected_output, input_str +@pytest.mark.parametrize( + ("input_str", "expected_output"), + isoformat_inputs_and_datetime_outputs, +) +def test_datetime_from_isoformat_with_py_36_or_lower(input_str, expected_output): + """ + `fromisoformat` was added in Python version 3.7 + """ + with mock.patch("sentry_sdk.utils.datetime") as datetime_mocked: + datetime_mocked.fromisoformat.side_effect = AttributeError() + datetime_mocked.strptime = datetime.strptime + assert datetime_from_isoformat(input_str) == expected_output, input_str + + @pytest.mark.parametrize( "env_var_value,strict,expected", [ @@ -346,6 +367,12 @@ def test_sanitize_url_and_split(url, expected_result): assert sanitized_url.fragment == expected_result.fragment +def test_sanitize_url_remove_authority_is_false(): + url = "https://usr:pwd@example.com/" + sanitized_url = sanitize_url(https://github.com/url,%20remove_authority=False) + assert sanitized_url == url + + @pytest.mark.parametrize( ("url", "sanitize", "expected_url", "expected_query", "expected_fragment"), [ @@ -629,6 +656,17 @@ def test_get_error_message(error, expected_result): assert get_error_message(exc_value) == expected_result(exc_value) +def test_safe_str_fails(): + class ExplodingStr: + def __str__(self): + raise Exception + + obj = ExplodingStr() + result = safe_str(obj) + + assert result == repr(obj) + + def test_installed_modules(): try: from importlib.metadata import distributions, version @@ -712,6 +750,20 @@ def test_default_release_empty_string(): assert release is None +def test_get_default_release_sentry_release_env(monkeypatch): + monkeypatch.setenv("SENTRY_RELEASE", "sentry-env-release") + assert get_default_release() == "sentry-env-release" + + +def test_get_default_release_other_release_env(monkeypatch): + monkeypatch.setenv("SOURCE_VERSION", "other-env-release") + + with mock.patch("sentry_sdk.utils.get_git_revision", return_value=""): + release = get_default_release() + + assert release == "other-env-release" + + def test_ensure_integration_enabled_integration_enabled(sentry_init): def original_function(): return "original" @@ -973,3 +1025,55 @@ def test_function(): ... sentry_sdk.utils.qualname_from_function(test_function) == "test_qualname_from_function_none_name..test_function" ) + + +def test_to_string_unicode_decode_error(): + class BadStr: + def __str__(self): + raise UnicodeDecodeError("utf-8", b"", 0, 1, "reason") + + obj = BadStr() + result = to_string(obj) + assert result == repr(obj)[1:-1] + + +def test_exc_info_from_error_dont_get_an_exc(): + class NotAnException: + pass + + with pytest.raises(ValueError) as exc: + exc_info_from_error(NotAnException()) + + assert "Expected Exception object to report, got Date: Tue, 20 May 2025 09:41:16 +0000 Subject: [PATCH 302/591] build(deps): bump codecov/codecov-action from 5.4.2 to 5.4.3 (#4397) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.4.2 to 5.4.3.
Release notes

Sourced from codecov/codecov-action's releases.

v5.4.3

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.2...v5.4.3

Changelog

Sourced from codecov/codecov-action's changelog.

v5.4.3

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.2..v5.4.3

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=codecov/codecov-action&package-manager=github_actions&previous-version=5.4.2&new-version=5.4.3)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 13 files changed, 19 insertions(+), 19 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index bc89cb9afe..4aa0f36b77 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -83,7 +83,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -158,7 +158,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 7763aa509d..114e904d4b 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 864583532d..2ac8d827fa 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 815b550027..460ffe1ad5 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -107,7 +107,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -206,7 +206,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index e28067841b..0e2c9ef166 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 41a77ffe34..3e0903e2c5 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index b741302de6..51ae8a8a81 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 7da9929435..05a8aaeda1 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 43b5e4a6a5..769a95c08a 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -75,7 +75,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index a6850256b2..43b5d8d56d 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -186,7 +186,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index b40027ddc7..67669c729b 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 1fbff47b65..c0438dc924 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 901e4808e4..705ae69a7f 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -91,7 +91,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.4.2 + uses: codecov/codecov-action@v5.4.3 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From b2d44006e4ab6293f48d1f1aaded74b61d79b0fc Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 20 May 2025 07:36:47 -0400 Subject: [PATCH 303/591] fix: Handle invalid `SENTRY_DEBUG` values properly (#4400) Previously, if the environment variable `SENTRY_DEBUG` would have an invalid value, such as "invalid," we would set `rv["debug"]` to `None`. However, since this value should be a boolean, it should instead be set to `False`. This change ensures that `rv["debug"]` always is a boolean. Where we previously would have set the value to `None`, we now set it to `False`. Other behavior remains unchanged. Ref https://github.com/getsentry/sentry-python/pull/4366/files#r2075294825 --- sentry_sdk/client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 9747c792d9..a03598a981 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -110,7 +110,7 @@ def _get_options(*args, **kwargs): rv["environment"] = os.environ.get("SENTRY_ENVIRONMENT") or "production" if rv["debug"] is None: - rv["debug"] = env_to_bool(os.environ.get("SENTRY_DEBUG", "False"), strict=True) + rv["debug"] = env_to_bool(os.environ.get("SENTRY_DEBUG"), strict=True) or False if rv["server_name"] is None and hasattr(socket, "gethostname"): rv["server_name"] = socket.gethostname() From ac7646b8334e4d083fbda6d491e81247324b2dee Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 20 May 2025 14:05:46 +0200 Subject: [PATCH 304/591] tests: Regenerate toxgen (#4403) Regular tox update --- tox.ini | 81 +++++++++++++++++++++++++++------------------------------ 1 file changed, 39 insertions(+), 42 deletions(-) diff --git a/tox.ini b/tox.ini index 332f541793..d9632b7cb5 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-05-06T10:23:50.156629+00:00 +# Last generated: 2025-05-20T07:21:06.821358+00:00 [tox] requires = @@ -136,9 +136,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.27.0 - {py3.8,py3.11,py3.12}-anthropic-v0.38.0 - {py3.8,py3.11,py3.12}-anthropic-v0.50.0 + {py3.8,py3.11,py3.12}-anthropic-v0.28.1 + {py3.8,py3.11,py3.12}-anthropic-v0.40.0 + {py3.8,py3.11,py3.12}-anthropic-v0.51.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 @@ -148,7 +148,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.31.4 # ~~~ DBs ~~~ @@ -157,7 +157,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.12.1 + {py3.9,py3.12,py3.13}-pymongo-v4.13.0 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -165,7 +165,7 @@ envlist = {py3.6,py3.8,py3.9}-sqlalchemy-v1.3.24 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 - {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.40 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.41 # ~~~ Flags ~~~ @@ -200,9 +200,9 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.228.0 - {py3.8,py3.12,py3.13}-strawberry-v0.247.2 - {py3.9,py3.12,py3.13}-strawberry-v0.266.0 + {py3.8,py3.11,py3.12}-strawberry-v0.229.2 + {py3.8,py3.12,py3.13}-strawberry-v0.249.0 + {py3.9,py3.12,py3.13}-strawberry-v0.269.0 # ~~~ Network ~~~ @@ -238,14 +238,14 @@ envlist = {py3.6,py3.7}-django-v1.11.29 {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 - {py3.8,py3.11,py3.12}-django-v4.2.20 + {py3.8,py3.11,py3.12}-django-v4.2.21 {py3.10,py3.11,py3.12}-django-v5.0.14 - {py3.10,py3.12,py3.13}-django-v5.2 + {py3.10,py3.12,py3.13}-django-v5.2.1 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 {py3.8,py3.12,py3.13}-flask-v3.0.3 - {py3.9,py3.12,py3.13}-flask-v3.1.0 + {py3.9,py3.12,py3.13}-flask-v3.1.1 {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 @@ -287,10 +287,9 @@ envlist = {py3.8,py3.10,py3.11}-starlite-v1.51.16 {py3.6,py3.7,py3.8}-tornado-v6.0.4 - {py3.6,py3.8,py3.9}-tornado-v6.1 {py3.7,py3.9,py3.10}-tornado-v6.2 {py3.8,py3.10,py3.11}-tornado-v6.4.2 - {py3.9,py3.12,py3.13}-tornado-v6.5b1 + {py3.9,py3.12,py3.13}-tornado-v6.5 # ~~~ Misc ~~~ @@ -301,9 +300,9 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.31 - {py3.9,py3.12,py3.13}-trytond-v7.6.0 + {py3.9,py3.12,py3.13}-trytond-v7.6.1 - {py3.7,py3.12,py3.13}-typer-v0.15.3 + {py3.7,py3.12,py3.13}-typer-v0.15.4 @@ -501,13 +500,13 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.27.0: anthropic==0.27.0 - anthropic-v0.38.0: anthropic==0.38.0 - anthropic-v0.50.0: anthropic==0.50.0 + anthropic-v0.28.1: anthropic==0.28.1 + anthropic-v0.40.0: anthropic==0.40.0 + anthropic-v0.51.0: anthropic==0.51.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.27.0: httpx<0.28.0 - anthropic-v0.38.0: httpx<0.28.0 + anthropic-v0.28.1: httpx<0.28.0 + anthropic-v0.40.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.8.1: cohere==5.8.1 @@ -517,7 +516,7 @@ deps = huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.25.2: huggingface_hub==0.25.2 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.30.2: huggingface_hub==0.30.2 + huggingface_hub-v0.31.4: huggingface_hub==0.31.4 # ~~~ DBs ~~~ @@ -526,7 +525,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.12.1: pymongo==4.12.1 + pymongo-v4.13.0: pymongo==4.13.0 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -535,7 +534,7 @@ deps = sqlalchemy-v1.3.24: sqlalchemy==1.3.24 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 - sqlalchemy-v2.0.40: sqlalchemy==2.0.40 + sqlalchemy-v2.0.41: sqlalchemy==2.0.41 # ~~~ Flags ~~~ @@ -579,13 +578,13 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.228.0: strawberry-graphql[fastapi,flask]==0.228.0 - strawberry-v0.247.2: strawberry-graphql[fastapi,flask]==0.247.2 - strawberry-v0.266.0: strawberry-graphql[fastapi,flask]==0.266.0 + strawberry-v0.229.2: strawberry-graphql[fastapi,flask]==0.229.2 + strawberry-v0.249.0: strawberry-graphql[fastapi,flask]==0.249.0 + strawberry-v0.269.0: strawberry-graphql[fastapi,flask]==0.269.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.228.0: pydantic<2.11 - strawberry-v0.247.2: pydantic<2.11 + strawberry-v0.229.2: pydantic<2.11 + strawberry-v0.249.0: pydantic<2.11 # ~~~ Network ~~~ @@ -628,17 +627,17 @@ deps = django-v1.11.29: django==1.11.29 django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 - django-v4.2.20: django==4.2.20 + django-v4.2.21: django==4.2.21 django-v5.0.14: django==5.0.14 - django-v5.2: django==5.2 + django-v5.2.1: django==5.2.1 django: psycopg2-binary django: djangorestframework django: pytest-django django: Werkzeug django-v3.2.25: pytest-asyncio - django-v4.2.20: pytest-asyncio + django-v4.2.21: pytest-asyncio django-v5.0.14: pytest-asyncio - django-v5.2: pytest-asyncio + django-v5.2.1: pytest-asyncio django-v2.2.28: six django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 @@ -650,14 +649,14 @@ deps = django-v2.2.28: pytest-django<4.0 django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] - django-v4.2.20: channels[daphne] + django-v4.2.21: channels[daphne] django-v5.0.14: channels[daphne] - django-v5.2: channels[daphne] + django-v5.2.1: channels[daphne] flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 flask-v3.0.3: flask==3.0.3 - flask-v3.1.0: flask==3.1.0 + flask-v3.1.1: flask==3.1.1 flask: flask-login flask: werkzeug flask-v1.1.4: werkzeug<2.1.0 @@ -739,13 +738,11 @@ deps = starlite: httpx<0.28 tornado-v6.0.4: tornado==6.0.4 - tornado-v6.1: tornado==6.1 tornado-v6.2: tornado==6.2 tornado-v6.4.2: tornado==6.4.2 - tornado-v6.5b1: tornado==6.5b1 + tornado-v6.5: tornado==6.5 tornado: pytest tornado-v6.0.4: pytest<8.2 - tornado-v6.1: pytest<8.2 tornado-v6.2: pytest<8.2 py3.6-tornado: aiocontextvars @@ -758,12 +755,12 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.31: trytond==7.0.31 - trytond-v7.6.0: trytond==7.6.0 + trytond-v7.6.1: trytond==7.6.1 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 - typer-v0.15.3: typer==0.15.3 + typer-v0.15.4: typer==0.15.4 From 9582a482fc9f4d84ee2a3ab6863eaed720821258 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 20 May 2025 14:57:41 +0200 Subject: [PATCH 305/591] fix(redis): Use `command_queue` instead of `command_stack` if available (#4404) This [commit](https://github.com/redis/redis-py/commit/91be4a0ff39556d5e270897740369d0b1cf9e044) in `redis-py` moved away from using the Pipeline's `command_stack` and is now using `_execution_strategy.command_queue` instead. We need to adapt to use the new command source if present. This fixes `redis-latest` in CI. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/integrations/redis/_sync_common.py | 8 +++++++- sentry_sdk/integrations/redis/utils.py | 10 +++++++--- 2 files changed, 14 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/redis/_sync_common.py b/sentry_sdk/integrations/redis/_sync_common.py index ef10e9e4f0..72f3eb7778 100644 --- a/sentry_sdk/integrations/redis/_sync_common.py +++ b/sentry_sdk/integrations/redis/_sync_common.py @@ -42,13 +42,19 @@ def sentry_patched_execute(self, *args, **kwargs): origin=SPAN_ORIGIN, ) as span: with capture_internal_exceptions(): + command_seq = None + try: + command_seq = self._execution_strategy.command_queue + except AttributeError: + command_seq = self.command_stack + set_db_data_fn(span, self) _set_pipeline_data( span, is_cluster, get_command_args_fn, False if is_cluster else self.transaction, - self.command_stack, + command_seq, ) return old_execute(self, *args, **kwargs) diff --git a/sentry_sdk/integrations/redis/utils.py b/sentry_sdk/integrations/redis/utils.py index 27fae1e8ca..cf230f6648 100644 --- a/sentry_sdk/integrations/redis/utils.py +++ b/sentry_sdk/integrations/redis/utils.py @@ -106,14 +106,18 @@ def _parse_rediscluster_command(command): def _set_pipeline_data( - span, is_cluster, get_command_args_fn, is_transaction, command_stack + span, + is_cluster, + get_command_args_fn, + is_transaction, + commands_seq, ): # type: (Span, bool, Any, bool, Sequence[Any]) -> None span.set_tag("redis.is_cluster", is_cluster) span.set_tag("redis.transaction", is_transaction) commands = [] - for i, arg in enumerate(command_stack): + for i, arg in enumerate(commands_seq): if i >= _MAX_NUM_COMMANDS: break @@ -123,7 +127,7 @@ def _set_pipeline_data( span.set_data( "redis.commands", { - "count": len(command_stack), + "count": len(commands_seq), "first_ten": commands, }, ) From 385c668272c375aa19bbf1860043b88bef7a529d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 20 May 2025 15:59:12 +0200 Subject: [PATCH 306/591] fix(grpc): Fix AttributeError when instrumenting with OTel (#4405) OTel also wraps `grpc.aio.server`, but expects the `interceptors` arg to be a list, while Sentry [turns it into a tuple](https://github.com/getsentry/sentry-python/blob/94414cdcdce2a311355967ee7b43e7035cf954ce/sentry_sdk/integrations/grpc/__init__.py#L130). Non-tuple sequences are actually only supported [starting](https://github.com/grpc/grpc/commit/96995801d878b4d39f329a51d664902e6125c07a) in grpc 1.42.0. So for older versions we need to still use a tuple. Fixes https://github.com/getsentry/sentry-python/issues/4389 --- sentry_sdk/integrations/grpc/__init__.py | 19 ++++++++++++++++++- 1 file changed, 18 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/grpc/__init__.py b/sentry_sdk/integrations/grpc/__init__.py index d9dcdddb55..4e15f95ae5 100644 --- a/sentry_sdk/integrations/grpc/__init__.py +++ b/sentry_sdk/integrations/grpc/__init__.py @@ -6,6 +6,7 @@ from grpc.aio import Server as AsyncServer from sentry_sdk.integrations import Integration +from sentry_sdk.utils import parse_version from .client import ClientInterceptor from .server import ServerInterceptor @@ -41,6 +42,8 @@ def __getitem__(self, _): P = ParamSpec("P") +GRPC_VERSION = parse_version(grpc.__version__) + def _wrap_channel_sync(func: Callable[P, Channel]) -> Callable[P, Channel]: "Wrapper for synchronous secure and insecure channel." @@ -127,7 +130,21 @@ def patched_aio_server( # type: ignore **kwargs: P.kwargs, ) -> Server: server_interceptor = AsyncServerInterceptor() - interceptors = (server_interceptor, *(interceptors or [])) + interceptors = [ + server_interceptor, + *(interceptors or []), + ] # type: Sequence[grpc.ServerInterceptor] + + try: + # We prefer interceptors as a list because of compatibility with + # opentelemetry https://github.com/getsentry/sentry-python/issues/4389 + # However, prior to grpc 1.42.0, only tuples were accepted, so we + # have no choice there. + if GRPC_VERSION is not None and GRPC_VERSION < (1, 42, 0): + interceptors = tuple(interceptors) + except Exception: + pass + return func(*args, interceptors=interceptors, **kwargs) # type: ignore return patched_aio_server # type: ignore From aad481d1a868d3c78d64d167267c02e8283a1064 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 2 Jun 2025 10:11:49 +0200 Subject: [PATCH 307/591] Fix CI, adapt to new redis-py release (#4431) Three new versions of things that need addressing: - pytest-asyncio 1.0 [removes](https://pytest-asyncio.readthedocs.io/en/latest/reference/changelog.html#id2) the deprecated `event_loop` fixture - mypy 1.16.0 finds some new issues - [redis 6.2.0](https://github.com/redis/redis-py/releases/tag/v6.2.0): - adds the `execution_strategy` abstraction to async cluster pipelines as well, so the commands are no longer accessible directly and need to be accessed via the `strategy` - renames the `_client` attribute on async cluster pipeline to `cluster_client` --- sentry_sdk/_types.py | 4 ++-- sentry_sdk/integrations/django/asgi.py | 4 ++-- sentry_sdk/integrations/redis/_async_common.py | 10 +++++++++- sentry_sdk/integrations/redis/redis_cluster.py | 14 +++++++++++--- sentry_sdk/integrations/tornado.py | 2 +- tests/integrations/aiohttp/test_aiohttp.py | 16 +--------------- 6 files changed, 26 insertions(+), 24 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 7da76e63dc..8336617a8d 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -170,7 +170,7 @@ class SDKInfo(TypedDict): "contexts": dict[str, dict[str, object]], "dist": str, "duration": Optional[float], - "environment": str, + "environment": Optional[str], "errors": list[dict[str, Any]], # TODO: We can expand on this type "event_id": str, "exception": dict[ @@ -188,7 +188,7 @@ class SDKInfo(TypedDict): "monitor_slug": Optional[str], "platform": Literal["python"], "profile": object, # Should be sentry_sdk.profiler.Profile, but we can't import that here due to circular imports - "release": str, + "release": Optional[str], "request": dict[str, object], "sdk": Mapping[str, object], "server_name": str, diff --git a/sentry_sdk/integrations/django/asgi.py b/sentry_sdk/integrations/django/asgi.py index 73a25acc9f..63a3f0b8f2 100644 --- a/sentry_sdk/integrations/django/asgi.py +++ b/sentry_sdk/integrations/django/asgi.py @@ -237,9 +237,9 @@ async def __acall__(self, *args, **kwargs): middleware_span = _check_middleware_span(old_method=f) if middleware_span is None: - return await f(*args, **kwargs) + return await f(*args, **kwargs) # type: ignore with middleware_span: - return await f(*args, **kwargs) + return await f(*args, **kwargs) # type: ignore return SentryASGIMixin diff --git a/sentry_sdk/integrations/redis/_async_common.py b/sentry_sdk/integrations/redis/_async_common.py index 196e85e74b..b96986fba3 100644 --- a/sentry_sdk/integrations/redis/_async_common.py +++ b/sentry_sdk/integrations/redis/_async_common.py @@ -41,13 +41,21 @@ async def _sentry_execute(self, *args, **kwargs): origin=SPAN_ORIGIN, ) as span: with capture_internal_exceptions(): + try: + command_seq = self._execution_strategy._command_queue + except AttributeError: + if is_cluster: + command_seq = self._command_stack + else: + command_seq = self.command_stack + set_db_data_fn(span, self) _set_pipeline_data( span, is_cluster, get_command_args_fn, False if is_cluster else self.is_transaction, - self._command_stack if is_cluster else self.command_stack, + command_seq, ) return await old_execute(self, *args, **kwargs) diff --git a/sentry_sdk/integrations/redis/redis_cluster.py b/sentry_sdk/integrations/redis/redis_cluster.py index 80cdc7235a..52936d1512 100644 --- a/sentry_sdk/integrations/redis/redis_cluster.py +++ b/sentry_sdk/integrations/redis/redis_cluster.py @@ -36,11 +36,19 @@ def _set_async_cluster_db_data(span, async_redis_cluster_instance): def _set_async_cluster_pipeline_db_data(span, async_redis_cluster_pipeline_instance): # type: (Span, AsyncClusterPipeline[Any]) -> None with capture_internal_exceptions(): + client = getattr(async_redis_cluster_pipeline_instance, "cluster_client", None) + if client is None: + # In older redis-py versions, the AsyncClusterPipeline had a `_client` + # attr but it is private so potentially problematic and mypy does not + # recognize it - see + # https://github.com/redis/redis-py/blame/v5.0.0/redis/asyncio/cluster.py#L1386 + client = ( + async_redis_cluster_pipeline_instance._client # type: ignore[attr-defined] + ) + _set_async_cluster_db_data( span, - # the AsyncClusterPipeline has always had a `_client` attr but it is private so potentially problematic and mypy - # does not recognize it - see https://github.com/redis/redis-py/blame/v5.0.0/redis/asyncio/cluster.py#L1386 - async_redis_cluster_pipeline_instance._client, # type: ignore[attr-defined] + client, ) diff --git a/sentry_sdk/integrations/tornado.py b/sentry_sdk/integrations/tornado.py index 3cd087524a..83fe5e94e8 100644 --- a/sentry_sdk/integrations/tornado.py +++ b/sentry_sdk/integrations/tornado.py @@ -76,7 +76,7 @@ async def sentry_execute_request_handler(self, *args, **kwargs): else: @coroutine # type: ignore - def sentry_execute_request_handler(self, *args, **kwargs): # type: ignore + def sentry_execute_request_handler(self, *args, **kwargs): # type: (RequestHandler, *Any, **Any) -> Any with _handle_request_impl(self): result = yield from old_execute(self, *args, **kwargs) diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index 06859b127f..47152f254c 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -6,11 +6,6 @@ import pytest -try: - import pytest_asyncio -except ImportError: - pytest_asyncio = None - from aiohttp import web, ClientSession from aiohttp.client import ServerDisconnectedError from aiohttp.web_request import Request @@ -27,14 +22,6 @@ from tests.conftest import ApproxDict -if pytest_asyncio is None: - # `loop` was deprecated in `pytest-aiohttp` - # in favor of `event_loop` from `pytest-asyncio` - @pytest.fixture - def event_loop(loop): - yield loop - - @pytest.mark.asyncio async def test_basic(sentry_init, aiohttp_client, capture_events): sentry_init(integrations=[AioHttpIntegration()]) @@ -490,7 +477,7 @@ async def hello(request): @pytest.mark.asyncio async def test_crumb_capture( - sentry_init, aiohttp_raw_server, aiohttp_client, event_loop, capture_events + sentry_init, aiohttp_raw_server, aiohttp_client, capture_events ): def before_breadcrumb(crumb, hint): crumb["data"]["extra"] = "foo" @@ -546,7 +533,6 @@ async def test_crumb_capture_client_error( sentry_init, aiohttp_raw_server, aiohttp_client, - event_loop, capture_events, status_code, level, From 7cf4ee409ca4a86675b1dcaddcbd90af0e04be2e Mon Sep 17 00:00:00 2001 From: Manabu Niseki Date: Mon, 2 Jun 2025 17:38:15 +0900 Subject: [PATCH 308/591] fix: fix ARQ integration error (#4427) (#4428) Fix #4427 by setting an empty list if `settings_cls.cron_jobs` is None. --- sentry_sdk/integrations/arq.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/arq.py b/sentry_sdk/integrations/arq.py index 1ea8e32fb3..b0b3d3f03e 100644 --- a/sentry_sdk/integrations/arq.py +++ b/sentry_sdk/integrations/arq.py @@ -214,7 +214,8 @@ def _sentry_create_worker(*args, **kwargs): ] if hasattr(settings_cls, "cron_jobs"): settings_cls.cron_jobs = [ - _get_arq_cron_job(cron_job) for cron_job in settings_cls.cron_jobs + _get_arq_cron_job(cron_job) + for cron_job in (settings_cls.cron_jobs or []) ] if "functions" in kwargs: From c2d5a768a780375e75a6f24beb1358f92eb5766f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 3 Jun 2025 16:32:40 +0200 Subject: [PATCH 309/591] tests: Regenerate tox.ini & fix CI (#4435) Regular tox update and fix some stuff pyspark 4.0.0 came out and it needs a different java version, so updated that too Closes https://github.com/getsentry/sentry-python/issues/4438 --- .github/workflows/test-integrations-tasks.yml | 10 ++ scripts/populate_tox/config.py | 4 +- scripts/populate_tox/tox.jinja | 2 + .../split_tox_gh_actions.py | 5 + .../templates/test_group.jinja | 9 ++ tests/integrations/django/asgi/test_asgi.py | 15 +++ tox.ini | 106 +++++++++--------- 7 files changed, 98 insertions(+), 53 deletions(-) diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 43b5d8d56d..91b47f90c6 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -46,6 +46,11 @@ jobs: allow-prereleases: true - name: Start Redis uses: supercharge/redis-github-action@1.8.0 + - name: Install Java + uses: actions/setup-java@v4 + with: + distribution: 'temurin' + java-version: '21' - name: Setup Test Env run: | pip install "coverage[toml]" tox @@ -135,6 +140,11 @@ jobs: allow-prereleases: true - name: Start Redis uses: supercharge/redis-github-action@1.8.0 + - name: Install Java + uses: actions/setup-java@v4 + with: + distribution: 'temurin' + java-version: '21' - name: Setup Test Env run: | pip install "coverage[toml]" tox diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 4d5d5b14ce..4664845c7b 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -58,14 +58,14 @@ "pytest-django", "Werkzeug", ], - ">=3.0": ["pytest-asyncio"], + ">=2.0": ["channels[daphne]"], ">=2.2,<3.1": ["six"], + ">=3.0": ["pytest-asyncio"], "<3.3": [ "djangorestframework>=3.0,<4.0", "Werkzeug<2.1.0", ], "<3.1": ["pytest-django<4.0"], - ">=2.0": ["channels[daphne]"], }, }, "dramatiq": { diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 2869da275b..3f3691147e 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -172,6 +172,7 @@ deps = # for justification of the upper bound on pytest {py3.6,py3.7}-gevent: pytest<7.0.0 {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest + gevent: pytest-asyncio # === Integrations === @@ -362,6 +363,7 @@ setenv = py3.6: COVERAGE_RCFILE=.coveragerc36 django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings + spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 common: TESTPATH=tests gevent: TESTPATH=tests diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 293af897c9..3fbc0ec1c5 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -47,6 +47,10 @@ "aws_lambda", } +FRAMEWORKS_NEEDING_JAVA = { + "spark", +} + # Frameworks grouped here will be tested together to not hog all GitHub runners. # If you add or remove a group, make sure to git rm the generated YAML file as # well. @@ -288,6 +292,7 @@ def render_template(group, frameworks, py_versions_pinned, py_versions_latest): "needs_docker": bool(set(frameworks) & FRAMEWORKS_NEEDING_DOCKER), "needs_postgres": bool(set(frameworks) & FRAMEWORKS_NEEDING_POSTGRES), "needs_redis": bool(set(frameworks) & FRAMEWORKS_NEEDING_REDIS), + "needs_java": bool(set(frameworks) & FRAMEWORKS_NEEDING_JAVA), "py_versions": { category: [f'"{version}"' for version in _normalize_py_versions(versions)] for category, versions in py_versions.items() diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 705ae69a7f..44f51f473e 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -48,6 +48,7 @@ with: python-version: {% raw %}${{ matrix.python-version }}{% endraw %} allow-prereleases: true + {% if needs_clickhouse %} - name: "Setup ClickHouse Server" uses: getsentry/action-clickhouse-in-ci@v1.6 @@ -58,6 +59,14 @@ uses: supercharge/redis-github-action@1.8.0 {% endif %} + {% if needs_java %} + - name: Install Java + uses: actions/setup-java@v4 + with: + distribution: 'temurin' + java-version: '21' + {% endif %} + - name: Setup Test Env run: | pip install "coverage[toml]" tox diff --git a/tests/integrations/django/asgi/test_asgi.py b/tests/integrations/django/asgi/test_asgi.py index 82eae30b1d..3c78ac3f38 100644 --- a/tests/integrations/django/asgi/test_asgi.py +++ b/tests/integrations/django/asgi/test_asgi.py @@ -30,6 +30,9 @@ @pytest.mark.parametrize("application", APPS) @pytest.mark.asyncio @pytest.mark.forked +@pytest.mark.skipif( + django.VERSION < (3, 0), reason="Django ASGI support shipped in 3.0" +) async def test_basic(sentry_init, capture_events, application): sentry_init( integrations=[DjangoIntegration()], @@ -579,6 +582,9 @@ async def test_asgi_request_body( "asyncio.iscoroutinefunction has been replaced in 3.12 by inspect.iscoroutinefunction" ), ) +@pytest.mark.skipif( + django.VERSION < (3, 0), reason="Django ASGI support shipped in 3.0" +) async def test_asgi_mixin_iscoroutinefunction_before_3_12(): sentry_asgi_mixin = _asgi_middleware_mixin_factory(lambda: None) @@ -636,6 +642,9 @@ def get_response(): ... @pytest.mark.parametrize("application", APPS) @pytest.mark.asyncio +@pytest.mark.skipif( + django.VERSION < (3, 1), reason="async views have been introduced in Django 3.1" +) async def test_async_view(sentry_init, capture_events, application): sentry_init( integrations=[DjangoIntegration()], @@ -655,6 +664,9 @@ async def test_async_view(sentry_init, capture_events, application): @pytest.mark.parametrize("application", APPS) @pytest.mark.asyncio +@pytest.mark.skipif( + django.VERSION < (3, 0), reason="Django ASGI support shipped in 3.0" +) async def test_transaction_http_method_default( sentry_init, capture_events, application ): @@ -687,6 +699,9 @@ async def test_transaction_http_method_default( @pytest.mark.parametrize("application", APPS) @pytest.mark.asyncio +@pytest.mark.skipif( + django.VERSION < (3, 0), reason="Django ASGI support shipped in 3.0" +) async def test_transaction_http_method_custom(sentry_init, capture_events, application): sentry_init( integrations=[ diff --git a/tox.ini b/tox.ini index d9632b7cb5..2fe4977a1c 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-05-20T07:21:06.821358+00:00 +# Last generated: 2025-06-03T13:53:53.259798+00:00 [tox] requires = @@ -138,7 +138,7 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.28.1 {py3.8,py3.11,py3.12}-anthropic-v0.40.0 - {py3.8,py3.11,py3.12}-anthropic-v0.51.0 + {py3.8,py3.11,py3.12}-anthropic-v0.52.2 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 @@ -148,7 +148,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.31.4 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.4 # ~~~ DBs ~~~ @@ -172,7 +172,7 @@ envlist = {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 - {py3.8,py3.12,py3.13}-launchdarkly-v9.11.0 + {py3.8,py3.12,py3.13}-launchdarkly-v9.11.1 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.1 @@ -180,10 +180,11 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 + {py3.7,py3.12,py3.13}-statsig-v0.58.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 - {py3.8,py3.12,py3.13}-unleash-v6.2.0 + {py3.8,py3.12,py3.13}-unleash-v6.2.1 # ~~~ GraphQL ~~~ @@ -193,8 +194,8 @@ envlist = {py3.9,py3.12,py3.13}-ariadne-v0.26.2 {py3.6,py3.9,py3.10}-gql-v3.4.1 - {py3.7,py3.11,py3.12}-gql-v3.5.2 - {py3.9,py3.12,py3.13}-gql-v3.6.0b4 + {py3.7,py3.11,py3.12}-gql-v3.5.3 + {py3.9,py3.12,py3.13}-gql-v4.0.0b0 {py3.6,py3.9,py3.10}-graphene-v3.3 {py3.8,py3.12,py3.13}-graphene-v3.4.3 @@ -202,26 +203,26 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.229.2 {py3.8,py3.12,py3.13}-strawberry-v0.249.0 - {py3.9,py3.12,py3.13}-strawberry-v0.269.0 + {py3.9,py3.12,py3.13}-strawberry-v0.270.5 # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 - {py3.7,py3.9,py3.10}-grpc-v1.44.0 - {py3.7,py3.10,py3.11}-grpc-v1.58.3 - {py3.9,py3.12,py3.13}-grpc-v1.71.0 - {py3.9,py3.12,py3.13}-grpc-v1.72.0rc1 + {py3.7,py3.9,py3.10}-grpc-v1.46.5 + {py3.7,py3.11,py3.12}-grpc-v1.60.2 + {py3.9,py3.12,py3.13}-grpc-v1.72.1 + {py3.9,py3.12,py3.13}-grpc-v1.73.0rc1 # ~~~ Tasks ~~~ {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 - {py3.8,py3.12,py3.13}-celery-v5.5.2 + {py3.8,py3.12,py3.13}-celery-v5.5.3 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 {py3.7,py3.10,py3.11}-dramatiq-v1.15.0 - {py3.8,py3.12,py3.13}-dramatiq-v1.17.1 + {py3.9,py3.12,py3.13}-dramatiq-v1.18.0 {py3.6,py3.7}-huey-v2.1.3 {py3.6,py3.7}-huey-v2.2.0 @@ -229,9 +230,8 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.3 {py3.8,py3.9}-spark-v3.0.3 - {py3.8,py3.9}-spark-v3.2.4 - {py3.8,py3.10,py3.11}-spark-v3.4.4 - {py3.8,py3.10,py3.11}-spark-v3.5.5 + {py3.8,py3.10,py3.11}-spark-v3.5.6 + {py3.9,py3.12,py3.13}-spark-v4.0.0 # ~~~ Web 1 ~~~ @@ -250,7 +250,7 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.46.2 + {py3.9,py3.12,py3.13}-starlette-v0.47.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 @@ -260,9 +260,9 @@ envlist = # ~~~ Web 2 ~~~ {py3.7}-aiohttp-v3.4.4 - {py3.7}-aiohttp-v3.6.3 - {py3.7,py3.9,py3.10}-aiohttp-v3.8.6 - {py3.9,py3.12,py3.13}-aiohttp-v3.11.18 + {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 + {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 + {py3.9,py3.12,py3.13}-aiohttp-v3.12.7 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.3 @@ -289,7 +289,7 @@ envlist = {py3.6,py3.7,py3.8}-tornado-v6.0.4 {py3.7,py3.9,py3.10}-tornado-v6.2 {py3.8,py3.10,py3.11}-tornado-v6.4.2 - {py3.9,py3.12,py3.13}-tornado-v6.5 + {py3.9,py3.12,py3.13}-tornado-v6.5.1 # ~~~ Misc ~~~ @@ -303,6 +303,7 @@ envlist = {py3.9,py3.12,py3.13}-trytond-v7.6.1 {py3.7,py3.12,py3.13}-typer-v0.15.4 + {py3.7,py3.12,py3.13}-typer-v0.16.0 @@ -333,6 +334,7 @@ deps = # for justification of the upper bound on pytest {py3.6,py3.7}-gevent: pytest<7.0.0 {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest + gevent: pytest-asyncio # === Integrations === @@ -502,7 +504,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.28.1: anthropic==0.28.1 anthropic-v0.40.0: anthropic==0.40.0 - anthropic-v0.51.0: anthropic==0.51.0 + anthropic-v0.52.2: anthropic==0.52.2 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.28.1: httpx<0.28.0 @@ -516,7 +518,7 @@ deps = huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.25.2: huggingface_hub==0.25.2 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.31.4: huggingface_hub==0.31.4 + huggingface_hub-v0.32.4: huggingface_hub==0.32.4 # ~~~ DBs ~~~ @@ -541,7 +543,7 @@ deps = launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 - launchdarkly-v9.11.0: launchdarkly-server-sdk==9.11.0 + launchdarkly-v9.11.1: launchdarkly-server-sdk==9.11.1 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.1: openfeature-sdk==0.8.1 @@ -549,11 +551,12 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 + statsig-v0.58.0: statsig==0.58.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 unleash-v6.1.0: UnleashClient==6.1.0 - unleash-v6.2.0: UnleashClient==6.2.0 + unleash-v6.2.1: UnleashClient==6.2.1 # ~~~ GraphQL ~~~ @@ -566,8 +569,8 @@ deps = ariadne: httpx gql-v3.4.1: gql[all]==3.4.1 - gql-v3.5.2: gql[all]==3.5.2 - gql-v3.6.0b4: gql[all]==3.6.0b4 + gql-v3.5.3: gql[all]==3.5.3 + gql-v4.0.0b0: gql[all]==4.0.0b0 graphene-v3.3: graphene==3.3 graphene-v3.4.3: graphene==3.4.3 @@ -580,7 +583,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.229.2: strawberry-graphql[fastapi,flask]==0.229.2 strawberry-v0.249.0: strawberry-graphql[fastapi,flask]==0.249.0 - strawberry-v0.269.0: strawberry-graphql[fastapi,flask]==0.269.0 + strawberry-v0.270.5: strawberry-graphql[fastapi,flask]==0.270.5 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.229.2: pydantic<2.11 @@ -589,10 +592,10 @@ deps = # ~~~ Network ~~~ grpc-v1.32.0: grpcio==1.32.0 - grpc-v1.44.0: grpcio==1.44.0 - grpc-v1.58.3: grpcio==1.58.3 - grpc-v1.71.0: grpcio==1.71.0 - grpc-v1.72.0rc1: grpcio==1.72.0rc1 + grpc-v1.46.5: grpcio==1.46.5 + grpc-v1.60.2: grpcio==1.60.2 + grpc-v1.72.1: grpcio==1.72.1 + grpc-v1.73.0rc1: grpcio==1.73.0rc1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -602,7 +605,7 @@ deps = # ~~~ Tasks ~~~ celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 - celery-v5.5.2: celery==5.5.2 + celery-v5.5.3: celery==5.5.3 celery: newrelic celery: redis py3.7-celery: importlib-metadata<5.0 @@ -610,7 +613,7 @@ deps = dramatiq-v1.9.0: dramatiq==1.9.0 dramatiq-v1.12.3: dramatiq==1.12.3 dramatiq-v1.15.0: dramatiq==1.15.0 - dramatiq-v1.17.1: dramatiq==1.17.1 + dramatiq-v1.18.0: dramatiq==1.18.0 huey-v2.1.3: huey==2.1.3 huey-v2.2.0: huey==2.2.0 @@ -618,9 +621,8 @@ deps = huey-v2.5.3: huey==2.5.3 spark-v3.0.3: pyspark==3.0.3 - spark-v3.2.4: pyspark==3.2.4 - spark-v3.4.4: pyspark==3.4.4 - spark-v3.5.5: pyspark==3.5.5 + spark-v3.5.6: pyspark==3.5.6 + spark-v4.0.0: pyspark==4.0.0 # ~~~ Web 1 ~~~ @@ -634,11 +636,16 @@ deps = django: djangorestframework django: pytest-django django: Werkzeug + django-v2.2.28: channels[daphne] + django-v3.2.25: channels[daphne] + django-v4.2.21: channels[daphne] + django-v5.0.14: channels[daphne] + django-v5.2.1: channels[daphne] + django-v2.2.28: six django-v3.2.25: pytest-asyncio django-v4.2.21: pytest-asyncio django-v5.0.14: pytest-asyncio django-v5.2.1: pytest-asyncio - django-v2.2.28: six django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 @@ -647,11 +654,6 @@ deps = django-v3.2.25: Werkzeug<2.1.0 django-v1.11.29: pytest-django<4.0 django-v2.2.28: pytest-django<4.0 - django-v2.2.28: channels[daphne] - django-v3.2.25: channels[daphne] - django-v4.2.21: channels[daphne] - django-v5.0.14: channels[daphne] - django-v5.2.1: channels[daphne] flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 @@ -665,7 +667,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.46.2: starlette==0.46.2 + starlette-v0.47.0: starlette==0.47.0 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -694,12 +696,12 @@ deps = # ~~~ Web 2 ~~~ aiohttp-v3.4.4: aiohttp==3.4.4 - aiohttp-v3.6.3: aiohttp==3.6.3 - aiohttp-v3.8.6: aiohttp==3.8.6 - aiohttp-v3.11.18: aiohttp==3.11.18 + aiohttp-v3.7.4: aiohttp==3.7.4 + aiohttp-v3.10.11: aiohttp==3.10.11 + aiohttp-v3.12.7: aiohttp==3.12.7 aiohttp: pytest-aiohttp - aiohttp-v3.8.6: pytest-asyncio - aiohttp-v3.11.18: pytest-asyncio + aiohttp-v3.10.11: pytest-asyncio + aiohttp-v3.12.7: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.3: bottle==0.13.3 @@ -740,7 +742,7 @@ deps = tornado-v6.0.4: tornado==6.0.4 tornado-v6.2: tornado==6.2 tornado-v6.4.2: tornado==6.4.2 - tornado-v6.5: tornado==6.5 + tornado-v6.5.1: tornado==6.5.1 tornado: pytest tornado-v6.0.4: pytest<8.2 tornado-v6.2: pytest<8.2 @@ -761,6 +763,7 @@ deps = trytond-v4.8.18: werkzeug<1.0 typer-v0.15.4: typer==0.15.4 + typer-v0.16.0: typer==0.16.0 @@ -771,6 +774,7 @@ setenv = py3.6: COVERAGE_RCFILE=.coveragerc36 django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings + spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 common: TESTPATH=tests gevent: TESTPATH=tests From a662a9de6da34237a80ce72e00ba46279091b792 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 5 Jun 2025 15:39:30 +0200 Subject: [PATCH 310/591] feat(logs): Add user attributes to logs (#4423) Example log with user attributes: https://sentry-sdks.sentry.io/explore/logs/trace/362b3309f7f9471eb504af407bf51932/?logsSortBys=-timestamp&project=4505239425777664&source=logs&tab=logs×tamp=1748448946 Closes #4422 --- sentry_sdk/client.py | 15 +++++++++++++++ tests/test_logs.py | 43 +++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 58 insertions(+) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index a03598a981..26baeea957 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -927,6 +927,21 @@ def _capture_experimental_log(self, current_scope, log): elif propagation_context is not None: log["trace_id"] = propagation_context.trace_id + # The user, if present, is always set on the isolation scope. + if self.should_send_default_pii() and isolation_scope._user is not None: + for log_attribute, user_attribute in ( + ("user.id", "id"), + ("user.name", "username"), + ("user.email", "email"), + ): + if ( + user_attribute in isolation_scope._user + and log_attribute not in log["attributes"] + ): + log["attributes"][log_attribute] = isolation_scope._user[ + user_attribute + ] + # If debug is enabled, log the log to the console debug = self.options.get("debug", False) if debug: diff --git a/tests/test_logs.py b/tests/test_logs.py index 7e8a72d30a..94c0f4ce6f 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -484,6 +484,49 @@ def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): raise AssertionError("200 logs were never flushed after five seconds") +def test_user_attributes(sentry_init, capture_envelopes): + """User attributes are sent if send_default_pii is True.""" + sentry_init(send_default_pii=True, _experiments={"enable_logs": True}) + + sentry_sdk.set_user({"id": "1", "email": "test@example.com", "username": "test"}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("Hello, world!") + + get_client().flush() + + logs = envelopes_to_logs(envelopes) + (log,) = logs + + # Check that all expected user attributes are present. + assert log["attributes"].items() >= { + ("user.id", "1"), + ("user.email", "test@example.com"), + ("user.name", "test"), + } + + +def test_user_attributes_no_pii(sentry_init, capture_envelopes): + """Ensure no user attributes are sent if send_default_pii is False.""" + sentry_init(_experiments={"enable_logs": True}) + + sentry_sdk.set_user({"id": "1", "email": "test@example.com", "username": "test"}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("Hello, world!") + + get_client().flush() + + logs = envelopes_to_logs(envelopes) + + (log,) = logs + assert "user.id" not in log["attributes"] + assert "user.email" not in log["attributes"] + assert "user.name" not in log["attributes"] + + @minimum_python_37 def test_auto_flush_logs_after_5s(sentry_init, capture_envelopes): """ From 7f8571cf03f57a45e8013a04385164a6b907d661 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 5 Jun 2025 15:39:43 +0200 Subject: [PATCH 311/591] ref: Remove `_capture_experimental_log` `scope` parameter (#4424) We are always just using the current scope anyway; it is less confusing if we eliminate the parameter Stacked on: - #4423 --- sentry_sdk/client.py | 13 ++++++++----- sentry_sdk/integrations/logging.py | 2 -- sentry_sdk/logger.py | 4 +--- 3 files changed, 9 insertions(+), 10 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 26baeea957..9d9a6473e8 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -8,6 +8,7 @@ from typing import TYPE_CHECKING, List, Dict, cast, overload import warnings +import sentry_sdk from sentry_sdk._compat import PY37, check_uwsgi_thread_support from sentry_sdk.utils import ( AnnotatedValue, @@ -215,8 +216,8 @@ def capture_event(self, *args, **kwargs): # type: (*Any, **Any) -> Optional[str] return None - def _capture_experimental_log(self, scope, log): - # type: (Scope, Log) -> None + def _capture_experimental_log(self, log): + # type: (Log) -> None pass def capture_session(self, *args, **kwargs): @@ -893,12 +894,14 @@ def capture_event( return return_value - def _capture_experimental_log(self, current_scope, log): - # type: (Scope, Log) -> None + def _capture_experimental_log(self, log): + # type: (Log) -> None logs_enabled = self.options["_experiments"].get("enable_logs", False) if not logs_enabled: return - isolation_scope = current_scope.get_isolation_scope() + + current_scope = sentry_sdk.get_current_scope() + isolation_scope = sentry_sdk.get_isolation_scope() log["attributes"]["sentry.sdk.name"] = SDK_INFO["name"] log["attributes"]["sentry.sdk.version"] = SDK_INFO["version"] diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 449a05fbf7..dddcc17334 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -352,7 +352,6 @@ def emit(self, record): def _capture_log_from_record(self, client, record): # type: (BaseClient, LogRecord) -> None - scope = sentry_sdk.get_current_scope() otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) project_root = client.options["project_root"] attrs = self._extra_from_record(record) # type: Any @@ -394,7 +393,6 @@ def _capture_log_from_record(self, client, record): # noinspection PyProtectedMember client._capture_experimental_log( - scope, { "severity_text": otel_severity_text, "severity_number": otel_severity_number, diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index c675c4d95d..2f5e859533 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -3,14 +3,13 @@ import time from typing import Any -from sentry_sdk import get_client, get_current_scope +from sentry_sdk import get_client from sentry_sdk.utils import safe_repr def _capture_log(severity_text, severity_number, template, **kwargs): # type: (str, int, str, **Any) -> None client = get_client() - scope = get_current_scope() attrs = { "sentry.message.template": template, @@ -36,7 +35,6 @@ def _capture_log(severity_text, severity_number, template, **kwargs): # noinspection PyProtectedMember client._capture_experimental_log( - scope, { "severity_text": severity_text, "severity_number": severity_number, From 4420c4d92c8c8939ee240800d8dddb5ffc7c858d Mon Sep 17 00:00:00 2001 From: Emmanuel Ferdman Date: Tue, 10 Jun 2025 14:35:22 +0300 Subject: [PATCH 312/591] Migrate to modern threading interface (#4452) ### PR Summary This small PR resolves the `threading` library warnings, which you can find in the [CI logs](https://github.com/getsentry/sentry-python/actions/runs/15490014338/job/43613162178#step:6:3523): ```python /home/runner/work/sentry-python/sentry-python/tests/conftest.py:608: DeprecationWarning: setDaemon() is deprecated, set the daemon attribute instead mock_server_thread.setDaemon(True) ``` Signed-off-by: Emmanuel Ferdman --- tests/conftest.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/conftest.py b/tests/conftest.py index b5f3f8b00e..6a33029d11 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -612,7 +612,7 @@ def create_mock_http_server(): mock_server_port = get_free_port() mock_server = HTTPServer(("localhost", mock_server_port), MockServerRequestHandler) mock_server_thread = Thread(target=mock_server.serve_forever) - mock_server_thread.setDaemon(True) + mock_server_thread.daemon = True mock_server_thread.start() return mock_server_port From 8ca298aadf19cc5e2508dae6b281bd37432a1dda Mon Sep 17 00:00:00 2001 From: Romain Girard Date: Tue, 10 Jun 2025 14:17:29 +0200 Subject: [PATCH 313/591] fix(logging): Strip log `record.name` for more robust matching (#4411) Hi! In the Python SDK, more specifically `sentry_sdk/integrations/logging.py`, a couple of loggers are ignored to avoid recursion errors. ```py _IGNORED_LOGGERS = set( ["sentry_sdk.errors", "urllib3.connectionpool", "urllib3.connection"] ) ``` Log records from these loggers are discarded, using an exact match on `record.name`. Unforunately, this breaks if the user modifies `record.name`, e.g. for formatting, which is what we were doing for log display (before becoming aware that Sentry relied on it after investigating an infinite recursion issue). ```py class _LengthFormatter(logging.Formatter): def format(self, record): """ Format a log record's header to a constant length """ if len(record.name) > _MAX_LOGGER_NAME_LENGTH: sep = "..." cut = _MAX_LOGGER_NAME_LENGTH // 3 - len(sep) record.name = record.name[:cut] + sep + record.name[-(_MAX_LOGGER_NAME_LENGTH - cut - len(sep)) :] record.name = record.name.ljust(_MAX_LOGGER_NAME_LENGTH) record.levelname = record.levelname.ljust(_MAX_LOGGER_LEVEL_NAME_LENGTH) return super().format(record) ``` As you can see, `record.name` is right-padded with blank spaces. We have found a workaround since, but given that it has taken us quite some time to find the issue, I thought that maybe it could affect others. This PR proposes matching `record.name.strip()` instead for increased robustness. --- sentry_sdk/integrations/logging.py | 7 +++++-- tests/integrations/logging/test_logging.py | 12 ++++++++++++ 2 files changed, 17 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index dddcc17334..75848bc985 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -119,7 +119,10 @@ def sentry_patched_callhandlers(self, record): # the integration. Otherwise we have a high chance of getting # into a recursion error when the integration is resolved # (this also is slower). - if ignored_loggers is not None and record.name not in ignored_loggers: + if ( + ignored_loggers is not None + and record.name.strip() not in ignored_loggers + ): integration = sentry_sdk.get_client().get_integration( LoggingIntegration ) @@ -164,7 +167,7 @@ def _can_record(self, record): # type: (LogRecord) -> bool """Prevents ignored loggers from recording""" for logger in _IGNORED_LOGGERS: - if fnmatch(record.name, logger): + if fnmatch(record.name.strip(), logger): return False return True diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index c08e960c00..557f3fc1b1 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -230,6 +230,18 @@ def test_ignore_logger(sentry_init, capture_events): assert not events +def test_ignore_logger_whitespace_padding(sentry_init, capture_events): + """Here we test insensitivity to whitespace padding of ignored loggers""" + sentry_init(integrations=[LoggingIntegration()], default_integrations=False) + events = capture_events() + + ignore_logger("testfoo") + + padded_logger = logging.getLogger(" testfoo ") + padded_logger.error("hi") + assert not events + + def test_ignore_logger_wildcard(sentry_init, capture_events): sentry_init(integrations=[LoggingIntegration()], default_integrations=False) events = capture_events() From 1433ec2af926b2a5f67f555533af3affdacc9f8d Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Tue, 10 Jun 2025 09:45:29 -0400 Subject: [PATCH 314/591] fix(logs): Don't gate user behind `send_default_pii` (#4453) ref https://github.com/getsentry/sentry-docs/pull/13974 --- sentry_sdk/client.py | 2 +- tests/test_logs.py | 26 +++----------------------- 2 files changed, 4 insertions(+), 24 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 9d9a6473e8..979ea92906 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -931,7 +931,7 @@ def _capture_experimental_log(self, log): log["trace_id"] = propagation_context.trace_id # The user, if present, is always set on the isolation scope. - if self.should_send_default_pii() and isolation_scope._user is not None: + if isolation_scope._user is not None: for log_attribute, user_attribute in ( ("user.id", "id"), ("user.name", "username"), diff --git a/tests/test_logs.py b/tests/test_logs.py index 94c0f4ce6f..dcbaba3c4f 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -484,9 +484,9 @@ def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): raise AssertionError("200 logs were never flushed after five seconds") -def test_user_attributes(sentry_init, capture_envelopes): - """User attributes are sent if send_default_pii is True.""" - sentry_init(send_default_pii=True, _experiments={"enable_logs": True}) +def test_log_user_attributes(sentry_init, capture_envelopes): + """User attributes are sent if enable_logs is True.""" + sentry_init(_experiments={"enable_logs": True}) sentry_sdk.set_user({"id": "1", "email": "test@example.com", "username": "test"}) envelopes = capture_envelopes() @@ -507,26 +507,6 @@ def test_user_attributes(sentry_init, capture_envelopes): } -def test_user_attributes_no_pii(sentry_init, capture_envelopes): - """Ensure no user attributes are sent if send_default_pii is False.""" - sentry_init(_experiments={"enable_logs": True}) - - sentry_sdk.set_user({"id": "1", "email": "test@example.com", "username": "test"}) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.warning("Hello, world!") - - get_client().flush() - - logs = envelopes_to_logs(envelopes) - - (log,) = logs - assert "user.id" not in log["attributes"] - assert "user.email" not in log["attributes"] - assert "user.name" not in log["attributes"] - - @minimum_python_37 def test_auto_flush_logs_after_5s(sentry_init, capture_envelopes): """ From 51db87ca6fd3c6cfc6eadd8ba800d72537a55d29 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 10 Jun 2025 16:26:32 +0200 Subject: [PATCH 315/591] feat(loguru): Sentry logs for Loguru (#4445) Allow to send Loguru logs to Sentry. We can't parametrize them nicely, but this is a good first step. Also: * Move some tests around. Tests specific to the stdlib logging integration were moved from the generic sentry logs tests to the logging integration tests. * Remove `@minimum_python_37` from some tests that don't need 3.7+. * Dedupe some code by moving it to a superclass (`_LoguruBaseHandler`) Closes https://github.com/getsentry/sentry-python/issues/4151 --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- CHANGELOG.md | 2 +- sentry_sdk/integrations/logging.py | 37 +-- sentry_sdk/integrations/loguru.py | 121 ++++++-- sentry_sdk/logger.py | 29 ++ tests/integrations/logging/test_logging.py | 197 ++++++++++++ tests/integrations/loguru/test_loguru.py | 341 ++++++++++++++++++++- tests/test_logs.py | 198 ------------ 7 files changed, 685 insertions(+), 240 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8d73e1bdd0..8dd7763121 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -122,7 +122,7 @@ sentry_sdk.init( dsn="...", _experiments={ - "enable_sentry_logs": True + "enable_logs": True } integrations=[ LoggingIntegration(sentry_logs_level=logging.ERROR), diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 75848bc985..62b1e09d64 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -5,6 +5,7 @@ import sentry_sdk from sentry_sdk.client import BaseClient +from sentry_sdk.logger import _log_level_to_otel from sentry_sdk.utils import ( safe_repr, to_string, @@ -14,7 +15,7 @@ ) from sentry_sdk.integrations import Integration -from typing import TYPE_CHECKING, Tuple +from typing import TYPE_CHECKING if TYPE_CHECKING: from collections.abc import MutableMapping @@ -36,6 +37,16 @@ logging.CRITICAL: "fatal", # CRITICAL is same as FATAL } +# Map logging level numbers to corresponding OTel level numbers +SEVERITY_TO_OTEL_SEVERITY = { + logging.CRITICAL: 21, # fatal + logging.ERROR: 17, # error + logging.WARNING: 13, # warn + logging.INFO: 9, # info + logging.DEBUG: 5, # debug +} + + # Capturing events from those loggers causes recursion errors. We cannot allow # the user to unconditionally create events from those loggers under any # circumstances. @@ -315,21 +326,6 @@ def _breadcrumb_from_record(self, record): } -def _python_level_to_otel(record_level): - # type: (int) -> Tuple[int, str] - for py_level, otel_severity_number, otel_severity_text in [ - (50, 21, "fatal"), - (40, 17, "error"), - (30, 13, "warn"), - (20, 9, "info"), - (10, 5, "debug"), - (5, 1, "trace"), - ]: - if record_level >= py_level: - return otel_severity_number, otel_severity_text - return 0, "default" - - class SentryLogsHandler(_BaseHandler): """ A logging handler that records Sentry logs for each Python log record. @@ -355,7 +351,9 @@ def emit(self, record): def _capture_log_from_record(self, client, record): # type: (BaseClient, LogRecord) -> None - otel_severity_number, otel_severity_text = _python_level_to_otel(record.levelno) + otel_severity_number, otel_severity_text = _log_level_to_otel( + record.levelno, SEVERITY_TO_OTEL_SEVERITY + ) project_root = client.options["project_root"] attrs = self._extra_from_record(record) # type: Any attrs["sentry.origin"] = "auto.logger.log" @@ -366,10 +364,7 @@ def _capture_log_from_record(self, client, record): for i, arg in enumerate(record.args): attrs[f"sentry.message.parameter.{i}"] = ( arg - if isinstance(arg, str) - or isinstance(arg, float) - or isinstance(arg, int) - or isinstance(arg, bool) + if isinstance(arg, (str, float, int, bool)) else safe_repr(arg) ) if record.lineno: diff --git a/sentry_sdk/integrations/loguru.py b/sentry_sdk/integrations/loguru.py index a71c4ac87f..df3ecf161a 100644 --- a/sentry_sdk/integrations/loguru.py +++ b/sentry_sdk/integrations/loguru.py @@ -1,22 +1,27 @@ import enum +import sentry_sdk from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.integrations.logging import ( BreadcrumbHandler, EventHandler, _BaseHandler, ) +from sentry_sdk.logger import _log_level_to_otel from typing import TYPE_CHECKING if TYPE_CHECKING: from logging import LogRecord - from typing import Optional, Any + from typing import Any, Optional try: import loguru from loguru import logger from loguru._defaults import LOGURU_FORMAT as DEFAULT_FORMAT + + if TYPE_CHECKING: + from loguru import Message except ImportError: raise DidNotEnable("LOGURU is not installed") @@ -31,6 +36,10 @@ class LoggingLevels(enum.IntEnum): CRITICAL = 50 +DEFAULT_LEVEL = LoggingLevels.INFO.value +DEFAULT_EVENT_LEVEL = LoggingLevels.ERROR.value + + SENTRY_LEVEL_FROM_LOGURU_LEVEL = { "TRACE": "DEBUG", "DEBUG": "DEBUG", @@ -41,8 +50,16 @@ class LoggingLevels(enum.IntEnum): "CRITICAL": "CRITICAL", } -DEFAULT_LEVEL = LoggingLevels.INFO.value -DEFAULT_EVENT_LEVEL = LoggingLevels.ERROR.value +# Map Loguru level numbers to corresponding OTel level numbers +SEVERITY_TO_OTEL_SEVERITY = { + LoggingLevels.CRITICAL: 21, # fatal + LoggingLevels.ERROR: 17, # error + LoggingLevels.WARNING: 13, # warn + LoggingLevels.SUCCESS: 11, # info + LoggingLevels.INFO: 9, # info + LoggingLevels.DEBUG: 5, # debug + LoggingLevels.TRACE: 1, # trace +} class LoguruIntegration(Integration): @@ -52,6 +69,7 @@ class LoguruIntegration(Integration): event_level = DEFAULT_EVENT_LEVEL # type: Optional[int] breadcrumb_format = DEFAULT_FORMAT event_format = DEFAULT_FORMAT + sentry_logs_level = DEFAULT_LEVEL # type: Optional[int] def __init__( self, @@ -59,12 +77,14 @@ def __init__( event_level=DEFAULT_EVENT_LEVEL, breadcrumb_format=DEFAULT_FORMAT, event_format=DEFAULT_FORMAT, + sentry_logs_level=DEFAULT_LEVEL, ): - # type: (Optional[int], Optional[int], str | loguru.FormatFunction, str | loguru.FormatFunction) -> None + # type: (Optional[int], Optional[int], str | loguru.FormatFunction, str | loguru.FormatFunction, Optional[int]) -> None LoguruIntegration.level = level LoguruIntegration.event_level = event_level LoguruIntegration.breadcrumb_format = breadcrumb_format LoguruIntegration.event_format = event_format + LoguruIntegration.sentry_logs_level = sentry_logs_level @staticmethod def setup_once(): @@ -83,8 +103,23 @@ def setup_once(): format=LoguruIntegration.event_format, ) + if LoguruIntegration.sentry_logs_level is not None: + logger.add( + loguru_sentry_logs_handler, + level=LoguruIntegration.sentry_logs_level, + ) + class _LoguruBaseHandler(_BaseHandler): + def __init__(self, *args, **kwargs): + # type: (*Any, **Any) -> None + if kwargs.get("level"): + kwargs["level"] = SENTRY_LEVEL_FROM_LOGURU_LEVEL.get( + kwargs.get("level", ""), DEFAULT_LEVEL + ) + + super().__init__(*args, **kwargs) + def _logging_to_event_level(self, record): # type: (LogRecord) -> str try: @@ -98,24 +133,72 @@ def _logging_to_event_level(self, record): class LoguruEventHandler(_LoguruBaseHandler, EventHandler): """Modified version of :class:`sentry_sdk.integrations.logging.EventHandler` to use loguru's level names.""" - def __init__(self, *args, **kwargs): - # type: (*Any, **Any) -> None - if kwargs.get("level"): - kwargs["level"] = SENTRY_LEVEL_FROM_LOGURU_LEVEL.get( - kwargs.get("level", ""), DEFAULT_LEVEL - ) - - super().__init__(*args, **kwargs) + pass class LoguruBreadcrumbHandler(_LoguruBaseHandler, BreadcrumbHandler): """Modified version of :class:`sentry_sdk.integrations.logging.BreadcrumbHandler` to use loguru's level names.""" - def __init__(self, *args, **kwargs): - # type: (*Any, **Any) -> None - if kwargs.get("level"): - kwargs["level"] = SENTRY_LEVEL_FROM_LOGURU_LEVEL.get( - kwargs.get("level", ""), DEFAULT_LEVEL - ) + pass - super().__init__(*args, **kwargs) + +def loguru_sentry_logs_handler(message): + # type: (Message) -> None + # This is intentionally a callable sink instead of a standard logging handler + # since otherwise we wouldn't get direct access to message.record + client = sentry_sdk.get_client() + + if not client.is_active(): + return + + if not client.options["_experiments"].get("enable_logs", False): + return + + record = message.record + + if ( + LoguruIntegration.sentry_logs_level is None + or record["level"].no < LoguruIntegration.sentry_logs_level + ): + return + + otel_severity_number, otel_severity_text = _log_level_to_otel( + record["level"].no, SEVERITY_TO_OTEL_SEVERITY + ) + + attrs = {"sentry.origin": "auto.logger.loguru"} # type: dict[str, Any] + + project_root = client.options["project_root"] + if record.get("file"): + if project_root is not None and record["file"].path.startswith(project_root): + attrs["code.file.path"] = record["file"].path[len(project_root) + 1 :] + else: + attrs["code.file.path"] = record["file"].path + + if record.get("line") is not None: + attrs["code.line.number"] = record["line"] + + if record.get("function"): + attrs["code.function.name"] = record["function"] + + if record.get("thread"): + attrs["thread.name"] = record["thread"].name + attrs["thread.id"] = record["thread"].id + + if record.get("process"): + attrs["process.pid"] = record["process"].id + attrs["process.executable.name"] = record["process"].name + + if record.get("name"): + attrs["logger.name"] = record["name"] + + client._capture_experimental_log( + { + "severity_text": otel_severity_text, + "severity_number": otel_severity_number, + "body": record["message"], + "attributes": attrs, + "time_unix_nano": int(record["time"].timestamp() * 1e9), + "trace_id": None, + } + ) diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index 2f5e859533..c18cf91ff2 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -6,6 +6,17 @@ from sentry_sdk import get_client from sentry_sdk.utils import safe_repr +OTEL_RANGES = [ + # ((severity level range), severity text) + # https://opentelemetry.io/docs/specs/otel/logs/data-model + ((1, 4), "trace"), + ((5, 8), "debug"), + ((9, 12), "info"), + ((13, 16), "warn"), + ((17, 20), "error"), + ((21, 24), "fatal"), +] + def _capture_log(severity_text, severity_number, template, **kwargs): # type: (str, int, str, **Any) -> None @@ -52,3 +63,21 @@ def _capture_log(severity_text, severity_number, template, **kwargs): warning = functools.partial(_capture_log, "warn", 13) error = functools.partial(_capture_log, "error", 17) fatal = functools.partial(_capture_log, "fatal", 21) + + +def _otel_severity_text(otel_severity_number): + # type: (int) -> str + for (lower, upper), severity in OTEL_RANGES: + if lower <= otel_severity_number <= upper: + return severity + + return "default" + + +def _log_level_to_otel(level, mapping): + # type: (int, dict[Any, int]) -> tuple[int, str] + for py_level, otel_severity_number in sorted(mapping.items(), reverse=True): + if level >= py_level: + return otel_severity_number, _otel_severity_text(otel_severity_number) + + return 0, "default" diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 557f3fc1b1..237373fc91 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -3,7 +3,10 @@ import pytest +from sentry_sdk import get_client +from sentry_sdk.consts import VERSION from sentry_sdk.integrations.logging import LoggingIntegration, ignore_logger +from tests.test_logs import envelopes_to_logs other_logger = logging.getLogger("testfoo") logger = logging.getLogger(__name__) @@ -295,3 +298,197 @@ def test_logging_dictionary_args(sentry_init, capture_events): == "the value of foo is bar, and the value of bar is baz" ) assert event["logentry"]["params"] == {"foo": "bar", "bar": "baz"} + + +def test_sentry_logs_warning(sentry_init, capture_envelopes): + """ + The python logger module should create 'warn' sentry logs if the flag is on. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("this is %s a template %s", "1", "2") + + get_client().flush() + logs = envelopes_to_logs(envelopes) + attrs = logs[0]["attributes"] + assert attrs["sentry.message.template"] == "this is %s a template %s" + assert "code.file.path" in attrs + assert "code.line.number" in attrs + assert attrs["logger.name"] == "test-logger" + assert attrs["sentry.environment"] == "production" + assert attrs["sentry.message.parameter.0"] == "1" + assert attrs["sentry.message.parameter.1"] == "2" + assert attrs["sentry.origin"] == "auto.logger.log" + assert logs[0]["severity_number"] == 13 + assert logs[0]["severity_text"] == "warn" + + +def test_sentry_logs_debug(sentry_init, capture_envelopes): + """ + The python logger module should not create 'debug' sentry logs if the flag is on by default + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.debug("this is %s a template %s", "1", "2") + get_client().flush() + + assert len(envelopes) == 0 + + +def test_no_log_infinite_loop(sentry_init, capture_envelopes): + """ + If 'debug' mode is true, and you set a low log level in the logging integration, there should be no infinite loops. + """ + sentry_init( + _experiments={"enable_logs": True}, + integrations=[LoggingIntegration(sentry_logs_level=logging.DEBUG)], + debug=True, + ) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.debug("this is %s a template %s", "1", "2") + get_client().flush() + + assert len(envelopes) == 1 + + +def test_logging_errors(sentry_init, capture_envelopes): + """ + The python logger module should be able to log errors without erroring + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.error(Exception("test exc 1")) + python_logger.error("error is %s", Exception("test exc 2")) + get_client().flush() + + error_event_1 = envelopes[0].items[0].payload.json + assert error_event_1["level"] == "error" + error_event_2 = envelopes[1].items[0].payload.json + assert error_event_2["level"] == "error" + + logs = envelopes_to_logs(envelopes) + assert logs[0]["severity_text"] == "error" + assert "sentry.message.template" not in logs[0]["attributes"] + assert "sentry.message.parameter.0" not in logs[0]["attributes"] + assert "code.line.number" in logs[0]["attributes"] + + assert logs[1]["severity_text"] == "error" + assert logs[1]["attributes"]["sentry.message.template"] == "error is %s" + assert logs[1]["attributes"]["sentry.message.parameter.0"] in ( + "Exception('test exc 2')", + "Exception('test exc 2',)", # py3.6 + ) + assert "code.line.number" in logs[1]["attributes"] + + assert len(logs) == 2 + + +def test_log_strips_project_root(sentry_init, capture_envelopes): + """ + The python logger should strip project roots from the log record path + """ + sentry_init( + _experiments={"enable_logs": True}, + project_root="/custom/test", + ) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.handle( + logging.LogRecord( + name="test-logger", + level=logging.WARN, + pathname="/custom/test/blah/path.py", + lineno=123, + msg="This is a test log with a custom pathname", + args=(), + exc_info=None, + ) + ) + get_client().flush() + + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + attrs = logs[0]["attributes"] + assert attrs["code.file.path"] == "blah/path.py" + + +def test_logger_with_all_attributes(sentry_init, capture_envelopes): + """ + The python logger should be able to log all attributes, including extra data. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning( + "log #%d", + 1, + extra={"foo": "bar", "numeric": 42, "more_complex": {"nested": "data"}}, + ) + get_client().flush() + + logs = envelopes_to_logs(envelopes) + + attributes = logs[0]["attributes"] + + assert "process.pid" in attributes + assert isinstance(attributes["process.pid"], int) + del attributes["process.pid"] + + assert "sentry.release" in attributes + assert isinstance(attributes["sentry.release"], str) + del attributes["sentry.release"] + + assert "server.address" in attributes + assert isinstance(attributes["server.address"], str) + del attributes["server.address"] + + assert "thread.id" in attributes + assert isinstance(attributes["thread.id"], int) + del attributes["thread.id"] + + assert "code.file.path" in attributes + assert isinstance(attributes["code.file.path"], str) + del attributes["code.file.path"] + + assert "code.function.name" in attributes + assert isinstance(attributes["code.function.name"], str) + del attributes["code.function.name"] + + assert "code.line.number" in attributes + assert isinstance(attributes["code.line.number"], int) + del attributes["code.line.number"] + + assert "process.executable.name" in attributes + assert isinstance(attributes["process.executable.name"], str) + del attributes["process.executable.name"] + + assert "thread.name" in attributes + assert isinstance(attributes["thread.name"], str) + del attributes["thread.name"] + + assert attributes.pop("sentry.sdk.name").startswith("sentry.python") + + # Assert on the remaining non-dynamic attributes. + assert attributes == { + "foo": "bar", + "numeric": 42, + "more_complex": "{'nested': 'data'}", + "logger.name": "test-logger", + "sentry.origin": "auto.logger.log", + "sentry.message.template": "log #%d", + "sentry.message.parameter.0": 1, + "sentry.environment": "production", + "sentry.sdk.version": VERSION, + "sentry.severity_number": 13, + "sentry.severity_text": "warn", + } diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 6be09b86dc..20d3230b49 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -1,8 +1,13 @@ +from unittest.mock import MagicMock, patch + import pytest from loguru import logger +from loguru._recattrs import RecordFile, RecordLevel import sentry_sdk +from sentry_sdk.consts import VERSION from sentry_sdk.integrations.loguru import LoguruIntegration, LoggingLevels +from tests.test_logs import envelopes_to_logs logger.remove(0) # don't print to console @@ -54,7 +59,7 @@ def test_just_log( formatted_message = ( " | " + "{:9}".format(level.name.upper()) - + "| tests.integrations.loguru.test_loguru:test_just_log:52 - test" + + "| tests.integrations.loguru.test_loguru:test_just_log:57 - test" ) if not created_event: @@ -127,3 +132,337 @@ def test_event_format(sentry_init, capture_events, uninstall_integration, reques (event,) = events assert event["logentry"]["message"] == formatted_message + + +def test_sentry_logs_warning( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + logger.warning("this is {} a {}", "just", "template") + + sentry_sdk.get_client().flush() + logs = envelopes_to_logs(envelopes) + + attrs = logs[0]["attributes"] + assert "code.file.path" in attrs + assert "code.line.number" in attrs + assert attrs["logger.name"] == "tests.integrations.loguru.test_loguru" + assert attrs["sentry.environment"] == "production" + assert attrs["sentry.origin"] == "auto.logger.loguru" + assert logs[0]["severity_number"] == 13 + assert logs[0]["severity_text"] == "warn" + + +def test_sentry_logs_debug( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + logger.debug("this is %s a template %s", "1", "2") + sentry_sdk.get_client().flush() + + assert len(envelopes) == 0 + + +def test_sentry_log_levels( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + integrations=[LoguruIntegration(sentry_logs_level=LoggingLevels.SUCCESS)], + _experiments={"enable_logs": True}, + ) + envelopes = capture_envelopes() + + logger.trace("this is a log") + logger.debug("this is a log") + logger.info("this is a log") + logger.success("this is a log") + logger.warning("this is a log") + logger.error("this is a log") + logger.critical("this is a log") + + sentry_sdk.get_client().flush() + logs = envelopes_to_logs(envelopes) + assert len(logs) == 4 + + assert logs[0]["severity_number"] == 11 + assert logs[0]["severity_text"] == "info" + assert logs[1]["severity_number"] == 13 + assert logs[1]["severity_text"] == "warn" + assert logs[2]["severity_number"] == 17 + assert logs[2]["severity_text"] == "error" + assert logs[3]["severity_number"] == 21 + assert logs[3]["severity_text"] == "fatal" + + +def test_disable_loguru_logs( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + integrations=[LoguruIntegration(sentry_logs_level=None)], + _experiments={"enable_logs": True}, + ) + envelopes = capture_envelopes() + + logger.trace("this is a log") + logger.debug("this is a log") + logger.info("this is a log") + logger.success("this is a log") + logger.warning("this is a log") + logger.error("this is a log") + logger.critical("this is a log") + + sentry_sdk.get_client().flush() + logs = envelopes_to_logs(envelopes) + assert len(logs) == 0 + + +def test_disable_sentry_logs( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + _experiments={"enable_logs": False}, + ) + envelopes = capture_envelopes() + + logger.trace("this is a log") + logger.debug("this is a log") + logger.info("this is a log") + logger.success("this is a log") + logger.warning("this is a log") + logger.error("this is a log") + logger.critical("this is a log") + + sentry_sdk.get_client().flush() + logs = envelopes_to_logs(envelopes) + assert len(logs) == 0 + + +def test_no_log_infinite_loop( + sentry_init, capture_envelopes, uninstall_integration, request +): + """ + In debug mode, there should be no infinite loops even when a low log level is set. + """ + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + _experiments={"enable_logs": True}, + integrations=[LoguruIntegration(sentry_logs_level=LoggingLevels.DEBUG)], + debug=True, + ) + envelopes = capture_envelopes() + + logger.debug("this is %s a template %s", "1", "2") + sentry_sdk.get_client().flush() + + assert len(envelopes) == 1 + + +def test_logging_errors(sentry_init, capture_envelopes, uninstall_integration, request): + """We're able to log errors without erroring.""" + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + logger.error(Exception("test exc 1")) + logger.error("error is %s", Exception("test exc 2")) + sentry_sdk.get_client().flush() + + error_event_1 = envelopes[0].items[0].payload.json + assert error_event_1["level"] == "error" + error_event_2 = envelopes[1].items[0].payload.json + assert error_event_2["level"] == "error" + + logs = envelopes_to_logs(envelopes) + assert logs[0]["severity_text"] == "error" + assert "code.line.number" in logs[0]["attributes"] + + assert logs[1]["severity_text"] == "error" + assert "code.line.number" in logs[1]["attributes"] + + assert len(logs) == 2 + + +def test_log_strips_project_root( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + _experiments={"enable_logs": True}, + project_root="/custom/test", + ) + envelopes = capture_envelopes() + + class FakeMessage: + def __init__(self, *args, **kwargs): + pass + + @property + def record(self): + return { + "elapsed": MagicMock(), + "exception": None, + "file": RecordFile(name="app.py", path="/custom/test/blah/path.py"), + "function": "", + "level": RecordLevel(name="ERROR", no=20, icon=""), + "line": 35, + "message": "some message", + "module": "app", + "name": "__main__", + "process": MagicMock(), + "thread": MagicMock(), + "time": MagicMock(), + "extra": MagicMock(), + } + + @record.setter + def record(self, val): + pass + + with patch("loguru._handler.Message", FakeMessage): + logger.error("some message") + + sentry_sdk.get_client().flush() + + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + attrs = logs[0]["attributes"] + assert attrs["code.file.path"] == "blah/path.py" + + +def test_log_keeps_full_path_if_not_in_project_root( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init( + _experiments={"enable_logs": True}, + project_root="/custom/test", + ) + envelopes = capture_envelopes() + + class FakeMessage: + def __init__(self, *args, **kwargs): + pass + + @property + def record(self): + return { + "elapsed": MagicMock(), + "exception": None, + "file": RecordFile(name="app.py", path="/blah/path.py"), + "function": "", + "level": RecordLevel(name="ERROR", no=20, icon=""), + "line": 35, + "message": "some message", + "module": "app", + "name": "__main__", + "process": MagicMock(), + "thread": MagicMock(), + "time": MagicMock(), + "extra": MagicMock(), + } + + @record.setter + def record(self, val): + pass + + with patch("loguru._handler.Message", FakeMessage): + logger.error("some message") + + sentry_sdk.get_client().flush() + + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + attrs = logs[0]["attributes"] + assert attrs["code.file.path"] == "/blah/path.py" + + +def test_logger_with_all_attributes( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + logger.warning("log #{}", 1) + sentry_sdk.get_client().flush() + + logs = envelopes_to_logs(envelopes) + + attributes = logs[0]["attributes"] + + assert "process.pid" in attributes + assert isinstance(attributes["process.pid"], int) + del attributes["process.pid"] + + assert "sentry.release" in attributes + assert isinstance(attributes["sentry.release"], str) + del attributes["sentry.release"] + + assert "server.address" in attributes + assert isinstance(attributes["server.address"], str) + del attributes["server.address"] + + assert "thread.id" in attributes + assert isinstance(attributes["thread.id"], int) + del attributes["thread.id"] + + assert "code.file.path" in attributes + assert isinstance(attributes["code.file.path"], str) + del attributes["code.file.path"] + + assert "code.function.name" in attributes + assert isinstance(attributes["code.function.name"], str) + del attributes["code.function.name"] + + assert "code.line.number" in attributes + assert isinstance(attributes["code.line.number"], int) + del attributes["code.line.number"] + + assert "process.executable.name" in attributes + assert isinstance(attributes["process.executable.name"], str) + del attributes["process.executable.name"] + + assert "thread.name" in attributes + assert isinstance(attributes["thread.name"], str) + del attributes["thread.name"] + + assert attributes.pop("sentry.sdk.name").startswith("sentry.python") + + # Assert on the remaining non-dynamic attributes. + assert attributes == { + "logger.name": "tests.integrations.loguru.test_loguru", + "sentry.origin": "auto.logger.loguru", + "sentry.environment": "production", + "sentry.sdk.version": VERSION, + "sentry.severity_number": 13, + "sentry.severity_text": "warn", + } diff --git a/tests/test_logs.py b/tests/test_logs.py index dcbaba3c4f..a2f412dcb0 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -9,7 +9,6 @@ import sentry_sdk.logger from sentry_sdk import get_client from sentry_sdk.envelope import Envelope -from sentry_sdk.integrations.logging import LoggingIntegration from sentry_sdk.types import Log from sentry_sdk.consts import SPANDATA, VERSION @@ -268,203 +267,6 @@ def test_logs_tied_to_spans(sentry_init, capture_envelopes): assert logs[0]["attributes"]["sentry.trace.parent_span_id"] == span.span_id -@minimum_python_37 -def test_logger_integration_warning(sentry_init, capture_envelopes): - """ - The python logger module should create 'warn' sentry logs if the flag is on. - """ - sentry_init(_experiments={"enable_logs": True}) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.warning("this is %s a template %s", "1", "2") - - get_client().flush() - logs = envelopes_to_logs(envelopes) - attrs = logs[0]["attributes"] - assert attrs["sentry.message.template"] == "this is %s a template %s" - assert "code.file.path" in attrs - assert "code.line.number" in attrs - assert attrs["logger.name"] == "test-logger" - assert attrs["sentry.environment"] == "production" - assert attrs["sentry.message.parameter.0"] == "1" - assert attrs["sentry.message.parameter.1"] == "2" - assert attrs["sentry.origin"] == "auto.logger.log" - assert logs[0]["severity_number"] == 13 - assert logs[0]["severity_text"] == "warn" - - -@minimum_python_37 -def test_logger_integration_debug(sentry_init, capture_envelopes): - """ - The python logger module should not create 'debug' sentry logs if the flag is on by default - """ - sentry_init(_experiments={"enable_logs": True}) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.debug("this is %s a template %s", "1", "2") - get_client().flush() - - assert len(envelopes) == 0 - - -@minimum_python_37 -def test_no_log_infinite_loop(sentry_init, capture_envelopes): - """ - If 'debug' mode is true, and you set a low log level in the logging integration, there should be no infinite loops. - """ - sentry_init( - _experiments={"enable_logs": True}, - integrations=[LoggingIntegration(sentry_logs_level=logging.DEBUG)], - debug=True, - ) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.debug("this is %s a template %s", "1", "2") - get_client().flush() - - assert len(envelopes) == 1 - - -@minimum_python_37 -def test_logging_errors(sentry_init, capture_envelopes): - """ - The python logger module should be able to log errors without erroring - """ - sentry_init(_experiments={"enable_logs": True}) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.error(Exception("test exc 1")) - python_logger.error("error is %s", Exception("test exc 2")) - get_client().flush() - - error_event_1 = envelopes[0].items[0].payload.json - assert error_event_1["level"] == "error" - error_event_2 = envelopes[1].items[0].payload.json - assert error_event_2["level"] == "error" - - logs = envelopes_to_logs(envelopes) - assert logs[0]["severity_text"] == "error" - assert "sentry.message.template" not in logs[0]["attributes"] - assert "sentry.message.parameter.0" not in logs[0]["attributes"] - assert "code.line.number" in logs[0]["attributes"] - - assert logs[1]["severity_text"] == "error" - assert logs[1]["attributes"]["sentry.message.template"] == "error is %s" - assert ( - logs[1]["attributes"]["sentry.message.parameter.0"] == "Exception('test exc 2')" - ) - assert "code.line.number" in logs[1]["attributes"] - - assert len(logs) == 2 - - -def test_log_strips_project_root(sentry_init, capture_envelopes): - """ - The python logger should strip project roots from the log record path - """ - sentry_init( - _experiments={"enable_logs": True}, - project_root="/custom/test", - ) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.handle( - logging.LogRecord( - name="test-logger", - level=logging.WARN, - pathname="/custom/test/blah/path.py", - lineno=123, - msg="This is a test log with a custom pathname", - args=(), - exc_info=None, - ) - ) - get_client().flush() - - logs = envelopes_to_logs(envelopes) - assert len(logs) == 1 - attrs = logs[0]["attributes"] - assert attrs["code.file.path"] == "blah/path.py" - - -def test_logger_with_all_attributes(sentry_init, capture_envelopes): - """ - The python logger should be able to log all attributes, including extra data. - """ - sentry_init(_experiments={"enable_logs": True}) - envelopes = capture_envelopes() - - python_logger = logging.Logger("test-logger") - python_logger.warning( - "log #%d", - 1, - extra={"foo": "bar", "numeric": 42, "more_complex": {"nested": "data"}}, - ) - get_client().flush() - - logs = envelopes_to_logs(envelopes) - - attributes = logs[0]["attributes"] - - assert "process.pid" in attributes - assert isinstance(attributes["process.pid"], int) - del attributes["process.pid"] - - assert "sentry.release" in attributes - assert isinstance(attributes["sentry.release"], str) - del attributes["sentry.release"] - - assert "server.address" in attributes - assert isinstance(attributes["server.address"], str) - del attributes["server.address"] - - assert "thread.id" in attributes - assert isinstance(attributes["thread.id"], int) - del attributes["thread.id"] - - assert "code.file.path" in attributes - assert isinstance(attributes["code.file.path"], str) - del attributes["code.file.path"] - - assert "code.function.name" in attributes - assert isinstance(attributes["code.function.name"], str) - del attributes["code.function.name"] - - assert "code.line.number" in attributes - assert isinstance(attributes["code.line.number"], int) - del attributes["code.line.number"] - - assert "process.executable.name" in attributes - assert isinstance(attributes["process.executable.name"], str) - del attributes["process.executable.name"] - - assert "thread.name" in attributes - assert isinstance(attributes["thread.name"], str) - del attributes["thread.name"] - - assert attributes.pop("sentry.sdk.name").startswith("sentry.python") - - # Assert on the remaining non-dynamic attributes. - assert attributes == { - "foo": "bar", - "numeric": 42, - "more_complex": "{'nested': 'data'}", - "logger.name": "test-logger", - "sentry.origin": "auto.logger.log", - "sentry.message.template": "log #%d", - "sentry.message.parameter.0": 1, - "sentry.environment": "production", - "sentry.sdk.version": VERSION, - "sentry.severity_number": 13, - "sentry.severity_text": "warn", - } - - def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): """ If you log >100 logs, it should automatically trigger a flush. From 949c4e817b87894b327784ed07b5fb62a6c91fe9 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 12 Jun 2025 09:46:34 +0000 Subject: [PATCH 316/591] release: 2.30.0 --- CHANGELOG.md | 21 +++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 24 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8dd7763121..5cc6dee907 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,26 @@ # Changelog +## 2.30.0 + +### Various fixes & improvements + +- feat(loguru): Sentry logs for Loguru (#4445) by @sentrivana +- fix(logs): Don't gate user behind `send_default_pii` (#4453) by @AbhiPrasad +- fix(logging): Strip log `record.name` for more robust matching (#4411) by @romaingd-spi +- Migrate to modern threading interface (#4452) by @emmanuel-ferdman +- ref: Remove `_capture_experimental_log` `scope` parameter (#4424) by @szokeasaurusrex +- feat(logs): Add user attributes to logs (#4423) by @szokeasaurusrex +- tests: Regenerate tox.ini & fix CI (#4435) by @sentrivana +- fix: fix ARQ integration error (#4427) (#4428) by @ninoseki +- Fix CI, adapt to new redis-py release (#4431) by @sentrivana +- fix(grpc): Fix AttributeError when instrumenting with OTel (#4405) by @sentrivana +- fix(redis): Use `command_queue` instead of `command_stack` if available (#4404) by @sentrivana +- tests: Regenerate toxgen (#4403) by @sentrivana +- fix: Handle invalid `SENTRY_DEBUG` values properly (#4400) by @szokeasaurusrex +- build(deps): bump codecov/codecov-action from 5.4.2 to 5.4.3 (#4397) by @dependabot +- Increase test coverage (#4393) by @mgaligniana +- tests(logs): avoid failures when running with integrations enabled (#4388) by @rominf + ## 2.29.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index e639185528..4e12abf550 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.29.1" +release = "2.30.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 241d197699..34ae5bdfd8 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1026,4 +1026,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.29.1" +VERSION = "2.30.0" diff --git a/setup.py b/setup.py index 92634bd2ef..ecb5dfa994 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.29.1", + version="2.30.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 719efd5c65d7df65f21296f4dc121ee7c8f5cd76 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 12 Jun 2025 11:51:24 +0200 Subject: [PATCH 317/591] Update CHANGELOG.md --- CHANGELOG.md | 28 +++++++++++++++++++++++----- 1 file changed, 23 insertions(+), 5 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5cc6dee907..ddeed9d687 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,22 +4,40 @@ ### Various fixes & improvements -- feat(loguru): Sentry logs for Loguru (#4445) by @sentrivana +- **New beta feature:** Sentry logs for Loguru (#4445) by @sentrivana + + We can now capture Loguru logs and send them to Sentry. + +```python +import sentry_sdk +from sentry_sdk.integrations.loguru import LoguruIntegration + +# Setup Sentry SDK to send Loguru log messages with a level of "error" or higher to Sentry +sentry_sdk.init( + _experiments={ + "enable_logs": True, + }, + integrations=[ + LoguruIntegration(sentry_logs_level=logging.ERROR), + ] +) +``` + - fix(logs): Don't gate user behind `send_default_pii` (#4453) by @AbhiPrasad - fix(logging): Strip log `record.name` for more robust matching (#4411) by @romaingd-spi - Migrate to modern threading interface (#4452) by @emmanuel-ferdman - ref: Remove `_capture_experimental_log` `scope` parameter (#4424) by @szokeasaurusrex - feat(logs): Add user attributes to logs (#4423) by @szokeasaurusrex -- tests: Regenerate tox.ini & fix CI (#4435) by @sentrivana - fix: fix ARQ integration error (#4427) (#4428) by @ninoseki -- Fix CI, adapt to new redis-py release (#4431) by @sentrivana - fix(grpc): Fix AttributeError when instrumenting with OTel (#4405) by @sentrivana - fix(redis): Use `command_queue` instead of `command_stack` if available (#4404) by @sentrivana -- tests: Regenerate toxgen (#4403) by @sentrivana - fix: Handle invalid `SENTRY_DEBUG` values properly (#4400) by @szokeasaurusrex -- build(deps): bump codecov/codecov-action from 5.4.2 to 5.4.3 (#4397) by @dependabot - Increase test coverage (#4393) by @mgaligniana - tests(logs): avoid failures when running with integrations enabled (#4388) by @rominf +- Fix CI, adapt to new redis-py release (#4431) by @sentrivana +- tests: Regenerate toxgen (#4403) by @sentrivana +- tests: Regenerate tox.ini & fix CI (#4435) by @sentrivana +- build(deps): bump codecov/codecov-action from 5.4.2 to 5.4.3 (#4397) by @dependabot ## 2.29.1 From 0e21fa2f0fbec007563f22d10bf6ec4a128faa25 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 13 Jun 2025 12:45:15 +0200 Subject: [PATCH 318/591] tests: Regenerate tox (#4457) (Yeah I'll automate this at some point) --- tox.ini | 78 ++++++++++++++++++++++++++++----------------------------- 1 file changed, 38 insertions(+), 40 deletions(-) diff --git a/tox.ini b/tox.ini index 2fe4977a1c..32e16dac3d 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-06-03T13:53:53.259798+00:00 +# Last generated: 2025-06-11T12:20:52.494394+00:00 [tox] requires = @@ -136,9 +136,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.28.1 - {py3.8,py3.11,py3.12}-anthropic-v0.40.0 - {py3.8,py3.11,py3.12}-anthropic-v0.52.2 + {py3.8,py3.11,py3.12}-anthropic-v0.29.2 + {py3.8,py3.11,py3.12}-anthropic-v0.42.0 + {py3.8,py3.11,py3.12}-anthropic-v0.54.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 @@ -148,7 +148,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.4 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 # ~~~ DBs ~~~ @@ -180,7 +180,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.58.0 + {py3.7,py3.12,py3.13}-statsig-v0.58.1 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -201,17 +201,16 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.229.2 - {py3.8,py3.12,py3.13}-strawberry-v0.249.0 - {py3.9,py3.12,py3.13}-strawberry-v0.270.5 + {py3.8,py3.11,py3.12}-strawberry-v0.230.0 + {py3.8,py3.12,py3.13}-strawberry-v0.251.0 + {py3.9,py3.12,py3.13}-strawberry-v0.273.0 # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.46.5 {py3.7,py3.11,py3.12}-grpc-v1.60.2 - {py3.9,py3.12,py3.13}-grpc-v1.72.1 - {py3.9,py3.12,py3.13}-grpc-v1.73.0rc1 + {py3.9,py3.12,py3.13}-grpc-v1.73.0 # ~~~ Tasks ~~~ @@ -238,9 +237,9 @@ envlist = {py3.6,py3.7}-django-v1.11.29 {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 - {py3.8,py3.11,py3.12}-django-v4.2.21 + {py3.8,py3.11,py3.12}-django-v4.2.23 {py3.10,py3.11,py3.12}-django-v5.0.14 - {py3.10,py3.12,py3.13}-django-v5.2.1 + {py3.10,py3.12,py3.13}-django-v5.2.3 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -262,7 +261,7 @@ envlist = {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.12.7 + {py3.9,py3.12,py3.13}-aiohttp-v3.12.12 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.3 @@ -299,8 +298,8 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.31 - {py3.9,py3.12,py3.13}-trytond-v7.6.1 + {py3.8,py3.11,py3.12}-trytond-v7.0.32 + {py3.9,py3.12,py3.13}-trytond-v7.6.2 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.0 @@ -502,13 +501,13 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.28.1: anthropic==0.28.1 - anthropic-v0.40.0: anthropic==0.40.0 - anthropic-v0.52.2: anthropic==0.52.2 + anthropic-v0.29.2: anthropic==0.29.2 + anthropic-v0.42.0: anthropic==0.42.0 + anthropic-v0.54.0: anthropic==0.54.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.28.1: httpx<0.28.0 - anthropic-v0.40.0: httpx<0.28.0 + anthropic-v0.29.2: httpx<0.28.0 + anthropic-v0.42.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.8.1: cohere==5.8.1 @@ -518,7 +517,7 @@ deps = huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.25.2: huggingface_hub==0.25.2 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.32.4: huggingface_hub==0.32.4 + huggingface_hub-v0.32.6: huggingface_hub==0.32.6 # ~~~ DBs ~~~ @@ -551,7 +550,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 - statsig-v0.58.0: statsig==0.58.0 + statsig-v0.58.1: statsig==0.58.1 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -581,21 +580,20 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.229.2: strawberry-graphql[fastapi,flask]==0.229.2 - strawberry-v0.249.0: strawberry-graphql[fastapi,flask]==0.249.0 - strawberry-v0.270.5: strawberry-graphql[fastapi,flask]==0.270.5 + strawberry-v0.230.0: strawberry-graphql[fastapi,flask]==0.230.0 + strawberry-v0.251.0: strawberry-graphql[fastapi,flask]==0.251.0 + strawberry-v0.273.0: strawberry-graphql[fastapi,flask]==0.273.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.229.2: pydantic<2.11 - strawberry-v0.249.0: pydantic<2.11 + strawberry-v0.230.0: pydantic<2.11 + strawberry-v0.251.0: pydantic<2.11 # ~~~ Network ~~~ grpc-v1.32.0: grpcio==1.32.0 grpc-v1.46.5: grpcio==1.46.5 grpc-v1.60.2: grpcio==1.60.2 - grpc-v1.72.1: grpcio==1.72.1 - grpc-v1.73.0rc1: grpcio==1.73.0rc1 + grpc-v1.73.0: grpcio==1.73.0 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -629,23 +627,23 @@ deps = django-v1.11.29: django==1.11.29 django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 - django-v4.2.21: django==4.2.21 + django-v4.2.23: django==4.2.23 django-v5.0.14: django==5.0.14 - django-v5.2.1: django==5.2.1 + django-v5.2.3: django==5.2.3 django: psycopg2-binary django: djangorestframework django: pytest-django django: Werkzeug django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] - django-v4.2.21: channels[daphne] + django-v4.2.23: channels[daphne] django-v5.0.14: channels[daphne] - django-v5.2.1: channels[daphne] + django-v5.2.3: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio - django-v4.2.21: pytest-asyncio + django-v4.2.23: pytest-asyncio django-v5.0.14: pytest-asyncio - django-v5.2.1: pytest-asyncio + django-v5.2.3: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 @@ -698,10 +696,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.12.7: aiohttp==3.12.7 + aiohttp-v3.12.12: aiohttp==3.12.12 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.12.7: pytest-asyncio + aiohttp-v3.12.12: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.3: bottle==0.13.3 @@ -756,8 +754,8 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.31: trytond==7.0.31 - trytond-v7.6.1: trytond==7.6.1 + trytond-v7.0.32: trytond==7.0.32 + trytond-v7.6.2: trytond==7.6.2 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From f71d223c6cecd4e99565c5e8e919ce9cf81707ff Mon Sep 17 00:00:00 2001 From: Abhijeet Prasad Date: Mon, 16 Jun 2025 10:17:57 -0400 Subject: [PATCH 319/591] feat(logs): Add support for dict args (#4478) resolves https://github.com/getsentry/sentry-python/issues/4477 This PR adds support for dict log arguments and adds (cursor generated) tests accordingly. --- sentry_sdk/integrations/logging.py | 7 ++ tests/integrations/logging/test_logging.py | 79 ++++++++++++++++++++++ 2 files changed, 86 insertions(+) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 62b1e09d64..a50512f622 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -367,6 +367,13 @@ def _capture_log_from_record(self, client, record): if isinstance(arg, (str, float, int, bool)) else safe_repr(arg) ) + elif isinstance(record.args, dict): + for key, value in record.args.items(): + attrs[f"sentry.message.parameter.{key}"] = ( + value + if isinstance(value, (str, float, int, bool)) + else safe_repr(value) + ) if record.lineno: attrs["code.line.number"] = record.lineno if record.pathname: diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 237373fc91..6ef4ae371b 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -492,3 +492,82 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): "sentry.severity_number": 13, "sentry.severity_text": "warn", } + + +def test_sentry_logs_named_parameters(sentry_init, capture_envelopes): + """ + The python logger module should capture named parameters from dictionary arguments in Sentry logs. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.info( + "%(source)s call completed, %(input_tk)i input tk, %(output_tk)i output tk (model %(model)s, cost $%(cost).4f)", + { + "source": "test_source", + "input_tk": 100, + "output_tk": 50, + "model": "gpt-4", + "cost": 0.0234, + }, + ) + + get_client().flush() + logs = envelopes_to_logs(envelopes) + + assert len(logs) == 1 + attrs = logs[0]["attributes"] + + # Check that the template is captured + assert ( + attrs["sentry.message.template"] + == "%(source)s call completed, %(input_tk)i input tk, %(output_tk)i output tk (model %(model)s, cost $%(cost).4f)" + ) + + # Check that dictionary arguments are captured as named parameters + assert attrs["sentry.message.parameter.source"] == "test_source" + assert attrs["sentry.message.parameter.input_tk"] == 100 + assert attrs["sentry.message.parameter.output_tk"] == 50 + assert attrs["sentry.message.parameter.model"] == "gpt-4" + assert attrs["sentry.message.parameter.cost"] == 0.0234 + + # Check other standard attributes + assert attrs["logger.name"] == "test-logger" + assert attrs["sentry.origin"] == "auto.logger.log" + assert logs[0]["severity_number"] == 9 # info level + assert logs[0]["severity_text"] == "info" + + +def test_sentry_logs_named_parameters_complex_values(sentry_init, capture_envelopes): + """ + The python logger module should handle complex values in named parameters using safe_repr. + """ + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + complex_object = {"nested": {"data": [1, 2, 3]}, "tuple": (4, 5, 6)} + python_logger.warning( + "Processing %(simple)s with %(complex)s data", + { + "simple": "simple_value", + "complex": complex_object, + }, + ) + + get_client().flush() + logs = envelopes_to_logs(envelopes) + + assert len(logs) == 1 + attrs = logs[0]["attributes"] + + # Check that simple values are kept as-is + assert attrs["sentry.message.parameter.simple"] == "simple_value" + + # Check that complex values are converted using safe_repr + assert "sentry.message.parameter.complex" in attrs + complex_param = attrs["sentry.message.parameter.complex"] + assert isinstance(complex_param, str) + assert "nested" in complex_param + assert "data" in complex_param From fedcb07dcf60b318ec62e5cf833b3c16b70ba707 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 17 Jun 2025 10:40:33 +0200 Subject: [PATCH 320/591] tests: Upper bound on fakeredis on old Python versions (#4482) --- scripts/populate_tox/tox.jinja | 2 +- tox.ini | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 3f3691147e..45f56e2f1f 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -295,7 +295,7 @@ deps = # Redis redis: fakeredis!=1.7.4 redis: pytest<8.0.0 - {py3.6,py3.7}-redis: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 + {py3.6,py3.7,py3.8}-redis: fakeredis<2.26.0 {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 diff --git a/tox.ini b/tox.ini index 32e16dac3d..3ba62e1a5c 100644 --- a/tox.ini +++ b/tox.ini @@ -456,7 +456,7 @@ deps = # Redis redis: fakeredis!=1.7.4 redis: pytest<8.0.0 - {py3.6,py3.7}-redis: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 + {py3.6,py3.7,py3.8}-redis: fakeredis<2.26.0 {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 From 449b2fa49417d6bf87ee2d245c2a55cdd26d9fe1 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 17 Jun 2025 12:26:29 +0200 Subject: [PATCH 321/591] fix(scope): Handle token reset `LookupError`s gracefully (#4481) We're surfacing internal SDK errors to users in https://github.com/getsentry/sentry-python/issues/4410. --- sentry_sdk/scope.py | 36 +++++++++++++++++++------ tests/test_scope.py | 64 +++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 92 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index f346569255..73bf43573e 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -1673,8 +1673,11 @@ def new_scope(): yield new_scope finally: - # restore original scope - _current_scope.reset(token) + try: + # restore original scope + _current_scope.reset(token) + except LookupError: + capture_internal_exception(sys.exc_info()) @contextmanager @@ -1708,8 +1711,11 @@ def use_scope(scope): yield scope finally: - # restore original scope - _current_scope.reset(token) + try: + # restore original scope + _current_scope.reset(token) + except LookupError: + capture_internal_exception(sys.exc_info()) @contextmanager @@ -1750,8 +1756,15 @@ def isolation_scope(): finally: # restore original scopes - _current_scope.reset(current_token) - _isolation_scope.reset(isolation_token) + try: + _current_scope.reset(current_token) + except LookupError: + capture_internal_exception(sys.exc_info()) + + try: + _isolation_scope.reset(isolation_token) + except LookupError: + capture_internal_exception(sys.exc_info()) @contextmanager @@ -1790,8 +1803,15 @@ def use_isolation_scope(isolation_scope): finally: # restore original scopes - _current_scope.reset(current_token) - _isolation_scope.reset(isolation_token) + try: + _current_scope.reset(current_token) + except LookupError: + capture_internal_exception(sys.exc_info()) + + try: + _isolation_scope.reset(isolation_token) + except LookupError: + capture_internal_exception(sys.exc_info()) def should_send_default_pii(): diff --git a/tests/test_scope.py b/tests/test_scope.py index 9b16dc4344..e645d84234 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -905,3 +905,67 @@ def test_last_event_id_cleared(sentry_init): Scope.get_isolation_scope().clear() assert Scope.last_event_id() is None, "last_event_id should be cleared" + + +@pytest.mark.tests_internal_exceptions +@pytest.mark.parametrize( + "scope_manager", + [ + new_scope, + use_scope, + ], +) +def test_handle_lookup_error_on_token_reset_current_scope(scope_manager): + with mock.patch("sentry_sdk.scope.capture_internal_exception") as mock_capture: + with mock.patch("sentry_sdk.scope._current_scope") as mock_token_var: + mock_token_var.reset.side_effect = LookupError() + + mock_token = mock.Mock() + mock_token_var.set.return_value = mock_token + + try: + if scope_manager == use_scope: + with scope_manager(Scope()): + pass + else: + with scope_manager(): + pass + + except Exception: + pytest.fail("Context manager should handle LookupError gracefully") + + mock_capture.assert_called_once() + mock_token_var.reset.assert_called_once_with(mock_token) + + +@pytest.mark.tests_internal_exceptions +@pytest.mark.parametrize( + "scope_manager", + [ + isolation_scope, + use_isolation_scope, + ], +) +def test_handle_lookup_error_on_token_reset_isolation_scope(scope_manager): + with mock.patch("sentry_sdk.scope.capture_internal_exception") as mock_capture: + with mock.patch("sentry_sdk.scope._current_scope") as mock_current_scope: + with mock.patch( + "sentry_sdk.scope._isolation_scope" + ) as mock_isolation_scope: + mock_isolation_scope.reset.side_effect = LookupError() + mock_current_token = mock.Mock() + mock_current_scope.set.return_value = mock_current_token + + try: + if scope_manager == use_isolation_scope: + with scope_manager(Scope()): + pass + else: + with scope_manager(): + pass + + except Exception: + pytest.fail("Context manager should handle LookupError gracefully") + + mock_capture.assert_called_once() + mock_current_scope.reset.assert_called_once_with(mock_current_token) From 6a58e5fb7cc8d6d794a70dc1f00761dce240e2b7 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 17 Jun 2025 15:31:53 +0200 Subject: [PATCH 322/591] tests: Regenerate tox (#4484) Regular tox update. This includes a fix for a new Bottle version which made one of our tests get stuck in a neverending `while` loop. --- tests/integrations/bottle/test_bottle.py | 57 ++++++++++-------------- tox.ini | 48 ++++++++++---------- 2 files changed, 47 insertions(+), 58 deletions(-) diff --git a/tests/integrations/bottle/test_bottle.py b/tests/integrations/bottle/test_bottle.py index 9cc436a229..363a9167e6 100644 --- a/tests/integrations/bottle/test_bottle.py +++ b/tests/integrations/bottle/test_bottle.py @@ -12,8 +12,6 @@ from werkzeug.test import Client from werkzeug.wrappers import Response -import sentry_sdk.integrations.bottle as bottle_sentry - @pytest.fixture(scope="function") def app(sentry_init): @@ -46,7 +44,7 @@ def inner(): def test_has_context(sentry_init, app, capture_events, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) events = capture_events() client = get_client() @@ -77,11 +75,7 @@ def test_transaction_style( capture_events, get_client, ): - sentry_init( - integrations=[ - bottle_sentry.BottleIntegration(transaction_style=transaction_style) - ] - ) + sentry_init(integrations=[BottleIntegration(transaction_style=transaction_style)]) events = capture_events() client = get_client() @@ -100,7 +94,7 @@ def test_transaction_style( def test_errors( sentry_init, capture_exceptions, capture_events, app, debug, catchall, get_client ): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) app.catchall = catchall set_debug(mode=debug) @@ -127,7 +121,7 @@ def index(): def test_large_json_request(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) data = {"foo": {"bar": "a" * 2000}} @@ -157,7 +151,7 @@ def index(): @pytest.mark.parametrize("data", [{}, []], ids=["empty-dict", "empty-list"]) def test_empty_json_request(sentry_init, capture_events, app, data, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) @app.route("/", method="POST") def index(): @@ -180,7 +174,7 @@ def index(): def test_medium_formdata_request(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) data = {"foo": "a" * 2000} @@ -209,9 +203,7 @@ def index(): def test_too_large_raw_request( sentry_init, input_char, capture_events, app, get_client ): - sentry_init( - integrations=[bottle_sentry.BottleIntegration()], max_request_body_size="small" - ) + sentry_init(integrations=[BottleIntegration()], max_request_body_size="small") data = input_char * 2000 @@ -239,9 +231,7 @@ def index(): def test_files_and_form(sentry_init, capture_events, app, get_client): - sentry_init( - integrations=[bottle_sentry.BottleIntegration()], max_request_body_size="always" - ) + sentry_init(integrations=[BottleIntegration()], max_request_body_size="always") data = {"foo": "a" * 2000, "file": (BytesIO(b"hello"), "hello.txt")} @@ -278,9 +268,7 @@ def index(): def test_json_not_truncated_if_max_request_body_size_is_always( sentry_init, capture_events, app, get_client ): - sentry_init( - integrations=[bottle_sentry.BottleIntegration()], max_request_body_size="always" - ) + sentry_init(integrations=[BottleIntegration()], max_request_body_size="always") data = { "key{}".format(i): "value{}".format(i) for i in range(MAX_DATABAG_BREADTH + 10) @@ -309,8 +297,8 @@ def index(): @pytest.mark.parametrize( "integrations", [ - [bottle_sentry.BottleIntegration()], - [bottle_sentry.BottleIntegration(), LoggingIntegration(event_level="ERROR")], + [BottleIntegration()], + [BottleIntegration(), LoggingIntegration(event_level="ERROR")], ], ) def test_errors_not_reported_twice( @@ -324,23 +312,24 @@ def test_errors_not_reported_twice( @app.route("/") def index(): - try: - 1 / 0 - except Exception as e: - logger.exception(e) - raise e + 1 / 0 events = capture_events() client = get_client() + with pytest.raises(ZeroDivisionError): - client.get("/") + try: + client.get("/") + except ZeroDivisionError as e: + logger.exception(e) + raise e assert len(events) == 1 def test_mount(app, capture_exceptions, capture_events, sentry_init, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) app.catchall = False @@ -367,7 +356,7 @@ def crashing_app(environ, start_response): def test_error_in_errorhandler(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) set_debug(False) app.catchall = True @@ -397,7 +386,7 @@ def error_handler(err): def test_bad_request_not_captured(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) events = capture_events() @app.route("/") @@ -412,7 +401,7 @@ def index(): def test_no_exception_on_redirect(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[bottle_sentry.BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()]) events = capture_events() @app.route("/") @@ -436,7 +425,7 @@ def test_span_origin( capture_events, ): sentry_init( - integrations=[bottle_sentry.BottleIntegration()], + integrations=[BottleIntegration()], traces_sample_rate=1.0, ) events = capture_events() diff --git a/tox.ini b/tox.ini index 3ba62e1a5c..c0c50c6029 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-06-11T12:20:52.494394+00:00 +# Last generated: 2025-06-17T08:49:27.078408+00:00 [tox] requires = @@ -146,9 +146,9 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.15.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 - {py3.8,py3.10,py3.11}-huggingface_hub-v0.25.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 + {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.0 # ~~~ DBs ~~~ @@ -157,7 +157,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.13.0 + {py3.9,py3.12,py3.13}-pymongo-v4.13.2 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -180,7 +180,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.58.1 + {py3.7,py3.12,py3.13}-statsig-v0.58.2 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -201,9 +201,9 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.230.0 - {py3.8,py3.12,py3.13}-strawberry-v0.251.0 - {py3.9,py3.12,py3.13}-strawberry-v0.273.0 + {py3.8,py3.11,py3.12}-strawberry-v0.231.1 + {py3.8,py3.12,py3.13}-strawberry-v0.253.1 + {py3.9,py3.12,py3.13}-strawberry-v0.274.0 # ~~~ Network ~~~ @@ -261,10 +261,10 @@ envlist = {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.12.12 + {py3.9,py3.12,py3.13}-aiohttp-v3.12.13 {py3.6,py3.7}-bottle-v0.12.25 - {py3.8,py3.12,py3.13}-bottle-v0.13.3 + {py3.8,py3.12,py3.13}-bottle-v0.13.4 {py3.6}-falcon-v1.4.1 {py3.6,py3.7}-falcon-v2.0.0 @@ -515,9 +515,9 @@ deps = cohere-v5.15.0: cohere==5.15.0 huggingface_hub-v0.22.2: huggingface_hub==0.22.2 - huggingface_hub-v0.25.2: huggingface_hub==0.25.2 - huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.32.6: huggingface_hub==0.32.6 + huggingface_hub-v0.26.5: huggingface_hub==0.26.5 + huggingface_hub-v0.30.2: huggingface_hub==0.30.2 + huggingface_hub-v0.33.0: huggingface_hub==0.33.0 # ~~~ DBs ~~~ @@ -526,7 +526,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.13.0: pymongo==4.13.0 + pymongo-v4.13.2: pymongo==4.13.2 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -550,7 +550,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 - statsig-v0.58.1: statsig==0.58.1 + statsig-v0.58.2: statsig==0.58.2 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -580,13 +580,13 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.230.0: strawberry-graphql[fastapi,flask]==0.230.0 - strawberry-v0.251.0: strawberry-graphql[fastapi,flask]==0.251.0 - strawberry-v0.273.0: strawberry-graphql[fastapi,flask]==0.273.0 + strawberry-v0.231.1: strawberry-graphql[fastapi,flask]==0.231.1 + strawberry-v0.253.1: strawberry-graphql[fastapi,flask]==0.253.1 + strawberry-v0.274.0: strawberry-graphql[fastapi,flask]==0.274.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.230.0: pydantic<2.11 - strawberry-v0.251.0: pydantic<2.11 + strawberry-v0.231.1: pydantic<2.11 + strawberry-v0.253.1: pydantic<2.11 # ~~~ Network ~~~ @@ -696,13 +696,13 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.12.12: aiohttp==3.12.12 + aiohttp-v3.12.13: aiohttp==3.12.13 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.12.12: pytest-asyncio + aiohttp-v3.12.13: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 - bottle-v0.13.3: bottle==0.13.3 + bottle-v0.13.4: bottle==0.13.4 bottle: werkzeug<2.1.0 falcon-v1.4.1: falcon==1.4.1 From 3f9acc4cf74da1543a2b8fc14799ed186ef58053 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 18 Jun 2025 17:29:32 +0200 Subject: [PATCH 323/591] fix(ci): Do not install newest tracerite (#4494) New release of `tracerite` (a transitive dependency in the sanic workflow) causes [this](https://github.com/getsentry/sentry-python/actions/runs/15735197921/job/44345942053?pr=4493) to happen. Related `tracerite` bug report: https://github.com/sanic-org/tracerite/issues/20 Once this is fixed, we can unpin. --- scripts/populate_tox/tox.jinja | 1 + tox.ini | 1 + 2 files changed, 2 insertions(+) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 45f56e2f1f..3386e2ae72 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -326,6 +326,7 @@ deps = # Sanic sanic: websockets<11.0 sanic: aiohttp + {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-sanic: tracerite<1.1.2 sanic-v{24.6}: sanic_testing sanic-latest: sanic_testing {py3.6}-sanic: aiocontextvars==0.2.1 diff --git a/tox.ini b/tox.ini index c0c50c6029..a94ecba825 100644 --- a/tox.ini +++ b/tox.ini @@ -487,6 +487,7 @@ deps = # Sanic sanic: websockets<11.0 sanic: aiohttp + {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-sanic: tracerite<1.1.2 sanic-v{24.6}: sanic_testing sanic-latest: sanic_testing {py3.6}-sanic: aiocontextvars==0.2.1 From d39599fc374b01a62fd702fa3adc59aac0f2b79c Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Thu, 19 Jun 2025 10:03:34 -0400 Subject: [PATCH 324/591] fix(profiling): Ensure profiler thread exits when needed (#4497) The soft exit wasn't properly shutting down the thread if another profiler started up too quickly. This ensures it is reused if possible but is properly shutdown if needed. Specifically, the shutdown allowed the profiler 1 cycle before actually shutting down. If another profiler is started during this cycle, it's possible the old profiler never shuts down. Resulting in multiple profilers running. Fixes #4489 --- sentry_sdk/profiler/continuous_profiler.py | 38 +++++++++++++++------- tests/profiler/test_continuous_profiler.py | 31 +++++++++++++++--- 2 files changed, 53 insertions(+), 16 deletions(-) diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 77ba60dbda..00dd29e36c 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -236,6 +236,7 @@ def __init__(self, frequency, options, sdk_info, capture_func): self.pid = None # type: Optional[int] self.running = False + self.soft_shutdown = False self.new_profiles = deque(maxlen=128) # type: Deque[ContinuousProfile] self.active_profiles = set() # type: Set[ContinuousProfile] @@ -317,7 +318,7 @@ def profiler_id(self): return self.buffer.profiler_id def make_sampler(self): - # type: () -> Callable[..., None] + # type: () -> Callable[..., bool] cwd = os.getcwd() cache = LRUCache(max_size=256) @@ -325,7 +326,7 @@ def make_sampler(self): if self.lifecycle == "trace": def _sample_stack(*args, **kwargs): - # type: (*Any, **Any) -> None + # type: (*Any, **Any) -> bool """ Take a sample of the stack on all the threads in the process. This should be called at a regular interval to collect samples. @@ -333,8 +334,7 @@ def _sample_stack(*args, **kwargs): # no profiles taking place, so we can stop early if not self.new_profiles and not self.active_profiles: - self.running = False - return + return True # This is the number of profiles we want to pop off. # It's possible another thread adds a new profile to @@ -357,7 +357,7 @@ def _sample_stack(*args, **kwargs): # For some reason, the frame we get doesn't have certain attributes. # When this happens, we abandon the current sample as it's bad. capture_internal_exception(sys.exc_info()) - return + return False # Move the new profiles into the active_profiles set. # @@ -374,9 +374,7 @@ def _sample_stack(*args, **kwargs): inactive_profiles = [] for profile in self.active_profiles: - if profile.active: - pass - else: + if not profile.active: # If a profile is marked inactive, we buffer it # to `inactive_profiles` so it can be removed. # We cannot remove it here as it would result @@ -389,10 +387,12 @@ def _sample_stack(*args, **kwargs): if self.buffer is not None: self.buffer.write(ts, sample) + return False + else: def _sample_stack(*args, **kwargs): - # type: (*Any, **Any) -> None + # type: (*Any, **Any) -> bool """ Take a sample of the stack on all the threads in the process. This should be called at a regular interval to collect samples. @@ -409,11 +409,13 @@ def _sample_stack(*args, **kwargs): # For some reason, the frame we get doesn't have certain attributes. # When this happens, we abandon the current sample as it's bad. capture_internal_exception(sys.exc_info()) - return + return False if self.buffer is not None: self.buffer.write(ts, sample) + return False + return _sample_stack def run(self): @@ -421,7 +423,7 @@ def run(self): last = time.perf_counter() while self.running: - self.sampler() + self.soft_shutdown = self.sampler() # some time may have elapsed since the last time # we sampled, so we need to account for that and @@ -430,6 +432,15 @@ def run(self): if elapsed < self.interval: thread_sleep(self.interval - elapsed) + # the soft shutdown happens here to give it a chance + # for the profiler to be reused + if self.soft_shutdown: + self.running = False + + # make sure to explicitly exit the profiler here or there might + # be multiple profilers at once + break + # after sleeping, make sure to take the current # timestamp so we can use it next iteration last = time.perf_counter() @@ -458,6 +469,8 @@ def __init__(self, frequency, options, sdk_info, capture_func): def ensure_running(self): # type: () -> None + self.soft_shutdown = False + pid = os.getpid() # is running on the right process @@ -532,6 +545,9 @@ def __init__(self, frequency, options, sdk_info, capture_func): def ensure_running(self): # type: () -> None + + self.soft_shutdown = False + pid = os.getpid() # is running on the right process diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 991f8bda5d..7283ec7164 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -459,33 +459,54 @@ def test_continuous_profiler_auto_start_and_stop_sampled( thread = threading.current_thread() + all_profiler_ids = set() + for _ in range(3): envelopes.clear() + profiler_ids = set() + with sentry_sdk.start_transaction(name="profiling 1"): - assert get_profiler_id() is not None, "profiler should be running" + profiler_id = get_profiler_id() + assert profiler_id is not None, "profiler should be running" + profiler_ids.add(profiler_id) with sentry_sdk.start_span(op="op"): time.sleep(0.1) - assert get_profiler_id() is not None, "profiler should be running" + profiler_id = get_profiler_id() + assert profiler_id is not None, "profiler should be running" + profiler_ids.add(profiler_id) + + time.sleep(0.03) # the profiler takes a while to stop in auto mode so if we start # a transaction immediately, it'll be part of the same chunk - assert get_profiler_id() is not None, "profiler should be running" + profiler_id = get_profiler_id() + assert profiler_id is not None, "profiler should be running" + profiler_ids.add(profiler_id) with sentry_sdk.start_transaction(name="profiling 2"): - assert get_profiler_id() is not None, "profiler should be running" + profiler_id = get_profiler_id() + assert profiler_id is not None, "profiler should be running" + profiler_ids.add(profiler_id) with sentry_sdk.start_span(op="op"): time.sleep(0.1) - assert get_profiler_id() is not None, "profiler should be running" + profiler_id = get_profiler_id() + assert profiler_id is not None, "profiler should be running" + profiler_ids.add(profiler_id) # wait at least 1 cycle for the profiler to stop time.sleep(0.2) assert get_profiler_id() is None, "profiler should not be running" + assert len(profiler_ids) == 1 + all_profiler_ids.add(profiler_ids.pop()) + assert_single_transaction_with_profile_chunks( envelopes, thread, max_chunks=1, transactions=2 ) + assert len(all_profiler_ids) == 3 + @pytest.mark.parametrize( "mode", From ae06ef177b320a3fb1400c225105e4bf3503c987 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 23 Jun 2025 12:37:54 +0200 Subject: [PATCH 325/591] fix(ci): Remove tracerite pin (almost) (#4504) This reverts commit 3f9acc4cf74da1543a2b8fc14799ed186ef58053. - tracerite 1.12 contained syntax that did not work on Python 3.x - tracerite 1.13 was then released, containing a fix - however, on Python 3.8 newest tracerite seems to be using importlib features that were only added in 3.9 (see below), so still pinning it to an older version there ``` Traceback: .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/_pytest/python.py:493: in importtestmodule mod = import_path( .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/_pytest/pathlib.py:587: in import_path importlib.import_module(module_name) ../../.pyenv/versions/3.8.18/lib/python3.8/importlib/__init__.py:127: in import_module return _bootstrap._gcd_import(name[level:], package, level) :1014: in _gcd_import ??? :991: in _find_and_load ??? :961: in _find_and_load_unlocked ??? :219: in _call_with_frames_removed ??? :1014: in _gcd_import ??? :991: in _find_and_load ??? :975: in _find_and_load_unlocked ??? :671: in _load_unlocked ??? :843: in exec_module ??? :219: in _call_with_frames_removed ??? tests/integrations/sanic/__init__.py:3: in import sanic .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/__init__.py:6: in from sanic.app import Sanic .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/app.py:58: in from sanic.application.state import ApplicationState, ServerStage .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/application/state.py:13: in from sanic.server.async_server import AsyncioServer .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/server/__init__.py:5: in from sanic.server.runners import serve .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/server/runners.py:6: in from sanic.config import Config .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/config.py:13: in from sanic.errorpages import DEFAULT_FORMAT, check_error_format .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/errorpages.py:27: in from sanic.pages.error import ErrorPage .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/sanic/pages/error.py:3: in import tracerite.html .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/tracerite/__init__.py:1: in from .html import html_traceback .tox/py3.8-sanic-v24.6/lib/python3.8/site-packages/tracerite/html.py:5: in from importlib.resources import files E ImportError: cannot import name 'files' from 'importlib.resources' (/Users/ivana/.pyenv/versions/3.8.18/lib/python3.8/importlib/resources.py) ``` --- scripts/populate_tox/tox.jinja | 2 +- tox.ini | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 3386e2ae72..f95a913fd9 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -326,10 +326,10 @@ deps = # Sanic sanic: websockets<11.0 sanic: aiohttp - {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-sanic: tracerite<1.1.2 sanic-v{24.6}: sanic_testing sanic-latest: sanic_testing {py3.6}-sanic: aiocontextvars==0.2.1 + {py3.8}-sanic: tracerite<1.1.2 sanic-v0.8: sanic~=0.8.0 sanic-v20: sanic~=20.0 sanic-v24.6: sanic~=24.6.0 diff --git a/tox.ini b/tox.ini index a94ecba825..7efbcb6d55 100644 --- a/tox.ini +++ b/tox.ini @@ -487,10 +487,10 @@ deps = # Sanic sanic: websockets<11.0 sanic: aiohttp - {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-sanic: tracerite<1.1.2 sanic-v{24.6}: sanic_testing sanic-latest: sanic_testing {py3.6}-sanic: aiocontextvars==0.2.1 + {py3.8}-sanic: tracerite<1.1.2 sanic-v0.8: sanic~=0.8.0 sanic-v20: sanic~=20.0 sanic-v24.6: sanic~=24.6.0 From 3e2994800dc99b07b016053878e90d5b64dbdeae Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Tue, 24 Jun 2025 12:17:39 +0200 Subject: [PATCH 326/591] Cursor generated rules (#4493) It also added performance and aws related files but I want to keep it simple for now. This adds: * quick reference * testing guide * project overview * core architecture * integration guide --- .cursor/rules/core-architecture.mdc | 122 +++++++++++++++++++++ .cursor/rules/integrations-guide.mdc | 158 +++++++++++++++++++++++++++ .cursor/rules/project-overview.mdc | 47 ++++++++ .cursor/rules/quick-reference.mdc | 51 +++++++++ .cursor/rules/testing-guide.mdc | 93 ++++++++++++++++ 5 files changed, 471 insertions(+) create mode 100644 .cursor/rules/core-architecture.mdc create mode 100644 .cursor/rules/integrations-guide.mdc create mode 100644 .cursor/rules/project-overview.mdc create mode 100644 .cursor/rules/quick-reference.mdc create mode 100644 .cursor/rules/testing-guide.mdc diff --git a/.cursor/rules/core-architecture.mdc b/.cursor/rules/core-architecture.mdc new file mode 100644 index 0000000000..885773f16d --- /dev/null +++ b/.cursor/rules/core-architecture.mdc @@ -0,0 +1,122 @@ +--- +description: +globs: +alwaysApply: false +--- +# Core Architecture + +## Scope and Client Pattern + +The Sentry SDK uses a **Scope and Client** pattern for managing state and context: + +### Scope +- [sentry_sdk/scope.py](mdc:sentry_sdk/scope.py) - Holds contextual data +- Holds a reference to the Client +- Contains tags, extra data, user info, breadcrumbs +- Thread-local storage for isolation + +### Client +- [sentry_sdk/client.py](mdc:sentry_sdk/client.py) - Handles event processing +- Manages transport and event serialization +- Applies sampling and filtering + +## Key Components + +### API Layer +- [sentry_sdk/api.py](mdc:sentry_sdk/api.py) - Public API functions +- `init()` - Initialize the SDK +- `capture_exception()` - Capture exceptions +- `capture_message()` - Capture custom messages +- `set_tag()`, `set_user()`, `set_context()` - Add context +- `start_transaction()` - Performance monitoring + +### Transport +- [sentry_sdk/transport.py](mdc:sentry_sdk/transport.py) - Event delivery +- `HttpTransport` - HTTP transport to Sentry servers +- Handles retries, rate limiting, and queuing + +### Integrations System +- [sentry_sdk/integrations/__init__.py](mdc:sentry_sdk/integrations/__init__.py) - Integration registry +- Base `Integration` class for all integrations +- Automatic setup and teardown +- Integration-specific configuration + +## Data Flow + +### Event Capture Flow +1. **Exception occurs** or **manual capture** called +2. **get_current_scope** gets the active current scope +2. **get_isolation_scope** gets the active isolation scope +3. **Scope data** (tags, user, context) is attached +4. **Client.process_event()** processes the event +5. **Sampling** and **filtering** applied +6. **Transport** sends to Sentry servers + +### Performance Monitoring Flow +1. **Transaction started** with `start_transaction()` +2. **Spans** created for operations within transaction with `start_span()` +3. **Timing data** collected automatically +4. **Transaction finished** and sent to Sentry + +## Context Management + +### Scope Stack +- **Global scope**: Default scope for the process +- **Isolation scope**: Isolated scope for specific operations, manages concurrency isolation +- **Current scope**: Active scope for current execution context + +### Scope Operations +- `configure_scope()` - Modify current scope +- `new_scope()` - Create isolated scope + +## Integration Architecture + +### Integration Lifecycle +1. **Registration**: Integration registered during `init()` +2. **Setup**: `setup_once()` called to install hooks +3. **Runtime**: Integration monitors and captures events +4. **Teardown**: Integration cleaned up on shutdown + +### Common Integration Patterns +- **Monkey patching**: Replace functions/methods with instrumented versions +- **Signal handlers**: Hook into framework signals/events +- **Middleware**: Add middleware to web frameworks +- **Exception handlers**: Catch and process exceptions + +### Integration Configuration +```python +# Example integration setup +sentry_sdk.init( + dsn="your-dsn", + integrations=[ + DjangoIntegration(), + CeleryIntegration(), + RedisIntegration(), + ], + traces_sample_rate=1.0, +) +``` + +## Error Handling + +### Exception Processing +- **Automatic capture**: Unhandled exceptions captured automatically +- **Manual capture**: Use `capture_exception()` for handled exceptions +- **Context preservation**: Stack traces, local variables, and context preserved + +### Breadcrumbs +- **Automatic breadcrumbs**: Framework operations logged automatically +- **Manual breadcrumbs**: Use `add_breadcrumb()` for custom events +- **Breadcrumb categories**: HTTP, database, navigation, etc. + +## Performance Monitoring + +### Transaction Tracking +- **Automatic transactions**: Web requests, background tasks +- **Custom transactions**: Use `start_transaction()` for custom operations +- **Span tracking**: Database queries, HTTP requests, custom operations +- **Performance data**: Timing, resource usage, custom measurements + +### Sampling +- **Transaction sampling**: Control percentage of transactions captured +- **Dynamic sampling**: Adjust sampling based on context diff --git a/.cursor/rules/integrations-guide.mdc b/.cursor/rules/integrations-guide.mdc new file mode 100644 index 0000000000..869a7f742a --- /dev/null +++ b/.cursor/rules/integrations-guide.mdc @@ -0,0 +1,158 @@ +--- +description: +globs: +alwaysApply: false +--- +# Integrations Guide + +## Integration Categories + +The Sentry Python SDK includes integrations for popular frameworks, libraries, and services: + +### Web Frameworks +- [sentry_sdk/integrations/django/](mdc:sentry_sdk/integrations/django) - Django web framework +- [sentry_sdk/integrations/flask/](mdc:sentry_sdk/integrations/flask) - Flask microframework +- [sentry_sdk/integrations/fastapi/](mdc:sentry_sdk/integrations/fastapi) - FastAPI framework +- [sentry_sdk/integrations/starlette/](mdc:sentry_sdk/integrations/starlette) - Starlette ASGI framework +- [sentry_sdk/integrations/sanic/](mdc:sentry_sdk/integrations/sanic) - Sanic async framework +- [sentry_sdk/integrations/tornado/](mdc:sentry_sdk/integrations/tornado) - Tornado web framework +- [sentry_sdk/integrations/pyramid/](mdc:sentry_sdk/integrations/pyramid) - Pyramid framework +- [sentry_sdk/integrations/bottle/](mdc:sentry_sdk/integrations/bottle) - Bottle microframework +- [sentry_sdk/integrations/chalice/](mdc:sentry_sdk/integrations/chalice) - AWS Chalice +- [sentry_sdk/integrations/quart/](mdc:sentry_sdk/integrations/quart) - Quart async framework +- [sentry_sdk/integrations/falcon/](mdc:sentry_sdk/integrations/falcon) - Falcon framework +- [sentry_sdk/integrations/litestar/](mdc:sentry_sdk/integrations/litestar) - Litestar framework +- [sentry_sdk/integrations/starlite/](mdc:sentry_sdk/integrations/starlite) - Starlite framework + +### Task Queues and Background Jobs +- [sentry_sdk/integrations/celery/](mdc:sentry_sdk/integrations/celery) - Celery task queue +- [sentry_sdk/integrations/rq/](mdc:sentry_sdk/integrations/rq) - Redis Queue +- [sentry_sdk/integrations/huey/](mdc:sentry_sdk/integrations/huey) - Huey task queue +- [sentry_sdk/integrations/arq/](mdc:sentry_sdk/integrations/arq) - Arq async task queue +- [sentry_sdk/integrations/dramatiq/](mdc:sentry_sdk/integrations/dramatiq) - Dramatiq task queue + +### Databases and Data Stores +- [sentry_sdk/integrations/sqlalchemy/](mdc:sentry_sdk/integrations/sqlalchemy) - SQLAlchemy ORM +- [sentry_sdk/integrations/asyncpg/](mdc:sentry_sdk/integrations/asyncpg) - AsyncPG PostgreSQL +- [sentry_sdk/integrations/pymongo/](mdc:sentry_sdk/integrations/pymongo) - PyMongo MongoDB +- [sentry_sdk/integrations/redis/](mdc:sentry_sdk/integrations/redis) - Redis client +- [sentry_sdk/integrations/clickhouse_driver/](mdc:sentry_sdk/integrations/clickhouse_driver) - ClickHouse driver + +### Cloud and Serverless +- [sentry_sdk/integrations/aws_lambda/](mdc:sentry_sdk/integrations/aws_lambda) - AWS Lambda +- [sentry_sdk/integrations/gcp/](mdc:sentry_sdk/integrations/gcp) - Google Cloud Platform +- [sentry_sdk/integrations/serverless/](mdc:sentry_sdk/integrations/serverless) - Serverless framework + +### HTTP and Networking +- [sentry_sdk/integrations/requests/](mdc:sentry_sdk/integrations/requests) - Requests HTTP library +- [sentry_sdk/integrations/httpx/](mdc:sentry_sdk/integrations/httpx) - HTTPX async HTTP client +- [sentry_sdk/integrations/aiohttp/](mdc:sentry_sdk/integrations/aiohttp) - aiohttp async HTTP +- [sentry_sdk/integrations/grpc/](mdc:sentry_sdk/integrations/grpc) - gRPC framework + +### AI and Machine Learning +- [sentry_sdk/integrations/openai/](mdc:sentry_sdk/integrations/openai) - OpenAI API +- [sentry_sdk/integrations/anthropic/](mdc:sentry_sdk/integrations/anthropic) - Anthropic Claude +- [sentry_sdk/integrations/cohere/](mdc:sentry_sdk/integrations/cohere) - Cohere AI +- [sentry_sdk/integrations/huggingface_hub/](mdc:sentry_sdk/integrations/huggingface_hub) - Hugging Face Hub +- [sentry_sdk/integrations/langchain/](mdc:sentry_sdk/integrations/langchain) - LangChain framework + +### GraphQL +- [sentry_sdk/integrations/graphene/](mdc:sentry_sdk/integrations/graphene) - Graphene GraphQL +- [sentry_sdk/integrations/ariadne/](mdc:sentry_sdk/integrations/ariadne) - Ariadne GraphQL +- [sentry_sdk/integrations/strawberry/](mdc:sentry_sdk/integrations/strawberry) - Strawberry GraphQL +- [sentry_sdk/integrations/gql/](mdc:sentry_sdk/integrations/gql) - GQL GraphQL client + +### Feature Flags and Configuration +- [sentry_sdk/integrations/launchdarkly/](mdc:sentry_sdk/integrations/launchdarkly) - LaunchDarkly +- [sentry_sdk/integrations/unleash/](mdc:sentry_sdk/integrations/unleash) - Unleash +- [sentry_sdk/integrations/statsig/](mdc:sentry_sdk/integrations/statsig) - Statsig +- [sentry_sdk/integrations/openfeature/](mdc:sentry_sdk/integrations/openfeature) - OpenFeature + +### Other Integrations +- [sentry_sdk/integrations/logging/](mdc:sentry_sdk/integrations/logging) - Python logging +- [sentry_sdk/integrations/loguru/](mdc:sentry_sdk/integrations/loguru) - Loguru logging +- [sentry_sdk/integrations/opentelemetry/](mdc:sentry_sdk/integrations/opentelemetry) - OpenTelemetry +- [sentry_sdk/integrations/ray/](mdc:sentry_sdk/integrations/ray) - Ray distributed computing +- [sentry_sdk/integrations/spark/](mdc:sentry_sdk/integrations/spark) - Apache Spark +- [sentry_sdk/integrations/beam/](mdc:sentry_sdk/integrations/beam) - Apache Beam + +## Integration Usage + +### Basic Integration Setup +```python +import sentry_sdk +from sentry_sdk.integrations.django import DjangoIntegration +from sentry_sdk.integrations.celery import CeleryIntegration + +sentry_sdk.init( + dsn="your-dsn", + integrations=[ + DjangoIntegration(), + CeleryIntegration(), + ], + traces_sample_rate=1.0, +) +``` + +### Integration Configuration +Most integrations accept configuration parameters: +```python +from sentry_sdk.integrations.django import DjangoIntegration +from sentry_sdk.integrations.redis import RedisIntegration + +sentry_sdk.init( + dsn="your-dsn", + integrations=[ + DjangoIntegration( + transaction_style="url", # Customize transaction naming + ), + RedisIntegration( + cache_prefixes=["myapp:"], # Filter cache operations + ), + ], +) +``` + +### Integration Testing +Each integration has corresponding tests in [tests/integrations/](mdc:tests/integrations): +- [tests/integrations/django/](mdc:tests/integrations/django) - Django integration tests +- [tests/integrations/flask/](mdc:tests/integrations/flask) - Flask integration tests +- [tests/integrations/celery/](mdc:tests/integrations/celery) - Celery integration tests + +## Integration Development + +### Creating New Integrations +1. **Create integration file** in [sentry_sdk/integrations/](mdc:sentry_sdk/integrations) +2. **Inherit from Integration base class** +3. **Implement setup_once() method** +4. **Add to integration registry** + +### Integration Base Class +```python +from sentry_sdk.integrations import Integration + +class MyIntegration(Integration): + identifier = "my_integration" + + def __init__(self, param=None): + self.param = param + + @staticmethod + def setup_once(): + # Install hooks, monkey patches, etc. + pass +``` + +### Common Integration Patterns +- **Monkey patching**: Replace functions with instrumented versions +- **Middleware**: Add middleware to web frameworks +- **Signal handlers**: Hook into framework signals +- **Exception handlers**: Catch and process exceptions +- **Context managers**: Add context to operations + +### Integration Best Practices +- **Zero configuration**: Work without user setup +- **Check integration status**: Use `sentry_sdk.get_client().get_integration()` +- **No side effects**: Don't alter library behavior +- **Graceful degradation**: Handle missing dependencies +- **Comprehensive testing**: Test all integration features diff --git a/.cursor/rules/project-overview.mdc b/.cursor/rules/project-overview.mdc new file mode 100644 index 0000000000..13fad83ae7 --- /dev/null +++ b/.cursor/rules/project-overview.mdc @@ -0,0 +1,47 @@ +--- +description: +globs: +alwaysApply: false +--- +# Sentry Python SDK - Project Overview + +## What is this project? + +The Sentry Python SDK is the official Python SDK for [Sentry](mdc:https://sentry.io), an error monitoring and performance monitoring platform. It helps developers capture errors, exceptions, traces and profiles from Python applications. + +## Key Files and Directories + +### Core SDK +- [sentry_sdk/__init__.py](mdc:sentry_sdk/__init__.py) - Main entry point, exports all public APIs +- [sentry_sdk/api.py](mdc:sentry_sdk/api.py) - Public API functions (init, capture_exception, etc.) +- [sentry_sdk/client.py](mdc:sentry_sdk/client.py) - Core client implementation +- [sentry_sdk/scope.py](mdc:sentry_sdk/scope.py) - Scope holds contextual metadata such as tags that are applied automatically to events and envelopes +- [sentry_sdk/transport.py](mdc:sentry_sdk/transport.py) - HTTP Transport that sends the envelopes to Sentry's servers +- [sentry_sdk/worker.py](mdc:sentry_sdk/worker.py) - Background threaded worker with a queue to manage transport requests +- [sentry_sdk/serializer.py](mdc:sentry_sdk/serializer.py) - Serializes the payload along with truncation logic + +### Integrations +- [sentry_sdk/integrations/](mdc:sentry_sdk/integrations) - Framework and library integrations + - [sentry_sdk/integrations/__init__.py](mdc:sentry_sdk/integrations/__init__.py) - Integration registry + - [sentry_sdk/integrations/django/](mdc:sentry_sdk/integrations/django) - Django framework integration + - [sentry_sdk/integrations/flask/](mdc:sentry_sdk/integrations/flask) - Flask framework integration + - [sentry_sdk/integrations/fastapi/](mdc:sentry_sdk/integrations/fastapi) - FastAPI integration + - [sentry_sdk/integrations/celery/](mdc:sentry_sdk/integrations/celery) - Celery task queue integration + - [sentry_sdk/integrations/aws_lambda/](mdc:sentry_sdk/integrations/aws_lambda) - AWS Lambda integration + +### Configuration and Setup +- [setup.py](mdc:setup.py) - Package configuration and dependencies +- [pyproject.toml](mdc:pyproject.toml) - Modern Python project configuration +- [tox.ini](mdc:tox.ini) - Test matrix configuration for multiple Python versions and integrations +- [requirements-*.txt](mdc:requirements-testing.txt) - Various dependency requirements + +### Documentation and Guides +- [README.md](mdc:README.md) - Project overview and quick start +- [CONTRIBUTING.md](mdc:CONTRIBUTING.md) - Development and contribution guidelines +- [MIGRATION_GUIDE.md](mdc:MIGRATION_GUIDE.md) - Migration from older versions +- [CHANGELOG.md](mdc:CHANGELOG.md) - Version history and changes + +### Testing +- [tests/](mdc:tests) - Comprehensive test suite + - [tests/integrations/](mdc:tests/integrations) - Integration-specific tests + - [tests/conftest.py](mdc:tests/conftest.py) - Pytest configuration and fixtures diff --git a/.cursor/rules/quick-reference.mdc b/.cursor/rules/quick-reference.mdc new file mode 100644 index 0000000000..453869fa83 --- /dev/null +++ b/.cursor/rules/quick-reference.mdc @@ -0,0 +1,51 @@ +--- +description: +globs: +alwaysApply: false +--- +# Quick Reference + +## Common Commands + +### Development Setup +```bash +make .venv +source .venv/bin/activate # Windows: .venv\Scripts\activate +``` + +### Testing + +Our test matrix is implemented in [tox](mdc:https://tox.wiki). +The following runs the whole test suite and takes a long time. + +```bash +source .venv/bin/activate +tox +``` + +Prefer testing a single environment instead while developing. + +```bash +tox -e py3.12-common +``` + +For running a single test, use the pattern: + +```bash +tox -e py3.12-common -- project/tests/test_file.py::TestClassName::test_method +``` + +For testing specific integrations, refer to the test matrix in [sentry_sdk/tox.ini](mdc:sentry_sdk/tox.ini) for finding an entry. +For example, to test django, use: + +```bash +tox -e py3.12-django-v5.2.3 +``` + +### Code Quality + +Our `linters` tox environment runs `black` for formatting, `flake8` for linting and `mypy` for type checking. + +```bash +tox -e linters +``` diff --git a/.cursor/rules/testing-guide.mdc b/.cursor/rules/testing-guide.mdc new file mode 100644 index 0000000000..e336bb337a --- /dev/null +++ b/.cursor/rules/testing-guide.mdc @@ -0,0 +1,93 @@ +--- +description: +globs: +alwaysApply: false +--- +# Testing Guide + +## Test Structure + +### Test Organization +- [tests/](mdc:tests) - Main test directory +- [tests/conftest.py](mdc:tests/conftest.py) - Shared pytest fixtures and configuration +- [tests/integrations/](mdc:tests/integrations) - Integration-specific tests +- [tests/tracing/](mdc:tests/tracing) - Performance monitoring tests +- [tests/utils/](mdc:tests/utils) - Utility and helper tests + +### Integration Test Structure +Each integration has its own test directory: +- [tests/integrations/django/](mdc:tests/integrations/django) - Django integration tests +- [tests/integrations/flask/](mdc:tests/integrations/flask) - Flask integration tests +- [tests/integrations/celery/](mdc:tests/integrations/celery) - Celery integration tests +- [tests/integrations/aws_lambda/](mdc:tests/integrations/aws_lambda) - AWS Lambda tests + +## Running Tests + +### Tox Testing Matrix + +The [tox.ini](mdc:tox.ini) file defines comprehensive test environments. +Always run tests via `tox` from the main `.venv`. + +```bash +source .venv/bin/activate + +# Run all tox environments, takes a long time +tox + +# Run specific environment +tox -e py3.11-django-v4.2 + +# Run environments for specific Python version +tox -e py3.11-* + +# Run environments for specific integration +tox -e *-django-* + +# Run a single test +tox -e py3.12-common -- project/tests/test_file.py::TestClassName::test_method +``` + +### Test Environment Categories +- **Common tests**: `{py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common` +- **Integration tests**: `{python_version}-{integration}-v{framework_version}` +- **Gevent tests**: `{py3.6,py3.8,py3.10,py3.11,py3.12}-gevent` + +## Writing Tests + +### Test File Structure +```python +import pytest +import sentry_sdk +from sentry_sdk.integrations.flask import FlaskIntegration + +def test_flask_integration(sentry_init, capture_events): + """Test Flask integration captures exceptions.""" + # Test setup + sentry_init(integrations=[FlaskIntegration()]) + events = capture_events() + + # Test execution + # ... test code ... + + # Assertions + assert len(events) == 1 + assert events[0]["exception"]["values"][0]["type"] == "ValueError" +``` + +### Common Test Patterns + +## Test Best Practices + +### Test Organization +- **One test per function**: Each test should verify one specific behavior +- **Descriptive names**: Use clear, descriptive test function names +- **Arrange-Act-Assert**: Structure tests with setup, execution, and verification +- **Isolation**: Each test should be independent and not affect others +- **No mocking**: Never use mocks in tests +- **Cleanup**: Ensure tests clean up after themselves + +## Fixtures +The most important fixtures for testing are: +- `sentry_init`: Use in the beginning of a test to simulate initializing the SDK +- `capture_events`: Intercept the events for testing event payload +- `capture_envelopes`: Intercept the envelopes for testing envelope headers and payload From ad2bbff928bef9464dc13099dae1fb200717aaf4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 24 Jun 2025 14:12:31 +0200 Subject: [PATCH 327/591] tests: Tox update (#4509) --- tox.ini | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/tox.ini b/tox.ini index 7efbcb6d55..f4aee13d02 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-06-17T08:49:27.078408+00:00 +# Last generated: 2025-06-24T07:19:36.122984+00:00 [tox] requires = @@ -138,7 +138,7 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.29.2 {py3.8,py3.11,py3.12}-anthropic-v0.42.0 - {py3.8,py3.11,py3.12}-anthropic-v0.54.0 + {py3.8,py3.11,py3.12}-anthropic-v0.55.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 @@ -180,7 +180,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.58.2 + {py3.7,py3.12,py3.13}-statsig-v0.58.3 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -203,7 +203,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.231.1 {py3.8,py3.12,py3.13}-strawberry-v0.253.1 - {py3.9,py3.12,py3.13}-strawberry-v0.274.0 + {py3.9,py3.12,py3.13}-strawberry-v0.275.2 # ~~~ Network ~~~ @@ -249,12 +249,12 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.47.0 + {py3.9,py3.12,py3.13}-starlette-v0.47.1 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 {py3.7,py3.10,py3.11}-fastapi-v0.103.2 - {py3.8,py3.12,py3.13}-fastapi-v0.115.12 + {py3.8,py3.12,py3.13}-fastapi-v0.115.13 # ~~~ Web 2 ~~~ @@ -504,7 +504,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.29.2: anthropic==0.29.2 anthropic-v0.42.0: anthropic==0.42.0 - anthropic-v0.54.0: anthropic==0.54.0 + anthropic-v0.55.0: anthropic==0.55.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.29.2: httpx<0.28.0 @@ -551,7 +551,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 - statsig-v0.58.2: statsig==0.58.2 + statsig-v0.58.3: statsig==0.58.3 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -583,7 +583,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.231.1: strawberry-graphql[fastapi,flask]==0.231.1 strawberry-v0.253.1: strawberry-graphql[fastapi,flask]==0.253.1 - strawberry-v0.274.0: strawberry-graphql[fastapi,flask]==0.274.0 + strawberry-v0.275.2: strawberry-graphql[fastapi,flask]==0.275.2 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.231.1: pydantic<2.11 @@ -666,7 +666,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.47.0: starlette==0.47.0 + starlette-v0.47.1: starlette==0.47.1 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -681,7 +681,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.91.0: fastapi==0.91.0 fastapi-v0.103.2: fastapi==0.103.2 - fastapi-v0.115.12: fastapi==0.115.12 + fastapi-v0.115.13: fastapi==0.115.13 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart From 7f507fd4e0cdbb1f071766431a6dcf3db77b7bb1 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 24 Jun 2025 14:36:58 +0200 Subject: [PATCH 328/591] ref(langchain): Greatly simplify `_wrap_configure` (#4479) Resolving #4443 requires some changes to this method, but the current `args`/`kwargs` business makes the method difficult to reason through. This PR simplifies the logic by listing out the parameters we need to access, so we don't need to access them through `args` and `kwargs`. We also cut down on the amount of branching and the amount of variables (`new_callbacks` vs `existing_callbacks`). Behavior does not change in this PR; we fix the #4443 bug in #4485, which is based on this PR --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- sentry_sdk/integrations/langchain.py | 90 +++++++++++++++------------- 1 file changed, 49 insertions(+), 41 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 431fc46bec..1064f29ffd 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -22,6 +22,7 @@ from langchain_core.callbacks import ( manager, BaseCallbackHandler, + Callbacks, ) from langchain_core.agents import AgentAction, AgentFinish except ImportError: @@ -416,50 +417,57 @@ def _wrap_configure(f): # type: (Callable[..., Any]) -> Callable[..., Any] @wraps(f) - def new_configure(*args, **kwargs): - # type: (Any, Any) -> Any + def new_configure( + callback_manager_cls, # type: type + inheritable_callbacks=None, # type: Callbacks + local_callbacks=None, # type: Callbacks + *args, # type: Any + **kwargs, # type: Any + ): + # type: (...) -> Any integration = sentry_sdk.get_client().get_integration(LangchainIntegration) if integration is None: - return f(*args, **kwargs) + return f( + callback_manager_cls, + inheritable_callbacks, + local_callbacks, + *args, + **kwargs, + ) - with capture_internal_exceptions(): - new_callbacks = [] # type: List[BaseCallbackHandler] - if "local_callbacks" in kwargs: - existing_callbacks = kwargs["local_callbacks"] - kwargs["local_callbacks"] = new_callbacks - elif len(args) > 2: - existing_callbacks = args[2] - args = ( - args[0], - args[1], - new_callbacks, - ) + args[3:] - else: - existing_callbacks = [] - - if existing_callbacks: - if isinstance(existing_callbacks, list): - for cb in existing_callbacks: - new_callbacks.append(cb) - elif isinstance(existing_callbacks, BaseCallbackHandler): - new_callbacks.append(existing_callbacks) - else: - logger.debug("Unknown callback type: %s", existing_callbacks) - - already_added = False - for callback in new_callbacks: - if isinstance(callback, SentryLangchainCallback): - already_added = True - - if not already_added: - new_callbacks.append( - SentryLangchainCallback( - integration.max_spans, - integration.include_prompts, - integration.tiktoken_encoding_name, - ) - ) - return f(*args, **kwargs) + callbacks_list = local_callbacks or [] + + if isinstance(callbacks_list, BaseCallbackHandler): + callbacks_list = [callbacks_list] + elif not isinstance(callbacks_list, list): + logger.debug("Unknown callback type: %s", callbacks_list) + # Just proceed with original function call + return f( + callback_manager_cls, + inheritable_callbacks, + local_callbacks, + *args, + **kwargs, + ) + + if not any(isinstance(cb, SentryLangchainCallback) for cb in callbacks_list): + # Avoid mutating the existing callbacks list + callbacks_list = [ + *callbacks_list, + SentryLangchainCallback( + integration.max_spans, + integration.include_prompts, + integration.tiktoken_encoding_name, + ), + ] + + return f( + callback_manager_cls, + inheritable_callbacks, + callbacks_list, + *args, + **kwargs, + ) return new_configure From 4a0e5ed544a37109da1a8b33bda50f4871920a85 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 24 Jun 2025 14:42:53 +0200 Subject: [PATCH 329/591] Support `openai-agents` (#4437) Add support for AI agents projects using `openai-agents` (https://pypi.org/project/openai-agents/) Docs PR is here: https://github.com/getsentry/sentry-docs/pull/14113 This integration: - records tracing data of agent invocation, tool execution, ai client requests to LLMs, and handoffs to other agents. - captures input and output to/from LLMs if `set_default_pii=True`. - is mostly compatible to the OpenTelememetry `gen_ai` semantic conventions. (input and output is not compatible because Sentry does not have Span events. This information is stored in arrays on the Span attributes. - Captures errors that happen during agent execution (like problems during interaction with the LLM. This integration does not: - Capture errors during function tool exection because this is very hard to patch (see comment in the code) Example span tree in Sentry.io: ![Screenshot 2025-06-24 at 12 15 17](https://github.com/user-attachments/assets/87199067-434f-4bb9-b563-5c6fc18c56cb) --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 8 + pyproject.toml | 4 + scripts/populate_tox/config.py | 6 + scripts/populate_tox/tox.jinja | 1 + .../split_tox_gh_actions.py | 1 + sentry_sdk/consts.py | 421 +++++++++---- sentry_sdk/integrations/__init__.py | 1 + .../integrations/openai_agents/__init__.py | 53 ++ .../integrations/openai_agents/consts.py | 1 + .../openai_agents/patches/__init__.py | 4 + .../openai_agents/patches/agent_run.py | 143 +++++ .../openai_agents/patches/models.py | 50 ++ .../openai_agents/patches/runner.py | 42 ++ .../openai_agents/patches/tools.py | 77 +++ .../openai_agents/spans/__init__.py | 5 + .../openai_agents/spans/agent_workflow.py | 21 + .../openai_agents/spans/ai_client.py | 38 ++ .../openai_agents/spans/execute_tool.py | 43 ++ .../openai_agents/spans/handoff.py | 19 + .../openai_agents/spans/invoke_agent.py | 34 + .../integrations/openai_agents/utils.py | 209 +++++++ tests/integrations/openai_agents/__init__.py | 3 + .../openai_agents/test_openai_agents.py | 580 ++++++++++++++++++ tox.ini | 8 +- 24 files changed, 1638 insertions(+), 134 deletions(-) create mode 100644 sentry_sdk/integrations/openai_agents/__init__.py create mode 100644 sentry_sdk/integrations/openai_agents/consts.py create mode 100644 sentry_sdk/integrations/openai_agents/patches/__init__.py create mode 100644 sentry_sdk/integrations/openai_agents/patches/agent_run.py create mode 100644 sentry_sdk/integrations/openai_agents/patches/models.py create mode 100644 sentry_sdk/integrations/openai_agents/patches/runner.py create mode 100644 sentry_sdk/integrations/openai_agents/patches/tools.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/__init__.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/agent_workflow.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/ai_client.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/execute_tool.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/handoff.py create mode 100644 sentry_sdk/integrations/openai_agents/spans/invoke_agent.py create mode 100644 sentry_sdk/integrations/openai_agents/utils.py create mode 100644 tests/integrations/openai_agents/__init__.py create mode 100644 tests/integrations/openai_agents/test_openai_agents.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 4aa0f36b77..e81d507d27 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -66,6 +66,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-openai-latest" + - name: Test openai_agents latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_agents-latest" - name: Test huggingface_hub latest run: | set -x # print commands that are executed @@ -141,6 +145,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai" + - name: Test openai_agents pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_agents" - name: Test huggingface_hub pinned run: | set -x # print commands that are executed diff --git a/pyproject.toml b/pyproject.toml index 5e16b30793..e5eae2c21f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -183,6 +183,10 @@ ignore_missing_imports = true module = "grpc.*" ignore_missing_imports = true +[[tool.mypy.overrides]] +module = "agents.*" +ignore_missing_imports = true + # # Tool: Flake8 # diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 4664845c7b..411d7fe666 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -139,6 +139,12 @@ "loguru": { "package": "loguru", }, + "openai_agents": { + "package": "openai-agents", + "deps": { + "*": ["pytest-asyncio"], + }, + }, "openfeature": { "package": "openfeature-sdk", }, diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index f95a913fd9..ac14bdb02a 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -400,6 +400,7 @@ setenv = litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru openai: TESTPATH=tests/integrations/openai + openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry potel: TESTPATH=tests/integrations/opentelemetry diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 3fbc0ec1c5..af1ff84cd6 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -63,6 +63,7 @@ "cohere", "langchain", "openai", + "openai_agents", "huggingface_hub", ], "Cloud": [ diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 34ae5bdfd8..53148a36df 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -108,16 +108,39 @@ class SPANDATA: See: https://develop.sentry.dev/sdk/performance/span-data-conventions/ """ + AI_CITATIONS = "ai.citations" + """ + References or sources cited by the AI model in its response. + Example: ["Smith et al. 2020", "Jones 2019"] + """ + + AI_DOCUMENTS = "ai.documents" + """ + Documents or content chunks used as context for the AI model. + Example: ["doc1.txt", "doc2.pdf"] + """ + + AI_FINISH_REASON = "ai.finish_reason" + """ + The reason why the model stopped generating. + Example: "length" + """ + AI_FREQUENCY_PENALTY = "ai.frequency_penalty" """ Used to reduce repetitiveness of generated tokens. Example: 0.5 """ - AI_PRESENCE_PENALTY = "ai.presence_penalty" + AI_FUNCTION_CALL = "ai.function_call" """ - Used to reduce repetitiveness of generated tokens. - Example: 0.5 + For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls + """ + + AI_GENERATION_ID = "ai.generation_id" + """ + Unique identifier for the completion. + Example: "gen_123abc" """ AI_INPUT_MESSAGES = "ai.input_messages" @@ -126,10 +149,9 @@ class SPANDATA: Example: [{"role": "user", "message": "hello"}] """ - AI_MODEL_ID = "ai.model_id" + AI_LOGIT_BIAS = "ai.logit_bias" """ - The unique descriptor of the model being execugted - Example: gpt-4 + For an AI model call, the logit bias """ AI_METADATA = "ai.metadata" @@ -138,28 +160,94 @@ class SPANDATA: Example: {"executed_function": "add_integers"} """ - AI_TAGS = "ai.tags" + AI_MODEL_ID = "ai.model_id" """ - Tags that describe an AI pipeline step. - Example: {"executed_function": "add_integers"} + The unique descriptor of the model being execugted + Example: gpt-4 + """ + + AI_PIPELINE_NAME = "ai.pipeline.name" + """ + Name of the AI pipeline or chain being executed. + Example: "qa-pipeline" + """ + + AI_PREAMBLE = "ai.preamble" + """ + For an AI model call, the preamble parameter. + Preambles are a part of the prompt used to adjust the model's overall behavior and conversation style. + Example: "You are now a clown." + """ + + AI_PRESENCE_PENALTY = "ai.presence_penalty" + """ + Used to reduce repetitiveness of generated tokens. + Example: 0.5 + """ + + AI_RAW_PROMPTING = "ai.raw_prompting" + """ + Minimize pre-processing done to the prompt sent to the LLM. + Example: true + """ + + AI_RESPONSE_FORMAT = "ai.response_format" + """ + For an AI model call, the format of the response + """ + + AI_RESPONSES = "ai.responses" + """ + The responses to an AI model call. Always as a list. + Example: ["hello", "world"] + """ + + AI_SEARCH_QUERIES = "ai.search_queries" + """ + Queries used to search for relevant context or documents. + Example: ["climate change effects", "renewable energy"] + """ + + AI_SEARCH_REQUIRED = "ai.is_search_required" + """ + Boolean indicating if the model needs to perform a search. + Example: true + """ + + AI_SEARCH_RESULTS = "ai.search_results" + """ + Results returned from search queries for context. + Example: ["Result 1", "Result 2"] + """ + + AI_SEED = "ai.seed" + """ + The seed, ideally models given the same seed and same other parameters will produce the exact same output. + Example: 123.45 """ AI_STREAMING = "ai.streaming" """ - Whether or not the AI model call's repsonse was streamed back asynchronously + Whether or not the AI model call's response was streamed back asynchronously Example: true """ + AI_TAGS = "ai.tags" + """ + Tags that describe an AI pipeline step. + Example: {"executed_function": "add_integers"} + """ + AI_TEMPERATURE = "ai.temperature" """ For an AI model call, the temperature parameter. Temperature essentially means how random the output will be. Example: 0.5 """ - AI_TOP_P = "ai.top_p" + AI_TEXTS = "ai.texts" """ - For an AI model call, the top_p parameter. Top_p essentially controls how random the output will be. - Example: 0.5 + Raw text inputs provided to the model. + Example: ["What is machine learning?"] """ AI_TOP_K = "ai.top_k" @@ -168,9 +256,10 @@ class SPANDATA: Example: 35 """ - AI_FUNCTION_CALL = "ai.function_call" + AI_TOP_P = "ai.top_p" """ - For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls + For an AI model call, the top_p parameter. Top_p essentially controls how random the output will be. + Example: 0.5 """ AI_TOOL_CALLS = "ai.tool_calls" @@ -183,168 +272,236 @@ class SPANDATA: For an AI model call, the functions that are available """ - AI_RESPONSE_FORMAT = "ai.response_format" + AI_WARNINGS = "ai.warnings" """ - For an AI model call, the format of the response + Warning messages generated during model execution. + Example: ["Token limit exceeded"] """ - AI_LOGIT_BIAS = "ai.logit_bias" + CACHE_HIT = "cache.hit" """ - For an AI model call, the logit bias + A boolean indicating whether the requested data was found in the cache. + Example: true """ - AI_PREAMBLE = "ai.preamble" + CACHE_ITEM_SIZE = "cache.item_size" """ - For an AI model call, the preamble parameter. - Preambles are a part of the prompt used to adjust the model's overall behavior and conversation style. - Example: "You are now a clown." + The size of the requested data in bytes. + Example: 58 """ - AI_RAW_PROMPTING = "ai.raw_prompting" + CACHE_KEY = "cache.key" """ - Minimize pre-processing done to the prompt sent to the LLM. - Example: true + The key of the requested data. + Example: template.cache.some_item.867da7e2af8e6b2f3aa7213a4080edb3 """ - AI_RESPONSES = "ai.responses" + + CODE_FILEPATH = "code.filepath" """ - The responses to an AI model call. Always as a list. - Example: ["hello", "world"] + The source code file name that identifies the code unit as uniquely as possible (preferably an absolute file path). + Example: "/app/myapplication/http/handler/server.py" """ - AI_SEED = "ai.seed" + CODE_FUNCTION = "code.function" """ - The seed, ideally models given the same seed and same other parameters will produce the exact same output. - Example: 123.45 + The method or function name, or equivalent (usually rightmost part of the code unit's name). + Example: "server_request" """ - AI_CITATIONS = "ai.citations" + CODE_LINENO = "code.lineno" """ - References or sources cited by the AI model in its response. - Example: ["Smith et al. 2020", "Jones 2019"] + The line number in `code.filepath` best representing the operation. It SHOULD point within the code unit named in `code.function`. + Example: 42 """ - AI_DOCUMENTS = "ai.documents" + CODE_NAMESPACE = "code.namespace" """ - Documents or content chunks used as context for the AI model. - Example: ["doc1.txt", "doc2.pdf"] + The "namespace" within which `code.function` is defined. Usually the qualified class or module name, such that `code.namespace` + some separator + `code.function` form a unique identifier for the code unit. + Example: "http.handler" """ - AI_SEARCH_QUERIES = "ai.search_queries" + DB_MONGODB_COLLECTION = "db.mongodb.collection" """ - Queries used to search for relevant context or documents. - Example: ["climate change effects", "renewable energy"] + The MongoDB collection being accessed within the database. + See: https://github.com/open-telemetry/semantic-conventions/blob/main/docs/database/mongodb.md#attributes + Example: public.users; customers """ - AI_SEARCH_RESULTS = "ai.search_results" + DB_NAME = "db.name" """ - Results returned from search queries for context. - Example: ["Result 1", "Result 2"] + The name of the database being accessed. For commands that switch the database, this should be set to the target database (even if the command fails). + Example: myDatabase """ - AI_GENERATION_ID = "ai.generation_id" + DB_OPERATION = "db.operation" """ - Unique identifier for the completion. - Example: "gen_123abc" + The name of the operation being executed, e.g. the MongoDB command name such as findAndModify, or the SQL keyword. + See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md + Example: findAndModify, HMSET, SELECT """ - AI_SEARCH_REQUIRED = "ai.is_search_required" + DB_SYSTEM = "db.system" """ - Boolean indicating if the model needs to perform a search. - Example: true + An identifier for the database management system (DBMS) product being used. + See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md + Example: postgresql """ - AI_FINISH_REASON = "ai.finish_reason" + DB_USER = "db.user" """ - The reason why the model stopped generating. - Example: "length" + The name of the database user used for connecting to the database. + See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md + Example: my_user """ - AI_PIPELINE_NAME = "ai.pipeline.name" + GEN_AI_AGENT_NAME = "gen_ai.agent.name" """ - Name of the AI pipeline or chain being executed. - Example: "qa-pipeline" + The name of the agent being used. + Example: "ResearchAssistant" """ - AI_TEXTS = "ai.texts" + GEN_AI_CHOICE = "gen_ai.choice" """ - Raw text inputs provided to the model. - Example: ["What is machine learning?"] + The model's response message. + Example: "The weather in Paris is rainy and overcast, with temperatures around 57°F" """ - AI_WARNINGS = "ai.warnings" + GEN_AI_OPERATION_NAME = "gen_ai.operation.name" """ - Warning messages generated during model execution. - Example: ["Token limit exceeded"] + The name of the operation being performed. + Example: "chat" """ - DB_NAME = "db.name" + GEN_AI_RESPONSE_TEXT = "gen_ai.response.text" """ - The name of the database being accessed. For commands that switch the database, this should be set to the target database (even if the command fails). - Example: myDatabase + The model's response text messages. + Example: ["The weather in Paris is rainy and overcast, with temperatures around 57°F", "The weather in London is sunny and warm, with temperatures around 65°F"] """ - DB_USER = "db.user" + GEN_AI_RESPONSE_TOOL_CALLS = "gen_ai.response.tool_calls" """ - The name of the database user used for connecting to the database. - See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md - Example: my_user + The tool calls in the model's response. + Example: [{"name": "get_weather", "arguments": {"location": "Paris"}}] """ - DB_OPERATION = "db.operation" + GEN_AI_REQUEST_AVAILABLE_TOOLS = "gen_ai.request.available_tools" """ - The name of the operation being executed, e.g. the MongoDB command name such as findAndModify, or the SQL keyword. - See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md - Example: findAndModify, HMSET, SELECT + The available tools for the model. + Example: [{"name": "get_weather", "description": "Get the weather for a given location"}, {"name": "get_news", "description": "Get the news for a given topic"}] """ - DB_SYSTEM = "db.system" + GEN_AI_REQUEST_FREQUENCY_PENALTY = "gen_ai.request.frequency_penalty" """ - An identifier for the database management system (DBMS) product being used. - See: https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md - Example: postgresql + The frequency penalty parameter used to reduce repetitiveness of generated tokens. + Example: 0.1 """ - DB_MONGODB_COLLECTION = "db.mongodb.collection" + GEN_AI_REQUEST_MAX_TOKENS = "gen_ai.request.max_tokens" """ - The MongoDB collection being accessed within the database. - See: https://github.com/open-telemetry/semantic-conventions/blob/main/docs/database/mongodb.md#attributes - Example: public.users; customers + The maximum number of tokens to generate in the response. + Example: 2048 """ - CACHE_HIT = "cache.hit" + GEN_AI_REQUEST_MESSAGES = "gen_ai.request.messages" """ - A boolean indicating whether the requested data was found in the cache. - Example: true + The messages passed to the model. The "content" can be a string or an array of objects. + Example: [{role: "system", "content: "Generate a random number."}, {"role": "user", "content": [{"text": "Generate a random number between 0 and 10.", "type": "text"}]}] """ - CACHE_ITEM_SIZE = "cache.item_size" + GEN_AI_REQUEST_MODEL = "gen_ai.request.model" """ - The size of the requested data in bytes. - Example: 58 + The model identifier being used for the request. + Example: "gpt-4-turbo-preview" """ - CACHE_KEY = "cache.key" + GEN_AI_REQUEST_PRESENCE_PENALTY = "gen_ai.request.presence_penalty" """ - The key of the requested data. - Example: template.cache.some_item.867da7e2af8e6b2f3aa7213a4080edb3 + The presence penalty parameter used to reduce repetitiveness of generated tokens. + Example: 0.1 """ - NETWORK_PEER_ADDRESS = "network.peer.address" + GEN_AI_REQUEST_TEMPERATURE = "gen_ai.request.temperature" """ - Peer address of the network connection - IP address or Unix domain socket name. - Example: 10.1.2.80, /tmp/my.sock, localhost + The temperature parameter used to control randomness in the output. + Example: 0.7 """ - NETWORK_PEER_PORT = "network.peer.port" + GEN_AI_REQUEST_TOP_P = "gen_ai.request.top_p" """ - Peer port number of the network connection. - Example: 6379 + The top_p parameter used to control diversity via nucleus sampling. + Example: 1.0 """ - HTTP_QUERY = "http.query" + GEN_AI_SYSTEM = "gen_ai.system" """ - The Query string present in the URL. - Example: ?foo=bar&bar=baz + The name of the AI system being used. + Example: "openai" + """ + + GEN_AI_TOOL_DESCRIPTION = "gen_ai.tool.description" + """ + The description of the tool being used. + Example: "Searches the web for current information about a topic" + """ + + GEN_AI_TOOL_INPUT = "gen_ai.tool.input" + """ + The input of the tool being used. + Example: {"location": "Paris"} + """ + + GEN_AI_TOOL_NAME = "gen_ai.tool.name" + """ + The name of the tool being used. + Example: "web_search" + """ + + GEN_AI_TOOL_OUTPUT = "gen_ai.tool.output" + """ + The output of the tool being used. + Example: "rainy, 57°F" + """ + + GEN_AI_TOOL_TYPE = "gen_ai.tool.type" + """ + The type of tool being used. + Example: "function" + """ + + GEN_AI_USAGE_INPUT_TOKENS = "gen_ai.usage.input_tokens" + """ + The number of tokens in the input. + Example: 150 + """ + + GEN_AI_USAGE_INPUT_TOKENS_CACHED = "gen_ai.usage.input_tokens.cached" + """ + The number of cached tokens in the input. + Example: 50 + """ + + GEN_AI_USAGE_OUTPUT_TOKENS = "gen_ai.usage.output_tokens" + """ + The number of tokens in the output. + Example: 250 + """ + + GEN_AI_USAGE_OUTPUT_TOKENS_REASONING = "gen_ai.usage.output_tokens.reasoning" + """ + The number of tokens used for reasoning in the output. + Example: 75 + """ + + GEN_AI_USAGE_TOTAL_TOKENS = "gen_ai.usage.total_tokens" + """ + The total number of tokens used (input + output). + Example: 400 + """ + + GEN_AI_USER_MESSAGE = "gen_ai.user.message" + """ + The user message passed to the model. + Example: "What's the weather in Paris?" """ HTTP_FRAGMENT = "http.fragment" @@ -359,6 +516,12 @@ class SPANDATA: Example: GET """ + HTTP_QUERY = "http.query" + """ + The Query string present in the URL. + Example: ?foo=bar&bar=baz + """ + HTTP_STATUS_CODE = "http.response.status_code" """ The HTTP status code as an integer. @@ -376,14 +539,14 @@ class SPANDATA: The message's identifier. """ - MESSAGING_MESSAGE_RETRY_COUNT = "messaging.message.retry.count" + MESSAGING_MESSAGE_RECEIVE_LATENCY = "messaging.message.receive.latency" """ - Number of retries/attempts to process a message. + The latency between when the task was enqueued and when it was started to be processed. """ - MESSAGING_MESSAGE_RECEIVE_LATENCY = "messaging.message.receive.latency" + MESSAGING_MESSAGE_RETRY_COUNT = "messaging.message.retry.count" """ - The latency between when the task was enqueued and when it was started to be processed. + Number of retries/attempts to process a message. """ MESSAGING_SYSTEM = "messaging.system" @@ -391,6 +554,24 @@ class SPANDATA: The messaging system's name, e.g. `kafka`, `aws_sqs` """ + NETWORK_PEER_ADDRESS = "network.peer.address" + """ + Peer address of the network connection - IP address or Unix domain socket name. + Example: 10.1.2.80, /tmp/my.sock, localhost + """ + + NETWORK_PEER_PORT = "network.peer.port" + """ + Peer port number of the network connection. + Example: 6379 + """ + + PROFILER_ID = "profiler_id" + """ + Label identifying the profiler id that the span occurred in. This should be a string. + Example: "5249fbada8d5416482c2f6e47e337372" + """ + SERVER_ADDRESS = "server.address" """ Name of the database host. @@ -416,30 +597,6 @@ class SPANDATA: Example: 16456 """ - CODE_FILEPATH = "code.filepath" - """ - The source code file name that identifies the code unit as uniquely as possible (preferably an absolute file path). - Example: "/app/myapplication/http/handler/server.py" - """ - - CODE_LINENO = "code.lineno" - """ - The line number in `code.filepath` best representing the operation. It SHOULD point within the code unit named in `code.function`. - Example: 42 - """ - - CODE_FUNCTION = "code.function" - """ - The method or function name, or equivalent (usually rightmost part of the code unit's name). - Example: "server_request" - """ - - CODE_NAMESPACE = "code.namespace" - """ - The "namespace" within which `code.function` is defined. Usually the qualified class or module name, such that `code.namespace` + some separator + `code.function` form a unique identifier for the code unit. - Example: "http.handler" - """ - THREAD_ID = "thread.id" """ Identifier of a thread from where the span originated. This should be a string. @@ -452,12 +609,6 @@ class SPANDATA: Example: "MainThread" """ - PROFILER_ID = "profiler_id" - """ - Label identifying the profiler id that the span occurred in. This should be a string. - Example: "5249fbada8d5416482c2f6e47e337372" - """ - class SPANSTATUS: """ @@ -497,6 +648,10 @@ class OP: FUNCTION = "function" FUNCTION_AWS = "function.aws" FUNCTION_GCP = "function.gcp" + GEN_AI_CHAT = "gen_ai.chat" + GEN_AI_EXECUTE_TOOL = "gen_ai.execute_tool" + GEN_AI_HANDOFF = "gen_ai.handoff" + GEN_AI_INVOKE_AGENT = "gen_ai.invoke_agent" GRAPHQL_EXECUTE = "graphql.execute" GRAPHQL_MUTATION = "graphql.mutation" GRAPHQL_PARSE = "graphql.parse" diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 118289950c..e2eadd523d 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -145,6 +145,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "launchdarkly": (9, 8, 0), "loguru": (0, 7, 0), "openai": (1, 0, 0), + "openai_agents": (0, 0, 19), "openfeature": (0, 7, 1), "quart": (0, 16, 0), "ray": (2, 7, 0), diff --git a/sentry_sdk/integrations/openai_agents/__init__.py b/sentry_sdk/integrations/openai_agents/__init__.py new file mode 100644 index 0000000000..06b6459441 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/__init__.py @@ -0,0 +1,53 @@ +from sentry_sdk.integrations import DidNotEnable, Integration + +from .patches import ( + _create_get_model_wrapper, + _create_get_all_tools_wrapper, + _create_run_wrapper, + _patch_agent_run, +) + +try: + import agents + +except ImportError: + raise DidNotEnable("OpenAI Agents not installed") + + +def _patch_runner(): + # type: () -> None + # Create the root span for one full agent run (including eventual handoffs) + # Note agents.run.DEFAULT_AGENT_RUNNER.run_sync is a wrapper around + # agents.run.DEFAULT_AGENT_RUNNER.run. It does not need to be wrapped separately. + # TODO-anton: Also patch streaming runner: agents.Runner.run_streamed + agents.run.DEFAULT_AGENT_RUNNER.run = _create_run_wrapper( + agents.run.DEFAULT_AGENT_RUNNER.run + ) + + # Creating the actual spans for each agent run. + _patch_agent_run() + + +def _patch_model(): + # type: () -> None + agents.run.AgentRunner._get_model = classmethod( + _create_get_model_wrapper(agents.run.AgentRunner._get_model), + ) + + +def _patch_tools(): + # type: () -> None + agents.run.AgentRunner._get_all_tools = classmethod( + _create_get_all_tools_wrapper(agents.run.AgentRunner._get_all_tools), + ) + + +class OpenAIAgentsIntegration(Integration): + identifier = "openai_agents" + + @staticmethod + def setup_once(): + # type: () -> None + _patch_tools() + _patch_model() + _patch_runner() diff --git a/sentry_sdk/integrations/openai_agents/consts.py b/sentry_sdk/integrations/openai_agents/consts.py new file mode 100644 index 0000000000..f5de978be0 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/consts.py @@ -0,0 +1 @@ +SPAN_ORIGIN = "auto.ai.openai_agents" diff --git a/sentry_sdk/integrations/openai_agents/patches/__init__.py b/sentry_sdk/integrations/openai_agents/patches/__init__.py new file mode 100644 index 0000000000..06bb1711f8 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/__init__.py @@ -0,0 +1,4 @@ +from .models import _create_get_model_wrapper # noqa: F401 +from .tools import _create_get_all_tools_wrapper # noqa: F401 +from .runner import _create_run_wrapper # noqa: F401 +from .agent_run import _patch_agent_run # noqa: F401 diff --git a/sentry_sdk/integrations/openai_agents/patches/agent_run.py b/sentry_sdk/integrations/openai_agents/patches/agent_run.py new file mode 100644 index 0000000000..084100878c --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/agent_run.py @@ -0,0 +1,143 @@ +from functools import wraps + +from sentry_sdk.integrations import DidNotEnable + +from ..spans import invoke_agent_span, update_invoke_agent_span, handoff_span + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Optional + + +try: + import agents +except ImportError: + raise DidNotEnable("OpenAI Agents not installed") + + +def _patch_agent_run(): + # type: () -> None + """ + Patches AgentRunner methods to create agent invocation spans. + This directly patches the execution flow to track when agents start and stop. + """ + + # Store original methods + original_run_single_turn = agents.run.AgentRunner._run_single_turn + original_execute_handoffs = agents._run_impl.RunImpl.execute_handoffs + original_execute_final_output = agents._run_impl.RunImpl.execute_final_output + + def _start_invoke_agent_span(context_wrapper, agent): + # type: (agents.RunContextWrapper, agents.Agent) -> None + """Start an agent invocation span""" + # Store the agent on the context wrapper so we can access it later + context_wrapper._sentry_current_agent = agent + invoke_agent_span(context_wrapper, agent) + + def _end_invoke_agent_span(context_wrapper, agent, output=None): + # type: (agents.RunContextWrapper, agents.Agent, Optional[Any]) -> None + """End the agent invocation span""" + # Clear the stored agent + if hasattr(context_wrapper, "_sentry_current_agent"): + delattr(context_wrapper, "_sentry_current_agent") + + update_invoke_agent_span(context_wrapper, agent, output) + + def _has_active_agent_span(context_wrapper): + # type: (agents.RunContextWrapper) -> bool + """Check if there's an active agent span for this context""" + return getattr(context_wrapper, "_sentry_current_agent", None) is not None + + def _get_current_agent(context_wrapper): + # type: (agents.RunContextWrapper) -> Optional[agents.Agent] + """Get the current agent from context wrapper""" + return getattr(context_wrapper, "_sentry_current_agent", None) + + @wraps( + original_run_single_turn.__func__ + if hasattr(original_run_single_turn, "__func__") + else original_run_single_turn + ) + async def patched_run_single_turn(cls, *args, **kwargs): + # type: (agents.Runner, *Any, **Any) -> Any + """Patched _run_single_turn that creates agent invocation spans""" + + agent = kwargs.get("agent") + context_wrapper = kwargs.get("context_wrapper") + should_run_agent_start_hooks = kwargs.get("should_run_agent_start_hooks") + + # Start agent span when agent starts (but only once per agent) + if should_run_agent_start_hooks and agent and context_wrapper: + # End any existing span for a different agent + if _has_active_agent_span(context_wrapper): + current_agent = _get_current_agent(context_wrapper) + if current_agent and current_agent != agent: + _end_invoke_agent_span(context_wrapper, current_agent) + + _start_invoke_agent_span(context_wrapper, agent) + + # Call original method with all the correct parameters + result = await original_run_single_turn(*args, **kwargs) + + return result + + @wraps( + original_execute_handoffs.__func__ + if hasattr(original_execute_handoffs, "__func__") + else original_execute_handoffs + ) + async def patched_execute_handoffs(cls, *args, **kwargs): + # type: (agents.Runner, *Any, **Any) -> Any + """Patched execute_handoffs that creates handoff spans and ends agent span for handoffs""" + + context_wrapper = kwargs.get("context_wrapper") + run_handoffs = kwargs.get("run_handoffs") + agent = kwargs.get("agent") + + # Create Sentry handoff span for the first handoff (agents library only processes the first one) + if run_handoffs: + first_handoff = run_handoffs[0] + handoff_agent_name = first_handoff.handoff.agent_name + handoff_span(context_wrapper, agent, handoff_agent_name) + + # Call original method with all parameters + try: + result = await original_execute_handoffs(*args, **kwargs) + + finally: + # End span for current agent after handoff processing is complete + if agent and context_wrapper and _has_active_agent_span(context_wrapper): + _end_invoke_agent_span(context_wrapper, agent) + + return result + + @wraps( + original_execute_final_output.__func__ + if hasattr(original_execute_final_output, "__func__") + else original_execute_final_output + ) + async def patched_execute_final_output(cls, *args, **kwargs): + # type: (agents.Runner, *Any, **Any) -> Any + """Patched execute_final_output that ends agent span for final outputs""" + + agent = kwargs.get("agent") + context_wrapper = kwargs.get("context_wrapper") + final_output = kwargs.get("final_output") + + # Call original method with all parameters + try: + result = await original_execute_final_output(*args, **kwargs) + finally: + # End span for current agent after final output processing is complete + if agent and context_wrapper and _has_active_agent_span(context_wrapper): + _end_invoke_agent_span(context_wrapper, agent, final_output) + + return result + + # Apply patches + agents.run.AgentRunner._run_single_turn = classmethod(patched_run_single_turn) + agents._run_impl.RunImpl.execute_handoffs = classmethod(patched_execute_handoffs) + agents._run_impl.RunImpl.execute_final_output = classmethod( + patched_execute_final_output + ) diff --git a/sentry_sdk/integrations/openai_agents/patches/models.py b/sentry_sdk/integrations/openai_agents/patches/models.py new file mode 100644 index 0000000000..e6f24da6a1 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/models.py @@ -0,0 +1,50 @@ +from functools import wraps + +from sentry_sdk.integrations import DidNotEnable + +from ..spans import ai_client_span, update_ai_client_span + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable + + +try: + import agents +except ImportError: + raise DidNotEnable("OpenAI Agents not installed") + + +def _create_get_model_wrapper(original_get_model): + # type: (Callable[..., Any]) -> Callable[..., Any] + """ + Wraps the agents.Runner._get_model method to wrap the get_response method of the model to create a AI client span. + """ + + @wraps( + original_get_model.__func__ + if hasattr(original_get_model, "__func__") + else original_get_model + ) + def wrapped_get_model(cls, agent, run_config): + # type: (agents.Runner, agents.Agent, agents.RunConfig) -> agents.Model + + model = original_get_model(agent, run_config) + original_get_response = model.get_response + + @wraps(original_get_response) + async def wrapped_get_response(*args, **kwargs): + # type: (*Any, **Any) -> Any + with ai_client_span(agent, kwargs) as span: + result = await original_get_response(*args, **kwargs) + + update_ai_client_span(span, agent, kwargs, result) + + return result + + model.get_response = wrapped_get_response + + return model + + return wrapped_get_model diff --git a/sentry_sdk/integrations/openai_agents/patches/runner.py b/sentry_sdk/integrations/openai_agents/patches/runner.py new file mode 100644 index 0000000000..e1e9a3b50c --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/runner.py @@ -0,0 +1,42 @@ +from functools import wraps + +import sentry_sdk + +from ..spans import agent_workflow_span +from ..utils import _capture_exception + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable + + +def _create_run_wrapper(original_func): + # type: (Callable[..., Any]) -> Callable[..., Any] + """ + Wraps the agents.Runner.run methods to create a root span for the agent workflow runs. + + Note agents.Runner.run_sync() is a wrapper around agents.Runner.run(), + so it does not need to be wrapped separately. + """ + + @wraps(original_func) + async def wrapper(*args, **kwargs): + # type: (*Any, **Any) -> Any + agent = args[0] + with agent_workflow_span(agent): + result = None + try: + result = await original_func(*args, **kwargs) + return result + except Exception as exc: + _capture_exception(exc) + + # It could be that there is a "invoke agent" span still open + current_span = sentry_sdk.get_current_span() + if current_span is not None and current_span.timestamp is None: + current_span.__exit__(None, None, None) + + raise exc from None + + return wrapper diff --git a/sentry_sdk/integrations/openai_agents/patches/tools.py b/sentry_sdk/integrations/openai_agents/patches/tools.py new file mode 100644 index 0000000000..b359d32678 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/tools.py @@ -0,0 +1,77 @@ +from functools import wraps + +from sentry_sdk.integrations import DidNotEnable + +from ..spans import execute_tool_span, update_execute_tool_span + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable + +try: + import agents +except ImportError: + raise DidNotEnable("OpenAI Agents not installed") + + +def _create_get_all_tools_wrapper(original_get_all_tools): + # type: (Callable[..., Any]) -> Callable[..., Any] + """ + Wraps the agents.Runner._get_all_tools method of the Runner class to wrap all function tools with Sentry instrumentation. + """ + + @wraps( + original_get_all_tools.__func__ + if hasattr(original_get_all_tools, "__func__") + else original_get_all_tools + ) + async def wrapped_get_all_tools(cls, agent, context_wrapper): + # type: (agents.Runner, agents.Agent, agents.RunContextWrapper) -> list[agents.Tool] + + # Get the original tools + tools = await original_get_all_tools(agent, context_wrapper) + + wrapped_tools = [] + for tool in tools: + # Wrap only the function tools (for now) + if tool.__class__.__name__ != "FunctionTool": + wrapped_tools.append(tool) + continue + + # Create a new FunctionTool with our wrapped invoke method + original_on_invoke = tool.on_invoke_tool + + def create_wrapped_invoke(current_tool, current_on_invoke): + # type: (agents.Tool, Callable[..., Any]) -> Callable[..., Any] + @wraps(current_on_invoke) + async def sentry_wrapped_on_invoke_tool(*args, **kwargs): + # type: (*Any, **Any) -> Any + with execute_tool_span(current_tool, *args, **kwargs) as span: + # We can not capture exceptions in tool execution here because + # `_on_invoke_tool` is swallowing the exception here: + # https://github.com/openai/openai-agents-python/blob/main/src/agents/tool.py#L409-L422 + # And because function_tool is a decorator with `default_tool_error_function` set as a default parameter + # I was unable to monkey patch it because those are evaluated at module import time + # and the SDK is too late to patch it. I was also unable to patch `_on_invoke_tool_impl` + # because it is nested inside this import time code. As if they made it hard to patch on purpose... + result = await current_on_invoke(*args, **kwargs) + update_execute_tool_span(span, agent, current_tool, result) + + return result + + return sentry_wrapped_on_invoke_tool + + wrapped_tool = agents.FunctionTool( + name=tool.name, + description=tool.description, + params_json_schema=tool.params_json_schema, + on_invoke_tool=create_wrapped_invoke(tool, original_on_invoke), + strict_json_schema=tool.strict_json_schema, + is_enabled=tool.is_enabled, + ) + wrapped_tools.append(wrapped_tool) + + return wrapped_tools + + return wrapped_get_all_tools diff --git a/sentry_sdk/integrations/openai_agents/spans/__init__.py b/sentry_sdk/integrations/openai_agents/spans/__init__.py new file mode 100644 index 0000000000..3bc453cafa --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/__init__.py @@ -0,0 +1,5 @@ +from .agent_workflow import agent_workflow_span # noqa: F401 +from .ai_client import ai_client_span, update_ai_client_span # noqa: F401 +from .execute_tool import execute_tool_span, update_execute_tool_span # noqa: F401 +from .handoff import handoff_span # noqa: F401 +from .invoke_agent import invoke_agent_span, update_invoke_agent_span # noqa: F401 diff --git a/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py b/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py new file mode 100644 index 0000000000..de2f28d41e --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py @@ -0,0 +1,21 @@ +import sentry_sdk + +from ..consts import SPAN_ORIGIN +from ..utils import _get_start_span_function + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + import agents + + +def agent_workflow_span(agent): + # type: (agents.Agent) -> sentry_sdk.tracing.Span + + # Create a transaction or a span if an transaction is already active + span = _get_start_span_function()( + name=f"{agent.name} workflow", + origin=SPAN_ORIGIN, + ) + + return span diff --git a/sentry_sdk/integrations/openai_agents/spans/ai_client.py b/sentry_sdk/integrations/openai_agents/spans/ai_client.py new file mode 100644 index 0000000000..30c5fd1dac --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/ai_client.py @@ -0,0 +1,38 @@ +import sentry_sdk +from sentry_sdk.consts import OP, SPANDATA + +from ..consts import SPAN_ORIGIN +from ..utils import ( + _set_agent_data, + _set_input_data, + _set_output_data, + _set_usage_data, +) + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from agents import Agent + from typing import Any + + +def ai_client_span(agent, get_response_kwargs): + # type: (Agent, dict[str, Any]) -> sentry_sdk.tracing.Span + # TODO-anton: implement other types of operations. Now "chat" is hardcoded. + span = sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + description=f"chat {agent.model}", + origin=SPAN_ORIGIN, + ) + # TODO-anton: remove hardcoded stuff and replace something that also works for embedding and so on + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + + return span + + +def update_ai_client_span(span, agent, get_response_kwargs, result): + # type: (sentry_sdk.tracing.Span, Agent, dict[str, Any], Any) -> None + _set_agent_data(span, agent) + _set_usage_data(span, result.usage) + _set_input_data(span, get_response_kwargs) + _set_output_data(span, result) diff --git a/sentry_sdk/integrations/openai_agents/spans/execute_tool.py b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py new file mode 100644 index 0000000000..e6e880b64c --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py @@ -0,0 +1,43 @@ +import sentry_sdk +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.scope import should_send_default_pii + +from ..consts import SPAN_ORIGIN +from ..utils import _set_agent_data + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + import agents + from typing import Any + + +def execute_tool_span(tool, *args, **kwargs): + # type: (agents.Tool, *Any, **Any) -> sentry_sdk.tracing.Span + span = sentry_sdk.start_span( + op=OP.GEN_AI_EXECUTE_TOOL, + name=f"execute_tool {tool.name}", + origin=SPAN_ORIGIN, + ) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "execute_tool") + + if tool.__class__.__name__ == "FunctionTool": + span.set_data(SPANDATA.GEN_AI_TOOL_TYPE, "function") + + span.set_data(SPANDATA.GEN_AI_TOOL_NAME, tool.name) + span.set_data(SPANDATA.GEN_AI_TOOL_DESCRIPTION, tool.description) + + if should_send_default_pii(): + input = args[1] + span.set_data(SPANDATA.GEN_AI_TOOL_INPUT, input) + + return span + + +def update_execute_tool_span(span, agent, tool, result): + # type: (sentry_sdk.tracing.Span, agents.Agent, agents.Tool, Any) -> None + _set_agent_data(span, agent) + + if should_send_default_pii(): + span.set_data(SPANDATA.GEN_AI_TOOL_OUTPUT, result) diff --git a/sentry_sdk/integrations/openai_agents/spans/handoff.py b/sentry_sdk/integrations/openai_agents/spans/handoff.py new file mode 100644 index 0000000000..78e6788c7d --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/handoff.py @@ -0,0 +1,19 @@ +import sentry_sdk +from sentry_sdk.consts import OP, SPANDATA + +from ..consts import SPAN_ORIGIN + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + import agents + + +def handoff_span(context, from_agent, to_agent_name): + # type: (agents.RunContextWrapper, agents.Agent, str) -> None + with sentry_sdk.start_span( + op=OP.GEN_AI_HANDOFF, + name=f"handoff from {from_agent.name} to {to_agent_name}", + origin=SPAN_ORIGIN, + ) as span: + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "handoff") diff --git a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py new file mode 100644 index 0000000000..549ade1246 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py @@ -0,0 +1,34 @@ +import sentry_sdk +from sentry_sdk.consts import OP, SPANDATA + +from ..consts import SPAN_ORIGIN +from ..utils import _set_agent_data + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + import agents + from typing import Any + + +def invoke_agent_span(context, agent): + # type: (agents.RunContextWrapper, agents.Agent) -> sentry_sdk.tracing.Span + span = sentry_sdk.start_span( + op=OP.GEN_AI_INVOKE_AGENT, + name=f"invoke_agent {agent.name}", + origin=SPAN_ORIGIN, + ) + span.__enter__() + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + + _set_agent_data(span, agent) + + return span + + +def update_invoke_agent_span(context, agent, output): + # type: (agents.RunContextWrapper, agents.Agent, Any) -> None + current_span = sentry_sdk.get_current_span() + if current_span: + current_span.__exit__(None, None, None) diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py new file mode 100644 index 0000000000..28dbd6bb75 --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -0,0 +1,209 @@ +import json +import sentry_sdk +from sentry_sdk.consts import SPANDATA +from sentry_sdk.integrations import DidNotEnable +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import event_from_exception + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any + from typing import Callable + from typing import Union + from agents import Usage + +try: + import agents + +except ImportError: + raise DidNotEnable("OpenAI Agents not installed") + + +def _capture_exception(exc): + # type: (Any) -> None + event, hint = event_from_exception( + exc, + client_options=sentry_sdk.get_client().options, + mechanism={"type": "openai_agents", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + +def _get_start_span_function(): + # type: () -> Callable[..., Any] + current_span = sentry_sdk.get_current_span() + transaction_exists = ( + current_span is not None and current_span.containing_transaction == current_span + ) + return sentry_sdk.start_span if transaction_exists else sentry_sdk.start_transaction + + +def _set_agent_data(span, agent): + # type: (sentry_sdk.tracing.Span, agents.Agent) -> None + span.set_data( + SPANDATA.GEN_AI_SYSTEM, "openai" + ) # See footnote for https://opentelemetry.io/docs/specs/semconv/registry/attributes/gen-ai/#gen-ai-system for explanation why. + + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, agent.name) + + if agent.model_settings.max_tokens: + span.set_data( + SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, agent.model_settings.max_tokens + ) + + if agent.model: + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, agent.model) + + if agent.model_settings.presence_penalty: + span.set_data( + SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + agent.model_settings.presence_penalty, + ) + + if agent.model_settings.temperature: + span.set_data( + SPANDATA.GEN_AI_REQUEST_TEMPERATURE, agent.model_settings.temperature + ) + + if agent.model_settings.top_p: + span.set_data(SPANDATA.GEN_AI_REQUEST_TOP_P, agent.model_settings.top_p) + + if agent.model_settings.frequency_penalty: + span.set_data( + SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + agent.model_settings.frequency_penalty, + ) + + if len(agent.tools) > 0: + span.set_data( + SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, + safe_serialize([vars(tool) for tool in agent.tools]), + ) + + +def _set_usage_data(span, usage): + # type: (sentry_sdk.tracing.Span, Usage) -> None + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, usage.input_tokens) + span.set_data( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED, + usage.input_tokens_details.cached_tokens, + ) + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, usage.output_tokens) + span.set_data( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING, + usage.output_tokens_details.reasoning_tokens, + ) + span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, usage.total_tokens) + + +def _set_input_data(span, get_response_kwargs): + # type: (sentry_sdk.tracing.Span, dict[str, Any]) -> None + if not should_send_default_pii(): + return + + messages_by_role = { + "system": [], + "user": [], + "assistant": [], + "tool": [], + } # type: (dict[str, list[Any]]) + system_instructions = get_response_kwargs.get("system_instructions") + if system_instructions: + messages_by_role["system"].append({"type": "text", "text": system_instructions}) + + for message in get_response_kwargs.get("input", []): + if "role" in message: + messages_by_role[message.get("role")].append( + {"type": "text", "text": message.get("content")} + ) + else: + if message.get("type") == "function_call": + messages_by_role["assistant"].append(message) + elif message.get("type") == "function_call_output": + messages_by_role["tool"].append(message) + + request_messages = [] + for role, messages in messages_by_role.items(): + if len(messages) > 0: + request_messages.append({"role": role, "content": messages}) + + span.set_data(SPANDATA.GEN_AI_REQUEST_MESSAGES, safe_serialize(request_messages)) + + +def _set_output_data(span, result): + # type: (sentry_sdk.tracing.Span, Any) -> None + if not should_send_default_pii(): + return + + output_messages = { + "response": [], + "tool": [], + } # type: (dict[str, list[Any]]) + + for output in result.output: + if output.type == "function_call": + output_messages["tool"].append(output.dict()) + elif output.type == "message": + for output_message in output.content: + try: + output_messages["response"].append(output_message.text) + except AttributeError: + # Unknown output message type, just return the json + output_messages["response"].append(output_message.dict()) + + if len(output_messages["tool"]) > 0: + span.set_data( + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, safe_serialize(output_messages["tool"]) + ) + + if len(output_messages["response"]) > 0: + span.set_data( + SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(output_messages["response"]) + ) + + +def safe_serialize(data): + # type: (Any) -> str + """Safely serialize to a readable string.""" + + def serialize_item(item): + # type: (Any) -> Union[str, dict[Any, Any], list[Any], tuple[Any, ...]] + if callable(item): + try: + module = getattr(item, "__module__", None) + qualname = getattr(item, "__qualname__", None) + name = getattr(item, "__name__", "anonymous") + + if module and qualname: + full_path = f"{module}.{qualname}" + elif module and name: + full_path = f"{module}.{name}" + else: + full_path = name + + return f"" + except Exception: + return f"" + elif isinstance(item, dict): + return {k: serialize_item(v) for k, v in item.items()} + elif isinstance(item, (list, tuple)): + return [serialize_item(x) for x in item] + elif hasattr(item, "__dict__"): + try: + attrs = { + k: serialize_item(v) + for k, v in vars(item).items() + if not k.startswith("_") + } + return f"<{type(item).__name__} {attrs}>" + except Exception: + return repr(item) + else: + return item + + try: + serialized = serialize_item(data) + return json.dumps(serialized, default=str) + except Exception: + return str(data) diff --git a/tests/integrations/openai_agents/__init__.py b/tests/integrations/openai_agents/__init__.py new file mode 100644 index 0000000000..6940e2bbbe --- /dev/null +++ b/tests/integrations/openai_agents/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("agents") diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py new file mode 100644 index 0000000000..ec606c8806 --- /dev/null +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -0,0 +1,580 @@ +import re +import pytest +from unittest.mock import MagicMock, patch +import os + +from sentry_sdk.integrations.openai_agents import OpenAIAgentsIntegration +from sentry_sdk.integrations.openai_agents.utils import safe_serialize + +import agents +from agents import ( + Agent, + ModelResponse, + Usage, + ModelSettings, +) +from agents.items import ( + ResponseOutputMessage, + ResponseOutputText, + ResponseFunctionToolCall, +) + +test_run_config = agents.RunConfig(tracing_disabled=True) + + +@pytest.fixture +def mock_usage(): + return Usage( + requests=1, + input_tokens=10, + output_tokens=20, + total_tokens=30, + input_tokens_details=MagicMock(cached_tokens=0), + output_tokens_details=MagicMock(reasoning_tokens=5), + ) + + +@pytest.fixture +def mock_model_response(mock_usage): + return ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_123", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="Hello, how can I help you?", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=mock_usage, + response_id="resp_123", + ) + + +@pytest.fixture +def test_agent(): + """Create a real Agent instance for testing.""" + return Agent( + name="test_agent", + instructions="You are a helpful test assistant.", + model="gpt-4", + model_settings=ModelSettings( + max_tokens=100, + temperature=0.7, + top_p=1.0, + presence_penalty=0.0, + frequency_penalty=0.0, + ), + ) + + +@pytest.mark.asyncio +async def test_agent_invocation_span( + sentry_init, capture_events, test_agent, mock_model_response +): + """ + Test that the integration creates spans for agent invocations. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.return_value = mock_model_response + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + result = await agents.Runner.run( + test_agent, "Test input", run_config=test_run_config + ) + + assert result is not None + assert result.final_output == "Hello, how can I help you?" + + (transaction,) = events + spans = transaction["spans"] + invoke_agent_span, ai_client_span = spans + + assert transaction["transaction"] == "test_agent workflow" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" + + assert invoke_agent_span["description"] == "invoke_agent test_agent" + assert invoke_agent_span["data"]["gen_ai.operation.name"] == "invoke_agent" + assert invoke_agent_span["data"]["gen_ai.system"] == "openai" + assert invoke_agent_span["data"]["gen_ai.agent.name"] == "test_agent" + assert invoke_agent_span["data"]["gen_ai.request.max_tokens"] == 100 + assert invoke_agent_span["data"]["gen_ai.request.model"] == "gpt-4" + assert invoke_agent_span["data"]["gen_ai.request.temperature"] == 0.7 + assert invoke_agent_span["data"]["gen_ai.request.top_p"] == 1.0 + + assert ai_client_span["description"] == "chat gpt-4" + assert ai_client_span["data"]["gen_ai.operation.name"] == "chat" + assert ai_client_span["data"]["gen_ai.system"] == "openai" + assert ai_client_span["data"]["gen_ai.agent.name"] == "test_agent" + assert ai_client_span["data"]["gen_ai.request.max_tokens"] == 100 + assert ai_client_span["data"]["gen_ai.request.model"] == "gpt-4" + assert ai_client_span["data"]["gen_ai.request.temperature"] == 0.7 + assert ai_client_span["data"]["gen_ai.request.top_p"] == 1.0 + + +def test_agent_invocation_span_sync( + sentry_init, capture_events, test_agent, mock_model_response +): + """ + Test that the integration creates spans for agent invocations. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.return_value = mock_model_response + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + result = agents.Runner.run_sync( + test_agent, "Test input", run_config=test_run_config + ) + + assert result is not None + assert result.final_output == "Hello, how can I help you?" + + (transaction,) = events + spans = transaction["spans"] + invoke_agent_span, ai_client_span = spans + + assert transaction["transaction"] == "test_agent workflow" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" + + assert invoke_agent_span["description"] == "invoke_agent test_agent" + assert invoke_agent_span["data"]["gen_ai.operation.name"] == "invoke_agent" + assert invoke_agent_span["data"]["gen_ai.system"] == "openai" + assert invoke_agent_span["data"]["gen_ai.agent.name"] == "test_agent" + assert invoke_agent_span["data"]["gen_ai.request.max_tokens"] == 100 + assert invoke_agent_span["data"]["gen_ai.request.model"] == "gpt-4" + assert invoke_agent_span["data"]["gen_ai.request.temperature"] == 0.7 + assert invoke_agent_span["data"]["gen_ai.request.top_p"] == 1.0 + + assert ai_client_span["description"] == "chat gpt-4" + assert ai_client_span["data"]["gen_ai.operation.name"] == "chat" + assert ai_client_span["data"]["gen_ai.system"] == "openai" + assert ai_client_span["data"]["gen_ai.agent.name"] == "test_agent" + assert ai_client_span["data"]["gen_ai.request.max_tokens"] == 100 + assert ai_client_span["data"]["gen_ai.request.model"] == "gpt-4" + assert ai_client_span["data"]["gen_ai.request.temperature"] == 0.7 + assert ai_client_span["data"]["gen_ai.request.top_p"] == 1.0 + + +@pytest.mark.asyncio +async def test_handoff_span(sentry_init, capture_events, mock_usage): + """ + Test that handoff spans are created when agents hand off to other agents. + """ + # Create two simple agents with a handoff relationship + secondary_agent = agents.Agent( + name="secondary_agent", + instructions="You are a secondary agent.", + model="gpt-4o-mini", + ) + + primary_agent = agents.Agent( + name="primary_agent", + instructions="You are a primary agent that hands off to secondary agent.", + model="gpt-4o-mini", + handoffs=[secondary_agent], + ) + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Mock two responses: + # 1. Primary agent calls handoff tool + # 2. Secondary agent provides final response + handoff_response = ModelResponse( + output=[ + ResponseFunctionToolCall( + id="call_handoff_123", + call_id="call_handoff_123", + name="transfer_to_secondary_agent", + type="function_call", + arguments="{}", + function=MagicMock( + name="transfer_to_secondary_agent", arguments="{}" + ), + ) + ], + usage=mock_usage, + response_id="resp_handoff_123", + ) + + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="I'm the specialist and I can help with that!", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=mock_usage, + response_id="resp_final_123", + ) + + mock_get_response.side_effect = [handoff_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + result = await agents.Runner.run( + primary_agent, + "Please hand off to secondary agent", + run_config=test_run_config, + ) + + assert result is not None + + (transaction,) = events + spans = transaction["spans"] + handoff_span = spans[2] + + # Verify handoff span was created + assert handoff_span is not None + assert ( + handoff_span["description"] == "handoff from primary_agent to secondary_agent" + ) + assert handoff_span["data"]["gen_ai.operation.name"] == "handoff" + + +@pytest.mark.asyncio +async def test_tool_execution_span(sentry_init, capture_events, test_agent): + """ + Test tool execution span creation. + """ + + @agents.function_tool + def simple_test_tool(message: str) -> str: + """A simple tool""" + return f"Tool executed with: {message}" + + # Create agent with the tool + agent_with_tool = test_agent.clone(tools=[simple_test_tool]) + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Create a mock response that includes tool calls + tool_call = ResponseFunctionToolCall( + id="call_123", + call_id="call_123", + name="simple_test_tool", + type="function_call", + arguments='{"message": "hello"}', + function=MagicMock( + name="simple_test_tool", arguments='{"message": "hello"}' + ), + ) + + # First response with tool call + tool_response = ModelResponse( + output=[tool_call], + usage=Usage( + requests=1, input_tokens=10, output_tokens=5, total_tokens=15 + ), + response_id="resp_tool_123", + ) + + # Second response with final answer + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="Task completed using the tool", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=Usage( + requests=1, input_tokens=15, output_tokens=10, total_tokens=25 + ), + response_id="resp_final_123", + ) + + # Return different responses on successive calls + mock_get_response.side_effect = [tool_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await agents.Runner.run( + agent_with_tool, + "Please use the simple test tool", + run_config=test_run_config, + ) + + (transaction,) = events + spans = transaction["spans"] + ( + agent_span, + ai_client_span1, + tool_span, + ai_client_span2, + ) = spans + + available_tools = safe_serialize( + [ + { + "name": "simple_test_tool", + "description": "A simple tool", + "params_json_schema": { + "properties": {"message": {"title": "Message", "type": "string"}}, + "required": ["message"], + "title": "simple_test_tool_args", + "type": "object", + "additionalProperties": False, + }, + "on_invoke_tool": "._create_function_tool.._on_invoke_tool>", + "strict_json_schema": True, + "is_enabled": True, + } + ] + ) + + assert transaction["transaction"] == "test_agent workflow" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" + + assert agent_span["description"] == "invoke_agent test_agent" + assert agent_span["origin"] == "auto.ai.openai_agents" + assert agent_span["data"]["gen_ai.agent.name"] == "test_agent" + assert agent_span["data"]["gen_ai.operation.name"] == "invoke_agent" + assert agent_span["data"]["gen_ai.request.available_tools"] == available_tools + assert agent_span["data"]["gen_ai.request.max_tokens"] == 100 + assert agent_span["data"]["gen_ai.request.model"] == "gpt-4" + assert agent_span["data"]["gen_ai.request.temperature"] == 0.7 + assert agent_span["data"]["gen_ai.request.top_p"] == 1.0 + assert agent_span["data"]["gen_ai.system"] == "openai" + + assert ai_client_span1["description"] == "chat gpt-4" + assert ai_client_span1["data"]["gen_ai.operation.name"] == "chat" + assert ai_client_span1["data"]["gen_ai.system"] == "openai" + assert ai_client_span1["data"]["gen_ai.agent.name"] == "test_agent" + assert ai_client_span1["data"]["gen_ai.request.available_tools"] == available_tools + assert ai_client_span1["data"]["gen_ai.request.max_tokens"] == 100 + assert ai_client_span1["data"]["gen_ai.request.messages"] == safe_serialize( + [ + { + "role": "system", + "content": [ + {"type": "text", "text": "You are a helpful test assistant."} + ], + }, + { + "role": "user", + "content": [ + {"type": "text", "text": "Please use the simple test tool"} + ], + }, + ] + ) + assert ai_client_span1["data"]["gen_ai.request.model"] == "gpt-4" + assert ai_client_span1["data"]["gen_ai.request.temperature"] == 0.7 + assert ai_client_span1["data"]["gen_ai.request.top_p"] == 1.0 + assert ai_client_span1["data"]["gen_ai.usage.input_tokens"] == 10 + assert ai_client_span1["data"]["gen_ai.usage.input_tokens.cached"] == 0 + assert ai_client_span1["data"]["gen_ai.usage.output_tokens"] == 5 + assert ai_client_span1["data"]["gen_ai.usage.output_tokens.reasoning"] == 0 + assert ai_client_span1["data"]["gen_ai.usage.total_tokens"] == 15 + assert re.sub( + r"SerializationIterator\(.*\)", + "NOT_CHECKED", + ai_client_span1["data"]["gen_ai.response.tool_calls"], + ) == safe_serialize( + [ + { + "arguments": '{"message": "hello"}', + "call_id": "call_123", + "name": "simple_test_tool", + "type": "function_call", + "id": "call_123", + "status": None, + "function": "NOT_CHECKED", + } + ] + ) + + assert tool_span["description"] == "execute_tool simple_test_tool" + assert tool_span["data"]["gen_ai.agent.name"] == "test_agent" + assert tool_span["data"]["gen_ai.operation.name"] == "execute_tool" + assert ( + re.sub( + "<.*>(,)", + r"'NOT_CHECKED'\1", + agent_span["data"]["gen_ai.request.available_tools"], + ) + == available_tools + ) + assert tool_span["data"]["gen_ai.request.max_tokens"] == 100 + assert tool_span["data"]["gen_ai.request.model"] == "gpt-4" + assert tool_span["data"]["gen_ai.request.temperature"] == 0.7 + assert tool_span["data"]["gen_ai.request.top_p"] == 1.0 + assert tool_span["data"]["gen_ai.system"] == "openai" + assert tool_span["data"]["gen_ai.tool.description"] == "A simple tool" + assert tool_span["data"]["gen_ai.tool.input"] == '{"message": "hello"}' + assert tool_span["data"]["gen_ai.tool.name"] == "simple_test_tool" + assert tool_span["data"]["gen_ai.tool.output"] == "Tool executed with: hello" + assert tool_span["data"]["gen_ai.tool.type"] == "function" + + assert ai_client_span2["description"] == "chat gpt-4" + assert ai_client_span2["data"]["gen_ai.agent.name"] == "test_agent" + assert ai_client_span2["data"]["gen_ai.operation.name"] == "chat" + assert ( + re.sub( + "<.*>(,)", + r"'NOT_CHECKED'\1", + agent_span["data"]["gen_ai.request.available_tools"], + ) + == available_tools + ) + assert ai_client_span2["data"]["gen_ai.request.max_tokens"] == 100 + assert re.sub( + r"SerializationIterator\(.*\)", + "NOT_CHECKED", + ai_client_span2["data"]["gen_ai.request.messages"], + ) == safe_serialize( + [ + { + "role": "system", + "content": [ + {"type": "text", "text": "You are a helpful test assistant."} + ], + }, + { + "role": "user", + "content": [ + {"type": "text", "text": "Please use the simple test tool"} + ], + }, + { + "role": "assistant", + "content": [ + { + "arguments": '{"message": "hello"}', + "call_id": "call_123", + "name": "simple_test_tool", + "type": "function_call", + "id": "call_123", + "function": "NOT_CHECKED", + } + ], + }, + { + "role": "tool", + "content": [ + { + "call_id": "call_123", + "output": "Tool executed with: hello", + "type": "function_call_output", + } + ], + }, + ] + ) + assert ai_client_span2["data"]["gen_ai.request.model"] == "gpt-4" + assert ai_client_span2["data"]["gen_ai.request.temperature"] == 0.7 + assert ai_client_span2["data"]["gen_ai.request.top_p"] == 1.0 + assert ai_client_span2["data"]["gen_ai.response.text"] == safe_serialize( + ["Task completed using the tool"] + ) + assert ai_client_span2["data"]["gen_ai.system"] == "openai" + assert ai_client_span2["data"]["gen_ai.usage.input_tokens.cached"] == 0 + assert ai_client_span2["data"]["gen_ai.usage.input_tokens"] == 15 + assert ai_client_span2["data"]["gen_ai.usage.output_tokens.reasoning"] == 0 + assert ai_client_span2["data"]["gen_ai.usage.output_tokens"] == 10 + assert ai_client_span2["data"]["gen_ai.usage.total_tokens"] == 25 + + +@pytest.mark.asyncio +async def test_error_handling(sentry_init, capture_events, test_agent): + """ + Test error handling in agent execution. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.side_effect = Exception("Model Error") + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + with pytest.raises(Exception, match="Model Error"): + await agents.Runner.run( + test_agent, "Test input", run_config=test_run_config + ) + + ( + error_event, + transaction, + ) = events + + assert error_event["exception"]["values"][0]["type"] == "Exception" + assert error_event["exception"]["values"][0]["value"] == "Model Error" + assert error_event["exception"]["values"][0]["mechanism"]["type"] == "openai_agents" + + spans = transaction["spans"] + (invoke_agent_span, ai_client_span) = spans + + assert transaction["transaction"] == "test_agent workflow" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" + + assert invoke_agent_span["description"] == "invoke_agent test_agent" + assert invoke_agent_span["origin"] == "auto.ai.openai_agents" + + assert ai_client_span["description"] == "chat gpt-4" + assert ai_client_span["origin"] == "auto.ai.openai_agents" + assert ai_client_span["tags"]["status"] == "internal_error" diff --git a/tox.ini b/tox.ini index f4aee13d02..5c993718d7 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-06-24T07:19:36.122984+00:00 +# Last generated: 2025-06-24T12:35:34.437673+00:00 [tox] requires = @@ -145,6 +145,8 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.11.4 {py3.9,py3.11,py3.12}-cohere-v5.15.0 + {py3.9,py3.11,py3.12}-openai_agents-v0.0.19 + {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 @@ -515,6 +517,9 @@ deps = cohere-v5.11.4: cohere==5.11.4 cohere-v5.15.0: cohere==5.15.0 + openai_agents-v0.0.19: openai-agents==0.0.19 + openai_agents: pytest-asyncio + huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 @@ -809,6 +814,7 @@ setenv = litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru openai: TESTPATH=tests/integrations/openai + openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry potel: TESTPATH=tests/integrations/opentelemetry From 15f1348ffd2ea5f1a098df56d1864d7faa9117e1 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 24 Jun 2025 13:55:38 +0000 Subject: [PATCH 330/591] release: 2.31.0 --- CHANGELOG.md | 17 +++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 20 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index ddeed9d687..80ef5dc6ea 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,22 @@ # Changelog +## 2.31.0 + +### Various fixes & improvements + +- Support `openai-agents` (#4437) by @antonpirker +- ref(langchain): Greatly simplify `_wrap_configure` (#4479) by @szokeasaurusrex +- tests: Tox update (#4509) by @sentrivana +- Cursor generated rules (#4493) by @sl0thentr0py +- fix(ci): Remove tracerite pin (almost) (#4504) by @sentrivana +- fix(profiling): Ensure profiler thread exits when needed (#4497) by @Zylphrex +- fix(ci): Do not install newest tracerite (#4494) by @sentrivana +- tests: Regenerate tox (#4484) by @sentrivana +- fix(scope): Handle token reset `LookupError`s gracefully (#4481) by @sentrivana +- tests: Upper bound on fakeredis on old Python versions (#4482) by @sentrivana +- feat(logs): Add support for dict args (#4478) by @AbhiPrasad +- tests: Regenerate tox (#4457) by @sentrivana + ## 2.30.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 4e12abf550..01b40ae828 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.30.0" +release = "2.31.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 53148a36df..7102eea0e7 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1181,4 +1181,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.30.0" +VERSION = "2.31.0" diff --git a/setup.py b/setup.py index ecb5dfa994..0662be384e 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.30.0", + version="2.31.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 9792e4f4ac0a529a42726b01f8c78f5fd3e218a5 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 24 Jun 2025 16:18:34 +0200 Subject: [PATCH 331/591] Updated changelog --- CHANGELOG.md | 41 +++++++++++++++++++++++++++++------------ 1 file changed, 29 insertions(+), 12 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 80ef5dc6ea..8bcd8ddc73 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,18 +4,35 @@ ### Various fixes & improvements -- Support `openai-agents` (#4437) by @antonpirker -- ref(langchain): Greatly simplify `_wrap_configure` (#4479) by @szokeasaurusrex -- tests: Tox update (#4509) by @sentrivana -- Cursor generated rules (#4493) by @sl0thentr0py -- fix(ci): Remove tracerite pin (almost) (#4504) by @sentrivana -- fix(profiling): Ensure profiler thread exits when needed (#4497) by @Zylphrex -- fix(ci): Do not install newest tracerite (#4494) by @sentrivana -- tests: Regenerate tox (#4484) by @sentrivana -- fix(scope): Handle token reset `LookupError`s gracefully (#4481) by @sentrivana -- tests: Upper bound on fakeredis on old Python versions (#4482) by @sentrivana -- feat(logs): Add support for dict args (#4478) by @AbhiPrasad -- tests: Regenerate tox (#4457) by @sentrivana +- **New Integration (BETA):** Add support for `openai-agents` (#4437) by @antonpirker + + We can now instrument AI agents that are created with the [OpenAI Agents SDK](https://openai.github.io/openai-agents-python/) out of the box. + +```python +import sentry_sdk +from sentry_sdk.integrations.openai_agents import OpenAIAgentsIntegration + +# Add the OpenAIAgentsIntegration to your sentry_sdk.init call: +sentry_sdk.init( + dsn="...", + integrations=[ + OpenAIAgentsIntegration(), + ] +) +``` + +For more information see the [OpenAI Agents integrations documentation](https://docs.sentry.io/platforms/python/integrations/openai-agents/). + +- Logs: Add support for `dict` arguments (#4478) by @AbhiPrasad +- Add Cursor generated rules (#4493) by @sl0thentr0py +- Greatly simplify Langchain integrations `_wrap_configure` (#4479) by @szokeasaurusrex +- Fix(ci): Remove tracerite pin (almost) (#4504) by @sentrivana +- Fix(profiling): Ensure profiler thread exits when needed (#4497) by @Zylphrex +- Fix(ci): Do not install newest `tracerite` (#4494) by @sentrivana +- Fix(scope): Handle token reset `LookupError`s gracefully (#4481) by @sentrivana +- Tests: Tox update (#4509) by @sentrivana +- Tests: Upper bound on fakeredis on old Python versions (#4482) by @sentrivana +- Tests: Regenerate tox (#4457) by @sentrivana ## 2.30.0 From 8b6e5adfc6a55c40b7e64b099ae51c2cdb244031 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 24 Jun 2025 17:30:30 +0200 Subject: [PATCH 332/591] feat(sessions): Add top-level start- and end session methods (#4474) Closes #4473. Co-authored-by: Cursor Agent --- sentry_sdk/__init__.py | 2 ++ sentry_sdk/api.py | 16 ++++++++++++++ tests/test_sessions.py | 49 ++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 67 insertions(+) diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index 9fd7253fc2..e03f3b4484 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -47,6 +47,8 @@ "trace", "monitor", "logger", + "start_session", + "end_session", ] # Initialize the debug support after everything is loaded diff --git a/sentry_sdk/api.py b/sentry_sdk/api.py index e56109cbd0..698a2085ab 100644 --- a/sentry_sdk/api.py +++ b/sentry_sdk/api.py @@ -82,6 +82,8 @@ def overload(x): "start_transaction", "trace", "monitor", + "start_session", + "end_session", ] @@ -450,3 +452,17 @@ def continue_trace( return get_isolation_scope().continue_trace( environ_or_headers, op, name, source, origin ) + + +@scopemethod +def start_session( + session_mode="application", # type: str +): + # type: (...) -> None + return get_isolation_scope().start_session(session_mode=session_mode) + + +@scopemethod +def end_session(): + # type: () -> None + return get_isolation_scope().end_session() diff --git a/tests/test_sessions.py b/tests/test_sessions.py index 9cad0b7252..731b188727 100644 --- a/tests/test_sessions.py +++ b/tests/test_sessions.py @@ -246,3 +246,52 @@ def test_no_thread_on_shutdown_no_errors_deprecated( sentry_sdk.flush() # If we reach this point without error, the test is successful. + + +def test_top_level_start_session_basic(sentry_init, capture_envelopes): + """Test that top-level start_session starts a session on the isolation scope.""" + sentry_init(release="test-release", environment="test-env") + envelopes = capture_envelopes() + + # Start a session using the top-level API + sentry_sdk.start_session() + + # End the session + sentry_sdk.end_session() + sentry_sdk.flush() + + # Check that we got a session envelope + assert len(envelopes) == 1 + sess = envelopes[0] + assert len(sess.items) == 1 + sess_event = sess.items[0].payload.json + + assert sess_event["attrs"] == { + "release": "test-release", + "environment": "test-env", + } + assert sess_event["status"] == "exited" + + +def test_top_level_start_session_with_mode(sentry_init, capture_envelopes): + """Test that top-level start_session accepts session_mode parameter.""" + sentry_init(release="test-release", environment="test-env") + envelopes = capture_envelopes() + + # Start a session with request mode + sentry_sdk.start_session(session_mode="request") + sentry_sdk.end_session() + sentry_sdk.flush() + + # Request mode sessions are aggregated + assert len(envelopes) == 1 + sess = envelopes[0] + assert len(sess.items) == 1 + sess_event = sess.items[0].payload.json + + assert sess_event["attrs"] == { + "release": "test-release", + "environment": "test-env", + } + # Request sessions show up as aggregates + assert "aggregates" in sess_event From dae02180dfb095cdbd8ed7e81544ef048482d70b Mon Sep 17 00:00:00 2001 From: Lev Vereshchagin Date: Wed, 25 Jun 2025 11:34:58 +0300 Subject: [PATCH 333/591] fix(Litestar): Apply `failed_request_status_codes` to exceptions raised in middleware (#4074) This is a fix for #4021: exceptions raised in middleware were sent without taking into account `failed_request_status_codes` value. See the test case for an example. --------- Co-authored-by: Anton Pirker Co-authored-by: Daniel Szoke --- sentry_sdk/integrations/asgi.py | 22 ++++++- sentry_sdk/integrations/litestar.py | 9 +++ tests/integrations/litestar/test_litestar.py | 66 +++++++++++++++++++- 3 files changed, 93 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index fc8ee29b1a..1b020ebbc0 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -145,6 +145,22 @@ def __init__( else: self.__call__ = self._run_asgi2 + def _capture_lifespan_exception(self, exc): + # type: (Exception) -> None + """Capture exceptions raise in application lifespan handlers. + + The separate function is needed to support overriding in derived integrations that use different catching mechanisms. + """ + return _capture_exception(exc=exc, mechanism_type=self.mechanism_type) + + def _capture_request_exception(self, exc): + # type: (Exception) -> None + """Capture exceptions raised in incoming request handlers. + + The separate function is needed to support overriding in derived integrations that use different catching mechanisms. + """ + return _capture_exception(exc=exc, mechanism_type=self.mechanism_type) + def _run_asgi2(self, scope): # type: (Any) -> Any async def inner(receive, send): @@ -158,7 +174,7 @@ async def _run_asgi3(self, scope, receive, send): return await self._run_app(scope, receive, send, asgi_version=3) async def _run_app(self, scope, receive, send, asgi_version): - # type: (Any, Any, Any, Any, int) -> Any + # type: (Any, Any, Any, int) -> Any is_recursive_asgi_middleware = _asgi_middleware_applied.get(False) is_lifespan = scope["type"] == "lifespan" if is_recursive_asgi_middleware or is_lifespan: @@ -169,7 +185,7 @@ async def _run_app(self, scope, receive, send, asgi_version): return await self.app(scope, receive, send) except Exception as exc: - _capture_exception(exc, mechanism_type=self.mechanism_type) + self._capture_lifespan_exception(exc) raise exc from None _asgi_middleware_applied.set(True) @@ -256,7 +272,7 @@ async def _sentry_wrapped_send(event): scope, receive, _sentry_wrapped_send ) except Exception as exc: - _capture_exception(exc, mechanism_type=self.mechanism_type) + self._capture_request_exception(exc) raise exc from None finally: _asgi_middleware_applied.set(False) diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 5f0b32b04e..4e15081cba 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -87,6 +87,15 @@ def __init__(self, app, span_origin=LitestarIntegration.origin): span_origin=span_origin, ) + def _capture_request_exception(self, exc): + # type: (Exception) -> None + """Avoid catching exceptions from request handlers. + + Those exceptions are already handled in Litestar.after_exception handler. + We still catch exceptions from application lifespan handlers. + """ + pass + def patch_app_init(): # type: () -> None diff --git a/tests/integrations/litestar/test_litestar.py b/tests/integrations/litestar/test_litestar.py index 4f642479e4..b064c17112 100644 --- a/tests/integrations/litestar/test_litestar.py +++ b/tests/integrations/litestar/test_litestar.py @@ -402,7 +402,7 @@ async def __call__(self, scope, receive, send): @parametrize_test_configurable_status_codes -def test_configurable_status_codes( +def test_configurable_status_codes_handler( sentry_init, capture_events, failed_request_status_codes, @@ -427,3 +427,67 @@ async def error() -> None: client.get("/error") assert len(events) == int(expected_error) + + +@parametrize_test_configurable_status_codes +def test_configurable_status_codes_middleware( + sentry_init, + capture_events, + failed_request_status_codes, + status_code, + expected_error, +): + integration_kwargs = ( + {"failed_request_status_codes": failed_request_status_codes} + if failed_request_status_codes is not None + else {} + ) + sentry_init(integrations=[LitestarIntegration(**integration_kwargs)]) + + events = capture_events() + + def create_raising_middleware(app): + async def raising_middleware(scope, receive, send): + raise HTTPException(status_code=status_code) + + return raising_middleware + + @get("/error") + async def error() -> None: ... + + app = Litestar([error], middleware=[create_raising_middleware]) + client = TestClient(app) + client.get("/error") + + assert len(events) == int(expected_error) + + +def test_catch_non_http_exceptions_in_middleware( + sentry_init, + capture_events, +): + sentry_init(integrations=[LitestarIntegration()]) + + events = capture_events() + + def create_raising_middleware(app): + async def raising_middleware(scope, receive, send): + raise RuntimeError("Too Hot") + + return raising_middleware + + @get("/error") + async def error() -> None: ... + + app = Litestar([error], middleware=[create_raising_middleware]) + client = TestClient(app) + + try: + client.get("/error") + except RuntimeError: + pass + + assert len(events) == 1 + event_exception = events[0]["exception"]["values"][0] + assert event_exception["type"] == "RuntimeError" + assert event_exception["value"] == "Too Hot" From 0a2d8585f18f1d135d1f04624b702ef46fd119bb Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Wed, 25 Jun 2025 11:37:39 +0200 Subject: [PATCH 334/591] fix(langchain): Ensure no duplicate `SentryLangchainCallback` (#4485) Ensure that `SentryLangchainCallback` does not get added twice by also checking the `inheritable_callbacks` Fixes https://github.com/getsentry/sentry-python/issues/4443 --- sentry_sdk/integrations/langchain.py | 10 ++- .../integrations/langchain/test_langchain.py | 78 ++++++++++++++++++- 2 files changed, 85 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 1064f29ffd..5f82401389 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -1,3 +1,4 @@ +import itertools from collections import OrderedDict from functools import wraps @@ -451,7 +452,14 @@ def new_configure( **kwargs, ) - if not any(isinstance(cb, SentryLangchainCallback) for cb in callbacks_list): + inheritable_callbacks_list = ( + inheritable_callbacks if isinstance(inheritable_callbacks, list) else [] + ) + + if not any( + isinstance(cb, SentryLangchainCallback) + for cb in itertools.chain(callbacks_list, inheritable_callbacks_list) + ): # Avoid mutating the existing callbacks list callbacks_list = [ *callbacks_list, diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 3f1b3b1da5..863e6daf4c 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -14,10 +14,15 @@ from langchain_core.callbacks import CallbackManagerForLLMRun from langchain_core.messages import BaseMessage, AIMessageChunk -from langchain_core.outputs import ChatGenerationChunk +from langchain_core.outputs import ChatGenerationChunk, ChatResult +from langchain_core.runnables import RunnableConfig +from langchain_core.language_models.chat_models import BaseChatModel from sentry_sdk import start_transaction -from sentry_sdk.integrations.langchain import LangchainIntegration +from sentry_sdk.integrations.langchain import ( + LangchainIntegration, + SentryLangchainCallback, +) from langchain.agents import tool, AgentExecutor, create_openai_tools_agent from langchain_core.prompts import ChatPromptTemplate, MessagesPlaceholder @@ -342,3 +347,72 @@ def test_span_origin(sentry_init, capture_events): assert event["contexts"]["trace"]["origin"] == "manual" for span in event["spans"]: assert span["origin"] == "auto.ai.langchain" + + +def test_manual_callback_no_duplication(sentry_init): + """ + Test that when a user manually provides a SentryLangchainCallback, + the integration doesn't create a duplicate callback. + """ + + # Track callback instances + tracked_callback_instances = set() + + class CallbackTrackingModel(BaseChatModel): + """Mock model that tracks callback instances for testing.""" + + def _generate( + self, + messages, + stop=None, + run_manager=None, + **kwargs, + ): + # Track all SentryLangchainCallback instances + if run_manager: + for handler in run_manager.handlers: + if isinstance(handler, SentryLangchainCallback): + tracked_callback_instances.add(id(handler)) + + for handler in run_manager.inheritable_handlers: + if isinstance(handler, SentryLangchainCallback): + tracked_callback_instances.add(id(handler)) + + return ChatResult( + generations=[ + ChatGenerationChunk(message=AIMessageChunk(content="Hello!")) + ], + llm_output={}, + ) + + @property + def _llm_type(self): + return "test_model" + + @property + def _identifying_params(self): + return {} + + sentry_init(integrations=[LangchainIntegration()]) + + # Create a manual SentryLangchainCallback + manual_callback = SentryLangchainCallback( + max_span_map_size=100, include_prompts=False + ) + + # Create RunnableConfig with the manual callback + config = RunnableConfig(callbacks=[manual_callback]) + + # Invoke the model with the config + llm = CallbackTrackingModel() + llm.invoke("Hello", config) + + # Verify that only ONE SentryLangchainCallback instance was used + assert len(tracked_callback_instances) == 1, ( + f"Expected exactly 1 SentryLangchainCallback instance, " + f"but found {len(tracked_callback_instances)}. " + f"This indicates callback duplication occurred." + ) + + # Verify the callback ID matches our manual callback + assert id(manual_callback) in tracked_callback_instances From 7804260fbf3ed8f797af95d2c0bdfcfeb85b0605 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Wed, 25 Jun 2025 11:47:41 +0200 Subject: [PATCH 335/591] fix(langchain): Make `span_map` an instance variable (#4476) `span_map` should be an instance variable; otherwise, separate instances of the `SentryLangchainCallback` share the same `span_map` object, which is clearly not intended here. Also, remove the `max_span_map_size` class variable, it is always set on the instance, and so not needed. Ref #4443 Co-authored-by: Cursor Agent --- sentry_sdk/integrations/langchain.py | 5 +---- tests/integrations/langchain/test_langchain.py | 12 ++++++++++++ 2 files changed, 13 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 5f82401389..0b8bbd8049 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -89,12 +89,9 @@ def __init__(self, span): class SentryLangchainCallback(BaseCallbackHandler): # type: ignore[misc] """Base callback handler that can be used to handle callbacks from langchain.""" - span_map = OrderedDict() # type: OrderedDict[UUID, WatchedSpan] - - max_span_map_size = 0 - def __init__(self, max_span_map_size, include_prompts, tiktoken_encoding_name=None): # type: (int, bool, Optional[str]) -> None + self.span_map = OrderedDict() # type: OrderedDict[UUID, WatchedSpan] self.max_span_map_size = max_span_map_size self.include_prompts = include_prompts diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 863e6daf4c..8ace6d4821 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -416,3 +416,15 @@ def _identifying_params(self): # Verify the callback ID matches our manual callback assert id(manual_callback) in tracked_callback_instances + + +def test_span_map_is_instance_variable(): + """Test that each SentryLangchainCallback instance has its own span_map.""" + # Create two separate callback instances + callback1 = SentryLangchainCallback(max_span_map_size=100, include_prompts=True) + callback2 = SentryLangchainCallback(max_span_map_size=100, include_prompts=True) + + # Verify they have different span_map instances + assert ( + callback1.span_map is not callback2.span_map + ), "span_map should be an instance variable, not shared between instances" From ab2e3f08b600b22a95c3313eddd66f733e2d133c Mon Sep 17 00:00:00 2001 From: svartalf Date: Wed, 25 Jun 2025 11:54:42 +0200 Subject: [PATCH 336/591] fix(integrations/ray): Correctly pass keyword arguments to ray.remote function (#4430) Monkey-patched implementation was passing the provided keyword arguments incorrectly due to a typo - "*kwargs" was used instead of "**kwargs" twice. Fixed integration started hitting an assert in the Ray codebase that requires for users to use "@ray.remote" decorator either with no arguments and no parentheses, or with some of the arguments provided. An additional wrapper function was added to support both scenarios. --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/integrations/ray.py | 132 +++++++++++++++-------------- tests/integrations/ray/test_ray.py | 14 ++- 2 files changed, 81 insertions(+), 65 deletions(-) diff --git a/sentry_sdk/integrations/ray.py b/sentry_sdk/integrations/ray.py index 0842b92265..8d6cdc1201 100644 --- a/sentry_sdk/integrations/ray.py +++ b/sentry_sdk/integrations/ray.py @@ -42,73 +42,81 @@ def _patch_ray_remote(): old_remote = ray.remote @functools.wraps(old_remote) - def new_remote(f, *args, **kwargs): - # type: (Callable[..., Any], *Any, **Any) -> Callable[..., Any] + def new_remote(f=None, *args, **kwargs): + # type: (Optional[Callable[..., Any]], *Any, **Any) -> Callable[..., Any] + if inspect.isclass(f): # Ray Actors # (https://docs.ray.io/en/latest/ray-core/actors.html) # are not supported # (Only Ray Tasks are supported) - return old_remote(f, *args, *kwargs) - - def _f(*f_args, _tracing=None, **f_kwargs): - # type: (Any, Optional[dict[str, Any]], Any) -> Any - """ - Ray Worker - """ - _check_sentry_initialized() - - transaction = sentry_sdk.continue_trace( - _tracing or {}, - op=OP.QUEUE_TASK_RAY, - name=qualname_from_function(f), - origin=RayIntegration.origin, - source=TransactionSource.TASK, - ) - - with sentry_sdk.start_transaction(transaction) as transaction: - try: - result = f(*f_args, **f_kwargs) - transaction.set_status(SPANSTATUS.OK) - except Exception: - transaction.set_status(SPANSTATUS.INTERNAL_ERROR) - exc_info = sys.exc_info() - _capture_exception(exc_info) - reraise(*exc_info) - - return result - - rv = old_remote(_f, *args, *kwargs) - old_remote_method = rv.remote - - def _remote_method_with_header_propagation(*args, **kwargs): - # type: (*Any, **Any) -> Any - """ - Ray Client - """ - with sentry_sdk.start_span( - op=OP.QUEUE_SUBMIT_RAY, - name=qualname_from_function(f), - origin=RayIntegration.origin, - ) as span: - tracing = { - k: v - for k, v in sentry_sdk.get_current_scope().iter_trace_propagation_headers() - } - try: - result = old_remote_method(*args, **kwargs, _tracing=tracing) - span.set_status(SPANSTATUS.OK) - except Exception: - span.set_status(SPANSTATUS.INTERNAL_ERROR) - exc_info = sys.exc_info() - _capture_exception(exc_info) - reraise(*exc_info) - - return result - - rv.remote = _remote_method_with_header_propagation - - return rv + return old_remote(f, *args, **kwargs) + + def wrapper(user_f): + # type: (Callable[..., Any]) -> Any + def new_func(*f_args, _tracing=None, **f_kwargs): + # type: (Any, Optional[dict[str, Any]], Any) -> Any + _check_sentry_initialized() + + transaction = sentry_sdk.continue_trace( + _tracing or {}, + op=OP.QUEUE_TASK_RAY, + name=qualname_from_function(user_f), + origin=RayIntegration.origin, + source=TransactionSource.TASK, + ) + + with sentry_sdk.start_transaction(transaction) as transaction: + try: + result = user_f(*f_args, **f_kwargs) + transaction.set_status(SPANSTATUS.OK) + except Exception: + transaction.set_status(SPANSTATUS.INTERNAL_ERROR) + exc_info = sys.exc_info() + _capture_exception(exc_info) + reraise(*exc_info) + + return result + + if f: + rv = old_remote(new_func) + else: + rv = old_remote(*args, **kwargs)(new_func) + old_remote_method = rv.remote + + def _remote_method_with_header_propagation(*args, **kwargs): + # type: (*Any, **Any) -> Any + """ + Ray Client + """ + with sentry_sdk.start_span( + op=OP.QUEUE_SUBMIT_RAY, + name=qualname_from_function(user_f), + origin=RayIntegration.origin, + ) as span: + tracing = { + k: v + for k, v in sentry_sdk.get_current_scope().iter_trace_propagation_headers() + } + try: + result = old_remote_method(*args, **kwargs, _tracing=tracing) + span.set_status(SPANSTATUS.OK) + except Exception: + span.set_status(SPANSTATUS.INTERNAL_ERROR) + exc_info = sys.exc_info() + _capture_exception(exc_info) + reraise(*exc_info) + + return result + + rv.remote = _remote_method_with_header_propagation + + return rv + + if f is not None: + return wrapper(f) + else: + return wrapper ray.remote = new_remote diff --git a/tests/integrations/ray/test_ray.py b/tests/integrations/ray/test_ray.py index 95ab4ad0fa..b5bdd473c4 100644 --- a/tests/integrations/ray/test_ray.py +++ b/tests/integrations/ray/test_ray.py @@ -59,7 +59,10 @@ def read_error_from_log(job_id): @pytest.mark.forked -def test_tracing_in_ray_tasks(): +@pytest.mark.parametrize( + "task_options", [{}, {"num_cpus": 0, "memory": 1024 * 1024 * 10}] +) +def test_tracing_in_ray_tasks(task_options): setup_sentry() ray.init( @@ -69,14 +72,19 @@ def test_tracing_in_ray_tasks(): } ) - # Setup ray task - @ray.remote def example_task(): with sentry_sdk.start_span(op="task", name="example task step"): ... return sentry_sdk.get_client().transport.envelopes + # Setup ray task, calling decorator directly instead of @, + # to accommodate for test parametrization + if task_options: + example_task = ray.remote(**task_options)(example_task) + else: + example_task = ray.remote(example_task) + with sentry_sdk.start_transaction(op="task", name="ray test transaction"): worker_envelopes = ray.get(example_task.remote()) From 546ce1f71023b651860d6b576024b9d93b4c9ab8 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 26 Jun 2025 14:21:09 +0200 Subject: [PATCH 337/591] Set tool span to failed if an error is raised in the tool (#4527) Co-authored-by: Ivana Kellyer --- .../integrations/openai_agents/spans/execute_tool.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/openai_agents/spans/execute_tool.py b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py index e6e880b64c..5f9e4cb340 100644 --- a/sentry_sdk/integrations/openai_agents/spans/execute_tool.py +++ b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py @@ -1,5 +1,5 @@ import sentry_sdk -from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.consts import OP, SPANDATA, SPANSTATUS from sentry_sdk.scope import should_send_default_pii from ..consts import SPAN_ORIGIN @@ -39,5 +39,10 @@ def update_execute_tool_span(span, agent, tool, result): # type: (sentry_sdk.tracing.Span, agents.Agent, agents.Tool, Any) -> None _set_agent_data(span, agent) + if isinstance(result, str) and result.startswith( + "An error occurred while running the tool" + ): + span.set_status(SPANSTATUS.INTERNAL_ERROR) + if should_send_default_pii(): span.set_data(SPANDATA.GEN_AI_TOOL_OUTPUT, result) From bca8816ac1f84fe4304682bd6de173fbf0c005a3 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 26 Jun 2025 12:27:44 +0000 Subject: [PATCH 338/591] release: 2.32.0 --- CHANGELOG.md | 11 +++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 14 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8bcd8ddc73..63cb761830 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,16 @@ # Changelog +## 2.32.0 + +### Various fixes & improvements + +- Set tool span to failed if an error is raised in the tool (#4527) by @antonpirker +- fix(integrations/ray): Correctly pass keyword arguments to ray.remote function (#4430) by @svartalf +- fix(langchain): Make `span_map` an instance variable (#4476) by @szokeasaurusrex +- fix(langchain): Ensure no duplicate `SentryLangchainCallback` (#4485) by @szokeasaurusrex +- fix(Litestar): Apply `failed_request_status_codes` to exceptions raised in middleware (#4074) by @vrslev +- feat(sessions): Add top-level start- and end session methods (#4474) by @szokeasaurusrex + ## 2.31.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 01b40ae828..ea5995ee36 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.31.0" +release = "2.32.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 7102eea0e7..01f72e2887 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1181,4 +1181,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.31.0" +VERSION = "2.32.0" diff --git a/setup.py b/setup.py index 0662be384e..ae86cab158 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.31.0", + version="2.32.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From c815a3245d10e45bebee5b47292deec438a4d4d2 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 26 Jun 2025 14:28:54 +0200 Subject: [PATCH 339/591] updated changelog --- CHANGELOG.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 63cb761830..fd4a98e717 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,12 +4,12 @@ ### Various fixes & improvements -- Set tool span to failed if an error is raised in the tool (#4527) by @antonpirker +- feat(sessions): Add top-level start- and end session methods (#4474) by @szokeasaurusrex +- feat(openai-agents): Set tool span to failed if an error is raised in the tool (#4527) by @antonpirker - fix(integrations/ray): Correctly pass keyword arguments to ray.remote function (#4430) by @svartalf - fix(langchain): Make `span_map` an instance variable (#4476) by @szokeasaurusrex - fix(langchain): Ensure no duplicate `SentryLangchainCallback` (#4485) by @szokeasaurusrex - fix(Litestar): Apply `failed_request_status_codes` to exceptions raised in middleware (#4074) by @vrslev -- feat(sessions): Add top-level start- and end session methods (#4474) by @szokeasaurusrex ## 2.31.0 From 2634a523b3416748cf952bc517641594b9b40bac Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 27 Jun 2025 08:35:14 +0200 Subject: [PATCH 340/591] Pin zope.event (#4531) zope.event [released](https://pypi.org/project/zope.event/#history) a new version recently that broke our gevent ci --- scripts/populate_tox/tox.jinja | 1 + tox.ini | 1 + 2 files changed, 2 insertions(+) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index ac14bdb02a..c67f4127d5 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -173,6 +173,7 @@ deps = {py3.6,py3.7}-gevent: pytest<7.0.0 {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest gevent: pytest-asyncio + {py3.10,py3.11}-gevent: zope.event<5.0.0 # === Integrations === diff --git a/tox.ini b/tox.ini index 5c993718d7..881fb44574 100644 --- a/tox.ini +++ b/tox.ini @@ -336,6 +336,7 @@ deps = {py3.6,py3.7}-gevent: pytest<7.0.0 {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest gevent: pytest-asyncio + {py3.10,py3.11}-gevent: zope.event<5.0.0 # === Integrations === From 4f6613ea5b750183907f0ccd3af22ed72ed8a859 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 8 Jul 2025 09:36:05 +0200 Subject: [PATCH 341/591] chore: Remove Lambda urllib3 pin on Python 3.10+ (#4549) See [botocore](https://github.com/boto/botocore/blob/0b28c3155f160a68fd901c9eb80a7a6a45a16216/setup.cfg#L8) --- requirements-aws-lambda-layer.txt | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/requirements-aws-lambda-layer.txt b/requirements-aws-lambda-layer.txt index 8986fdafc0..8a6ff63aa7 100644 --- a/requirements-aws-lambda-layer.txt +++ b/requirements-aws-lambda-layer.txt @@ -1,7 +1,8 @@ certifi - -# In Lambda functions botocore is used, and botocore is not -# yet supporting urllib3 1.27.0 never mind 2+. +urllib3 +# In Lambda functions botocore is used, and botocore has +# restrictions on urllib3 +# https://github.com/boto/botocore/blob/develop/setup.cfg # So we pin this here to make our Lambda layer work with # Lambda Function using Python 3.7+ -urllib3<1.27 +urllib3<1.27; python_version < "3.10" From 987824e98259245bfc18d23c818332030231e2df Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 8 Jul 2025 09:37:05 +0200 Subject: [PATCH 342/591] tests: Tox update (#4555) --- tox.ini | 58 ++++++++++++++++++++++++++++++--------------------------- 1 file changed, 31 insertions(+), 27 deletions(-) diff --git a/tox.ini b/tox.ini index 881fb44574..3b3081b3cb 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-06-24T12:35:34.437673+00:00 +# Last generated: 2025-07-08T06:07:54.743036+00:00 [tox] requires = @@ -136,9 +136,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.29.2 - {py3.8,py3.11,py3.12}-anthropic-v0.42.0 - {py3.8,py3.11,py3.12}-anthropic-v0.55.0 + {py3.8,py3.11,py3.12}-anthropic-v0.30.1 + {py3.8,py3.11,py3.12}-anthropic-v0.44.0 + {py3.8,py3.11,py3.12}-anthropic-v0.57.1 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.8.1 @@ -146,11 +146,12 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.15.0 {py3.9,py3.11,py3.12}-openai_agents-v0.0.19 + {py3.9,py3.12,py3.13}-openai_agents-v0.1.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.0 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.2 # ~~~ DBs ~~~ @@ -186,7 +187,8 @@ envlist = {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 - {py3.8,py3.12,py3.13}-unleash-v6.2.1 + {py3.8,py3.12,py3.13}-unleash-v6.2.2 + {py3.8,py3.12,py3.13}-unleash-v6.3.0 # ~~~ GraphQL ~~~ @@ -205,14 +207,14 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.231.1 {py3.8,py3.12,py3.13}-strawberry-v0.253.1 - {py3.9,py3.12,py3.13}-strawberry-v0.275.2 + {py3.9,py3.12,py3.13}-strawberry-v0.275.5 # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.46.5 {py3.7,py3.11,py3.12}-grpc-v1.60.2 - {py3.9,py3.12,py3.13}-grpc-v1.73.0 + {py3.9,py3.12,py3.13}-grpc-v1.73.1 # ~~~ Tasks ~~~ @@ -241,7 +243,7 @@ envlist = {py3.6,py3.9,py3.10}-django-v3.2.25 {py3.8,py3.11,py3.12}-django-v4.2.23 {py3.10,py3.11,py3.12}-django-v5.0.14 - {py3.10,py3.12,py3.13}-django-v5.2.3 + {py3.10,py3.12,py3.13}-django-v5.2.4 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -256,7 +258,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 {py3.7,py3.10,py3.11}-fastapi-v0.103.2 - {py3.8,py3.12,py3.13}-fastapi-v0.115.13 + {py3.8,py3.12,py3.13}-fastapi-v0.116.0 # ~~~ Web 2 ~~~ @@ -300,8 +302,8 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.32 - {py3.9,py3.12,py3.13}-trytond-v7.6.2 + {py3.8,py3.11,py3.12}-trytond-v7.0.33 + {py3.9,py3.12,py3.13}-trytond-v7.6.3 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.0 @@ -505,13 +507,13 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.29.2: anthropic==0.29.2 - anthropic-v0.42.0: anthropic==0.42.0 - anthropic-v0.55.0: anthropic==0.55.0 + anthropic-v0.30.1: anthropic==0.30.1 + anthropic-v0.44.0: anthropic==0.44.0 + anthropic-v0.57.1: anthropic==0.57.1 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.29.2: httpx<0.28.0 - anthropic-v0.42.0: httpx<0.28.0 + anthropic-v0.30.1: httpx<0.28.0 + anthropic-v0.44.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.8.1: cohere==5.8.1 @@ -519,12 +521,13 @@ deps = cohere-v5.15.0: cohere==5.15.0 openai_agents-v0.0.19: openai-agents==0.0.19 + openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents: pytest-asyncio huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.33.0: huggingface_hub==0.33.0 + huggingface_hub-v0.33.2: huggingface_hub==0.33.2 # ~~~ DBs ~~~ @@ -562,7 +565,8 @@ deps = unleash-v6.0.1: UnleashClient==6.0.1 unleash-v6.1.0: UnleashClient==6.1.0 - unleash-v6.2.1: UnleashClient==6.2.1 + unleash-v6.2.2: UnleashClient==6.2.2 + unleash-v6.3.0: UnleashClient==6.3.0 # ~~~ GraphQL ~~~ @@ -589,7 +593,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.231.1: strawberry-graphql[fastapi,flask]==0.231.1 strawberry-v0.253.1: strawberry-graphql[fastapi,flask]==0.253.1 - strawberry-v0.275.2: strawberry-graphql[fastapi,flask]==0.275.2 + strawberry-v0.275.5: strawberry-graphql[fastapi,flask]==0.275.5 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.231.1: pydantic<2.11 @@ -600,7 +604,7 @@ deps = grpc-v1.32.0: grpcio==1.32.0 grpc-v1.46.5: grpcio==1.46.5 grpc-v1.60.2: grpcio==1.60.2 - grpc-v1.73.0: grpcio==1.73.0 + grpc-v1.73.1: grpcio==1.73.1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -636,7 +640,7 @@ deps = django-v3.2.25: django==3.2.25 django-v4.2.23: django==4.2.23 django-v5.0.14: django==5.0.14 - django-v5.2.3: django==5.2.3 + django-v5.2.4: django==5.2.4 django: psycopg2-binary django: djangorestframework django: pytest-django @@ -645,12 +649,12 @@ deps = django-v3.2.25: channels[daphne] django-v4.2.23: channels[daphne] django-v5.0.14: channels[daphne] - django-v5.2.3: channels[daphne] + django-v5.2.4: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio django-v4.2.23: pytest-asyncio django-v5.0.14: pytest-asyncio - django-v5.2.3: pytest-asyncio + django-v5.2.4: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 @@ -687,7 +691,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.91.0: fastapi==0.91.0 fastapi-v0.103.2: fastapi==0.103.2 - fastapi-v0.115.13: fastapi==0.115.13 + fastapi-v0.116.0: fastapi==0.116.0 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -761,8 +765,8 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.32: trytond==7.0.32 - trytond-v7.6.2: trytond==7.6.2 + trytond-v7.0.33: trytond==7.0.33 + trytond-v7.6.3: trytond==7.6.3 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 7e4053afea9a5e742977f47efb585d5f1a0edd8c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 8 Jul 2025 12:16:49 +0200 Subject: [PATCH 343/591] toxgen: Detect correct sentry-sdk (#4558) Sometimes after switching branches, `importlib.metadata("sentry-sdk")` would point to the wrong metadata (SDK 2.x instead of 3.x), which in turn meant toxgen would think the SDK supports different Python versions than it should. Closes https://github.com/getsentry/sentry-python/issues/4556 --- scripts/generate-test-files.sh | 11 ++++++----- scripts/populate_tox/populate_tox.py | 12 ++++++++++-- 2 files changed, 16 insertions(+), 7 deletions(-) diff --git a/scripts/generate-test-files.sh b/scripts/generate-test-files.sh index 40e279cdf4..d1e0a7602c 100755 --- a/scripts/generate-test-files.sh +++ b/scripts/generate-test-files.sh @@ -6,12 +6,13 @@ set -xe cd "$(dirname "$0")" +rm -rf toxgen.venv python -m venv toxgen.venv . toxgen.venv/bin/activate -pip install -e .. -pip install -r populate_tox/requirements.txt -pip install -r split_tox_gh_actions/requirements.txt +toxgen.venv/bin/pip install -e .. +toxgen.venv/bin/pip install -r populate_tox/requirements.txt +toxgen.venv/bin/pip install -r split_tox_gh_actions/requirements.txt -python populate_tox/populate_tox.py -python split_tox_gh_actions/split_tox_gh_actions.py +toxgen.venv/bin/python populate_tox/populate_tox.py +toxgen.venv/bin/python split_tox_gh_actions/split_tox_gh_actions.py diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 0aeb0f02ef..3d9e247b4f 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -10,7 +10,7 @@ from bisect import bisect_left from collections import defaultdict from datetime import datetime, timedelta, timezone # noqa: F401 -from importlib.metadata import metadata +from importlib.metadata import PackageMetadata, distributions from packaging.specifiers import SpecifierSet from packaging.version import Version from pathlib import Path @@ -88,6 +88,13 @@ } +def _fetch_sdk_metadata() -> PackageMetadata: + (dist,) = distributions( + name="sentry-sdk", path=[Path(__file__).parent.parent.parent] + ) + return dist.metadata + + def fetch_url(/service/url: str) -> Optional[dict]: for attempt in range(3): pypi_data = requests.get(url) @@ -583,8 +590,9 @@ def main(fail_on_changes: bool = False) -> None: ) global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION + meta = _fetch_sdk_metadata() sdk_python_versions = _parse_python_versions_from_classifiers( - metadata("sentry-sdk").get_all("Classifier") + meta.get_all("Classifier") ) MIN_PYTHON_VERSION = sdk_python_versions[0] MAX_PYTHON_VERSION = sdk_python_versions[-1] From a7b2d678a12658da0b0a686839a6ba9977b2c9f2 Mon Sep 17 00:00:00 2001 From: Simon Roth <39389607+srothh@users.noreply.github.com> Date: Tue, 8 Jul 2025 12:34:44 +0200 Subject: [PATCH 344/591] test(loguru): Remove hardcoded line number in test_just_log (#4552) Use regex to check line number instead of hard coding it Fixes GH-4454 https://github.com/getsentry/sentry-python/issues/4454 Changed the loguru test_just_log test to use a regex for the log format instead of a hard-coded line number. --- tests/integrations/loguru/test_loguru.py | 13 +++++++------ 1 file changed, 7 insertions(+), 6 deletions(-) diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 20d3230b49..c120d1d7e2 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -1,4 +1,5 @@ from unittest.mock import MagicMock, patch +import re import pytest from loguru import logger @@ -56,10 +57,10 @@ def test_just_log( getattr(logger, level.name.lower())("test") - formatted_message = ( - " | " - + "{:9}".format(level.name.upper()) - + "| tests.integrations.loguru.test_loguru:test_just_log:57 - test" + expected_pattern = ( + r" \| " + + r"{:9}".format(level.name.upper()) + + r"\| tests\.integrations\.loguru\.test_loguru:test_just_log:\d+ - test" ) if not created_event: @@ -72,7 +73,7 @@ def test_just_log( (breadcrumb,) = breadcrumbs assert breadcrumb["level"] == expected_sentry_level assert breadcrumb["category"] == "tests.integrations.loguru.test_loguru" - assert breadcrumb["message"][23:] == formatted_message + assert re.fullmatch(expected_pattern, breadcrumb["message"][23:]) else: assert not breadcrumbs @@ -85,7 +86,7 @@ def test_just_log( (event,) = events assert event["level"] == expected_sentry_level assert event["logger"] == "tests.integrations.loguru.test_loguru" - assert event["logentry"]["message"][23:] == formatted_message + assert re.fullmatch(expected_pattern, event["logentry"]["message"][23:]) def test_breadcrumb_format(sentry_init, capture_events, uninstall_integration, request): From 5709b25e817b9c64d729254de53d9ffd96a187ec Mon Sep 17 00:00:00 2001 From: Buck Evan Date: Wed, 9 Jul 2025 09:15:23 -0500 Subject: [PATCH 345/591] fix: shut down "session flusher" more promptly (#4561) Currently this thread will wait for `sleep(self.flush_interval)` to end before checking `self._running`. We can do better by using an Event. Now it will wait for `self.flush_interval` or until `self.__shutdown_requested` is set, whichever is shorter. --- sentry_sdk/sessions.py | 17 ++++++++--------- 1 file changed, 8 insertions(+), 9 deletions(-) diff --git a/sentry_sdk/sessions.py b/sentry_sdk/sessions.py index eaeb915e7b..a5dd589ee9 100644 --- a/sentry_sdk/sessions.py +++ b/sentry_sdk/sessions.py @@ -1,7 +1,6 @@ import os -import time import warnings -from threading import Thread, Lock +from threading import Thread, Lock, Event from contextlib import contextmanager import sentry_sdk @@ -162,7 +161,7 @@ def __init__( self._thread_lock = Lock() self._aggregate_lock = Lock() self._thread_for_pid = None # type: Optional[int] - self._running = True + self.__shutdown_requested = Event() def flush(self): # type: (...) -> None @@ -208,10 +207,10 @@ def _ensure_running(self): def _thread(): # type: (...) -> None - while self._running: - time.sleep(self.flush_interval) - if self._running: - self.flush() + running = True + while running: + running = not self.__shutdown_requested.wait(self.flush_interval) + self.flush() thread = Thread(target=_thread) thread.daemon = True @@ -220,7 +219,7 @@ def _thread(): except RuntimeError: # Unfortunately at this point the interpreter is in a state that no # longer allows us to spawn a thread and we have to bail. - self._running = False + self.__shutdown_requested.set() return None self._thread = thread @@ -271,7 +270,7 @@ def add_session( def kill(self): # type: (...) -> None - self._running = False + self.__shutdown_requested.set() def __del__(self): # type: (...) -> None From 1df6c9a9848db0a92a0ec35aa3e6c38c7a2c6b08 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 10 Jul 2025 11:48:35 +0200 Subject: [PATCH 346/591] Fix custom model name (#4569) The `model` parameter can be a string or an object of the model itself. If it is a model, get the models name from the `.model` attribute. --- .../openai_agents/spans/ai_client.py | 3 +- .../integrations/openai_agents/utils.py | 3 +- .../openai_agents/test_openai_agents.py | 54 +++++++++++++++++++ 3 files changed, 58 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/openai_agents/spans/ai_client.py b/sentry_sdk/integrations/openai_agents/spans/ai_client.py index 30c5fd1dac..d325ae86e3 100644 --- a/sentry_sdk/integrations/openai_agents/spans/ai_client.py +++ b/sentry_sdk/integrations/openai_agents/spans/ai_client.py @@ -19,9 +19,10 @@ def ai_client_span(agent, get_response_kwargs): # type: (Agent, dict[str, Any]) -> sentry_sdk.tracing.Span # TODO-anton: implement other types of operations. Now "chat" is hardcoded. + model_name = agent.model.model if hasattr(agent.model, "model") else agent.model span = sentry_sdk.start_span( op=OP.GEN_AI_CHAT, - description=f"chat {agent.model}", + description=f"chat {model_name}", origin=SPAN_ORIGIN, ) # TODO-anton: remove hardcoded stuff and replace something that also works for embedding and so on diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index 28dbd6bb75..dc66521c83 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -53,7 +53,8 @@ def _set_agent_data(span, agent): ) if agent.model: - span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, agent.model) + model_name = agent.model.model if hasattr(agent.model, "model") else agent.model + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) if agent.model_settings.presence_penalty: span.set_data( diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index ec606c8806..37a066aeca 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -74,6 +74,24 @@ def test_agent(): ) +@pytest.fixture +def test_agent_custom_model(): + """Create a real Agent instance for testing.""" + return Agent( + name="test_agent_custom_model", + instructions="You are a helpful test assistant.", + # the model could be agents.OpenAIChatCompletionsModel() + model=MagicMock(model="my-custom-model"), + model_settings=ModelSettings( + max_tokens=100, + temperature=0.7, + top_p=1.0, + presence_penalty=0.0, + frequency_penalty=0.0, + ), + ) + + @pytest.mark.asyncio async def test_agent_invocation_span( sentry_init, capture_events, test_agent, mock_model_response @@ -128,6 +146,42 @@ async def test_agent_invocation_span( assert ai_client_span["data"]["gen_ai.request.top_p"] == 1.0 +@pytest.mark.asyncio +async def test_client_span_custom_model( + sentry_init, capture_events, test_agent_custom_model, mock_model_response +): + """ + Test that the integration uses the correct model name if a custom model is used. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.return_value = mock_model_response + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + result = await agents.Runner.run( + test_agent_custom_model, "Test input", run_config=test_run_config + ) + + assert result is not None + assert result.final_output == "Hello, how can I help you?" + + (transaction,) = events + spans = transaction["spans"] + _, ai_client_span = spans + + assert ai_client_span["description"] == "chat my-custom-model" + assert ai_client_span["data"]["gen_ai.request.model"] == "my-custom-model" + + def test_agent_invocation_span_sync( sentry_init, capture_events, test_agent, mock_model_response ): From 6f71a1bb1db952cf12a71fef2e47971850f2a773 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 10 Jul 2025 15:11:11 +0200 Subject: [PATCH 347/591] Use `span.data` instead of `measurements` for token usage (#4567) Store AI token usage in `span.data` instead of deprecated `measurements`. In `relay` there is already code in place that copies the data from the deprecated `span.measurements` to `span.data` and uses `span.data` for calculating the cost of token usage. So this PR can be deployed in a minor without risk. See also `relay` PR: https://github.com/getsentry/relay/pull/4768 --- sentry_sdk/ai/monitoring.py | 10 ++-- .../integrations/anthropic/test_anthropic.py | 46 +++++++++---------- tests/integrations/cohere/test_cohere.py | 16 +++---- .../huggingface_hub/test_huggingface_hub.py | 4 +- .../integrations/langchain/test_langchain.py | 4 +- tests/integrations/openai/test_openai.py | 32 ++++++------- 6 files changed, 58 insertions(+), 54 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index ed33acd0f1..d3154f0631 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -102,15 +102,19 @@ def record_token_usage( ai_pipeline_name = get_ai_pipeline_name() if ai_pipeline_name: span.set_data(SPANDATA.AI_PIPELINE_NAME, ai_pipeline_name) + if prompt_tokens is not None: - span.set_measurement("ai_prompt_tokens_used", value=prompt_tokens) + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, prompt_tokens) + if completion_tokens is not None: - span.set_measurement("ai_completion_tokens_used", value=completion_tokens) + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, completion_tokens) + if ( total_tokens is None and prompt_tokens is not None and completion_tokens is not None ): total_tokens = prompt_tokens + completion_tokens + if total_tokens is not None: - span.set_measurement("ai_total_tokens_used", total_tokens) + span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, total_tokens) diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index 9ab0f879d1..e6e1a40aa9 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -125,9 +125,9 @@ def test_nonstreaming_create_message( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.input_tokens"] == 10 + assert span["data"]["gen_ai.usage.output_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 assert span["data"][SPANDATA.AI_STREAMING] is False @@ -193,9 +193,9 @@ async def test_nonstreaming_create_message_async( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.input_tokens"] == 10 + assert span["data"]["gen_ai.usage.output_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 assert span["data"][SPANDATA.AI_STREAMING] is False @@ -293,9 +293,9 @@ def test_streaming_create_message( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 30 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 40 + assert span["data"]["gen_ai.usage.input_tokens"] == 10 + assert span["data"]["gen_ai.usage.output_tokens"] == 30 + assert span["data"]["gen_ai.usage.total_tokens"] == 40 assert span["data"][SPANDATA.AI_STREAMING] is True @@ -396,9 +396,9 @@ async def test_streaming_create_message_async( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 30 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 40 + assert span["data"]["gen_ai.usage.input_tokens"] == 10 + assert span["data"]["gen_ai.usage.output_tokens"] == 30 + assert span["data"]["gen_ai.usage.total_tokens"] == 40 assert span["data"][SPANDATA.AI_STREAMING] is True @@ -525,9 +525,9 @@ def test_streaming_create_message_with_input_json_delta( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 366 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 51 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 417 + assert span["data"]["gen_ai.usage.input_tokens"] == 366 + assert span["data"]["gen_ai.usage.output_tokens"] == 51 + assert span["data"]["gen_ai.usage.total_tokens"] == 417 assert span["data"][SPANDATA.AI_STREAMING] is True @@ -662,9 +662,9 @@ async def test_streaming_create_message_with_input_json_delta_async( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 366 - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 51 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 417 + assert span["data"]["gen_ai.usage.input_tokens"] == 366 + assert span["data"]["gen_ai.usage.output_tokens"] == 51 + assert span["data"]["gen_ai.usage.total_tokens"] == 417 assert span["data"][SPANDATA.AI_STREAMING] is True @@ -807,10 +807,10 @@ def test_add_ai_data_to_span_with_input_json_delta(sentry_init): content_blocks=["{'test': 'data',", "'more': 'json'}"], ) - assert span._data.get(SPANDATA.AI_RESPONSES) == [ + assert span._data.get("ai.responses") == [ {"type": "text", "text": "{'test': 'data','more': 'json'}"} ] - assert span._data.get(SPANDATA.AI_STREAMING) is True - assert span._measurements.get("ai_prompt_tokens_used")["value"] == 10 - assert span._measurements.get("ai_completion_tokens_used")["value"] == 20 - assert span._measurements.get("ai_total_tokens_used")["value"] == 30 + assert span._data.get("ai.streaming") is True + assert span._data.get("gen_ai.usage.input_tokens") == 10 + assert span._data.get("gen_ai.usage.output_tokens") == 20 + assert span._data.get("gen_ai.usage.total_tokens") == 30 diff --git a/tests/integrations/cohere/test_cohere.py b/tests/integrations/cohere/test_cohere.py index 6c1185a28e..f13a77ae90 100644 --- a/tests/integrations/cohere/test_cohere.py +++ b/tests/integrations/cohere/test_cohere.py @@ -64,9 +64,9 @@ def test_nonstreaming_chat( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 # noinspection PyTypeChecker @@ -135,9 +135,9 @@ def test_streaming_chat(sentry_init, capture_events, send_default_pii, include_p assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 def test_bad_chat(sentry_init, capture_events): @@ -199,8 +199,8 @@ def test_embed(sentry_init, capture_events, send_default_pii, include_prompts): else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 + assert span["data"]["gen_ai.usage.input_tokens"] == 10 + assert span["data"]["gen_ai.usage.total_tokens"] == 10 def test_span_origin_chat(sentry_init, capture_events): diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index ee47cc7e56..540fd675b9 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -75,7 +75,7 @@ def test_nonstreaming_chat_completion( assert SPANDATA.AI_RESPONSES not in span["data"] if details_arg: - assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 + assert span["data"]["gen_ai.usage.total_tokens"] == 10 @pytest.mark.parametrize( @@ -134,7 +134,7 @@ def test_streaming_chat_completion( assert SPANDATA.AI_RESPONSES not in span["data"] if details_arg: - assert span["measurements"]["ai_total_tokens_used"]["value"] == 10 + assert span["data"]["gen_ai.usage.total_tokens"] == 10 def test_bad_chat_completion(sentry_init, capture_events): diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 8ace6d4821..a50a2849c3 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -186,8 +186,8 @@ def test_langchain_agent( assert len(list(x for x in tx["spans"] if x["op"] == "ai.run.langchain")) > 0 if use_unknown_llm_type: - assert "ai_prompt_tokens_used" in chat_spans[0]["measurements"] - assert "ai_total_tokens_used" in chat_spans[0]["measurements"] + assert "gen_ai.usage.input_tokens" in chat_spans[0]["data"] + assert "gen_ai.usage.total_tokens" in chat_spans[0]["data"] else: # important: to avoid double counting, we do *not* measure # tokens used if we have an explicit integration (e.g. OpenAI) diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 3fdc138f39..39195de277 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -90,9 +90,9 @@ def test_nonstreaming_chat_completion( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 @pytest.mark.asyncio @@ -132,9 +132,9 @@ async def test_nonstreaming_chat_completion_async( assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] assert SPANDATA.AI_RESPONSES not in span["data"] - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 10 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 def tiktoken_encoding_if_installed(): @@ -228,9 +228,9 @@ def test_streaming_chat_completion( try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 2 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 1 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 3 + assert span["data"]["gen_ai.usage.output_tokens"] == 2 + assert span["data"]["gen_ai.usage.input_tokens"] == 1 + assert span["data"]["gen_ai.usage.total_tokens"] == 3 except ImportError: pass # if tiktoken is not installed, we can't guarantee token usage will be calculated properly @@ -324,9 +324,9 @@ async def test_streaming_chat_completion_async( try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import - assert span["measurements"]["ai_completion_tokens_used"]["value"] == 2 - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 1 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 3 + assert span["data"]["gen_ai.usage.output_tokens"] == 2 + assert span["data"]["gen_ai.usage.input_tokens"] == 1 + assert span["data"]["gen_ai.usage.total_tokens"] == 3 except ImportError: pass # if tiktoken is not installed, we can't guarantee token usage will be calculated properly @@ -409,8 +409,8 @@ def test_embeddings_create( else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 @pytest.mark.asyncio @@ -457,8 +457,8 @@ async def test_embeddings_create_async( else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert span["measurements"]["ai_prompt_tokens_used"]["value"] == 20 - assert span["measurements"]["ai_total_tokens_used"]["value"] == 30 + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 @pytest.mark.parametrize( From 30ad1b26c72ae08e00d04fa538e49941c03c29e0 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 11 Jul 2025 13:41:38 +0200 Subject: [PATCH 348/591] Remove print statements from excepthook test (#4573) --- tests/integrations/excepthook/test_excepthook.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/tests/integrations/excepthook/test_excepthook.py b/tests/integrations/excepthook/test_excepthook.py index 82fe6c6861..745f62d818 100644 --- a/tests/integrations/excepthook/test_excepthook.py +++ b/tests/integrations/excepthook/test_excepthook.py @@ -42,7 +42,6 @@ def capture_envelope(self, envelope): subprocess.check_output([sys.executable, str(app)], stderr=subprocess.STDOUT) output = excinfo.value.output - print(output) assert b"ZeroDivisionError" in output assert b"LOL" in output @@ -86,7 +85,6 @@ def capture_envelope(self, envelope): subprocess.check_output([sys.executable, str(app)], stderr=subprocess.STDOUT) output = excinfo.value.output - print(output) assert b"ZeroDivisionError" in output assert b"LOL" in output From 710227aebfaa37a944504e54ca5189c7a7b16cad Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 11 Jul 2025 14:18:05 +0200 Subject: [PATCH 349/591] Fix pytest collection warning (#4574) fixes ``` PytestCollectionWarning: cannot collect test class 'TestSpanClientReports' because it has a __init__ constructor (from: tests/test_client.py) class TestSpanClientReports: ``` --- tests/test_client.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/tests/test_client.py b/tests/test_client.py index 2986920452..60b15615c8 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -1342,6 +1342,8 @@ class TestSpanClientReports: Tests for client reports related to spans. """ + __test__ = False + @staticmethod def span_dropper(spans_to_drop): """ From 7d7027a586f0864a7553ce4e504f1ede8f3af470 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 11 Jul 2025 14:18:15 +0200 Subject: [PATCH 350/591] Remove forked marker in client uwsgi test (#4575) don't think this one is needed at all part of #4538 --- tests/test_client.py | 1 - 1 file changed, 1 deletion(-) diff --git a/tests/test_client.py b/tests/test_client.py index 60b15615c8..9c6dbfe740 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -1297,7 +1297,6 @@ def test_error_sampler(_, sentry_init, capture_events, test_config): assert len(test_config.sampler_function_mock.call_args[0]) == 2 -@pytest.mark.forked @pytest.mark.parametrize( "opt,missing_flags", [ From 1cba56ae838117ea3dcb23ac358e3a9ca13f0d25 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 11 Jul 2025 14:18:23 +0200 Subject: [PATCH 351/591] Remove all forked markers in test_api (#4576) again see no reason why we need these part of #4538 --- tests/test_api.py | 10 ---------- 1 file changed, 10 deletions(-) diff --git a/tests/test_api.py b/tests/test_api.py index 08c295a5c4..acc33cdf4c 100644 --- a/tests/test_api.py +++ b/tests/test_api.py @@ -24,7 +24,6 @@ from sentry_sdk.client import Client, NonRecordingClient -@pytest.mark.forked def test_get_current_span(): fake_scope = mock.MagicMock() fake_scope.span = mock.MagicMock() @@ -34,7 +33,6 @@ def test_get_current_span(): assert get_current_span(fake_scope) is None -@pytest.mark.forked def test_get_current_span_default_hub(sentry_init): sentry_init() @@ -47,7 +45,6 @@ def test_get_current_span_default_hub(sentry_init): assert get_current_span() == fake_span -@pytest.mark.forked def test_get_current_span_default_hub_with_transaction(sentry_init): sentry_init() @@ -57,7 +54,6 @@ def test_get_current_span_default_hub_with_transaction(sentry_init): assert get_current_span() == new_transaction -@pytest.mark.forked def test_traceparent_with_tracing_enabled(sentry_init): sentry_init(traces_sample_rate=1.0) @@ -69,7 +65,6 @@ def test_traceparent_with_tracing_enabled(sentry_init): assert get_traceparent() == expected_traceparent -@pytest.mark.forked def test_traceparent_with_tracing_disabled(sentry_init): sentry_init() @@ -81,7 +76,6 @@ def test_traceparent_with_tracing_disabled(sentry_init): assert get_traceparent() == expected_traceparent -@pytest.mark.forked def test_baggage_with_tracing_disabled(sentry_init): sentry_init(release="1.0.0", environment="dev") propagation_context = get_isolation_scope()._propagation_context @@ -93,7 +87,6 @@ def test_baggage_with_tracing_disabled(sentry_init): assert get_baggage() == expected_baggage -@pytest.mark.forked def test_baggage_with_tracing_enabled(sentry_init): sentry_init(traces_sample_rate=1.0, release="1.0.0", environment="dev") with start_transaction() as transaction: @@ -103,7 +96,6 @@ def test_baggage_with_tracing_enabled(sentry_init): assert re.match(expected_baggage_re, get_baggage()) -@pytest.mark.forked def test_continue_trace(sentry_init): sentry_init() @@ -130,7 +122,6 @@ def test_continue_trace(sentry_init): } -@pytest.mark.forked def test_is_initialized(): assert not is_initialized() @@ -139,7 +130,6 @@ def test_is_initialized(): assert is_initialized() -@pytest.mark.forked def test_get_client(): client = get_client() assert client is not None From c31ba06e17eff6daef50c3a7aa738fa7bca7f04f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 15 Jul 2025 10:46:58 +0200 Subject: [PATCH 352/591] tests: Regenerate tox.ini (#4583) --- scripts/populate_tox/populate_tox.py | 3 +- tox.ini | 42 +++++++++++++++------------- 2 files changed, 24 insertions(+), 21 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 3d9e247b4f..3ca5ab18c8 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -508,7 +508,8 @@ def _compare_min_version_with_defined( ): print( f" Integration defines {defined_min_version} as minimum " - f"version, but the effective minimum version is {releases[0]}." + f"version, but the effective minimum version based on metadata " + f"is {releases[0]}." ) diff --git a/tox.ini b/tox.ini index 3b3081b3cb..8af16d640e 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-07-08T06:07:54.743036+00:00 +# Last generated: 2025-07-15T08:21:43.713048+00:00 [tox] requires = @@ -141,9 +141,9 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.57.1 {py3.9,py3.10,py3.11}-cohere-v5.4.0 - {py3.9,py3.11,py3.12}-cohere-v5.8.1 - {py3.9,py3.11,py3.12}-cohere-v5.11.4 - {py3.9,py3.11,py3.12}-cohere-v5.15.0 + {py3.9,py3.11,py3.12}-cohere-v5.9.4 + {py3.9,py3.11,py3.12}-cohere-v5.13.12 + {py3.9,py3.11,py3.12}-cohere-v5.16.1 {py3.9,py3.11,py3.12}-openai_agents-v0.0.19 {py3.9,py3.12,py3.13}-openai_agents-v0.1.0 @@ -151,7 +151,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.4 # ~~~ DBs ~~~ @@ -175,7 +175,7 @@ envlist = {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 - {py3.8,py3.12,py3.13}-launchdarkly-v9.11.1 + {py3.9,py3.12,py3.13}-launchdarkly-v9.12.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.1 @@ -183,7 +183,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.58.3 + {py3.7,py3.12,py3.13}-statsig-v0.59.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -207,7 +207,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.231.1 {py3.8,py3.12,py3.13}-strawberry-v0.253.1 - {py3.9,py3.12,py3.13}-strawberry-v0.275.5 + {py3.9,py3.12,py3.13}-strawberry-v0.276.0 # ~~~ Network ~~~ @@ -215,6 +215,7 @@ envlist = {py3.7,py3.9,py3.10}-grpc-v1.46.5 {py3.7,py3.11,py3.12}-grpc-v1.60.2 {py3.9,py3.12,py3.13}-grpc-v1.73.1 + {py3.9,py3.12,py3.13}-grpc-v1.74.0rc1 # ~~~ Tasks ~~~ @@ -258,14 +259,14 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 {py3.7,py3.10,py3.11}-fastapi-v0.103.2 - {py3.8,py3.12,py3.13}-fastapi-v0.116.0 + {py3.8,py3.12,py3.13}-fastapi-v0.116.1 # ~~~ Web 2 ~~~ {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.12.13 + {py3.9,py3.12,py3.13}-aiohttp-v3.12.14 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -516,9 +517,9 @@ deps = anthropic-v0.44.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 - cohere-v5.8.1: cohere==5.8.1 - cohere-v5.11.4: cohere==5.11.4 - cohere-v5.15.0: cohere==5.15.0 + cohere-v5.9.4: cohere==5.9.4 + cohere-v5.13.12: cohere==5.13.12 + cohere-v5.16.1: cohere==5.16.1 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 @@ -527,7 +528,7 @@ deps = huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.33.2: huggingface_hub==0.33.2 + huggingface_hub-v0.33.4: huggingface_hub==0.33.4 # ~~~ DBs ~~~ @@ -552,7 +553,7 @@ deps = launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 - launchdarkly-v9.11.1: launchdarkly-server-sdk==9.11.1 + launchdarkly-v9.12.0: launchdarkly-server-sdk==9.12.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.1: openfeature-sdk==0.8.1 @@ -560,7 +561,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 - statsig-v0.58.3: statsig==0.58.3 + statsig-v0.59.0: statsig==0.59.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -593,7 +594,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.231.1: strawberry-graphql[fastapi,flask]==0.231.1 strawberry-v0.253.1: strawberry-graphql[fastapi,flask]==0.253.1 - strawberry-v0.275.5: strawberry-graphql[fastapi,flask]==0.275.5 + strawberry-v0.276.0: strawberry-graphql[fastapi,flask]==0.276.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.231.1: pydantic<2.11 @@ -605,6 +606,7 @@ deps = grpc-v1.46.5: grpcio==1.46.5 grpc-v1.60.2: grpcio==1.60.2 grpc-v1.73.1: grpcio==1.73.1 + grpc-v1.74.0rc1: grpcio==1.74.0rc1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -691,7 +693,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.91.0: fastapi==0.91.0 fastapi-v0.103.2: fastapi==0.103.2 - fastapi-v0.116.0: fastapi==0.116.0 + fastapi-v0.116.1: fastapi==0.116.1 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -707,10 +709,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.12.13: aiohttp==3.12.13 + aiohttp-v3.12.14: aiohttp==3.12.14 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.12.13: pytest-asyncio + aiohttp-v3.12.14: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.4: bottle==0.13.4 From a1f62bada1771dd398a89fd384fc421736e5eb84 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 15 Jul 2025 13:58:57 +0200 Subject: [PATCH 353/591] feat(langchain): Support `BaseCallbackManager` (#4486) While implementing #4479, I noticed that our Langchain integration lacks support for the `local_callbacks` having type `BaseCallbackManager`, which according to the type hint is possible. This change adds support for this case. Fixes #4537 --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- sentry_sdk/integrations/langchain.py | 57 +++++--- .../integrations/langchain/test_langchain.py | 131 +++++++++++++++++- 2 files changed, 168 insertions(+), 20 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 0b8bbd8049..4fcc6a571d 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -23,6 +23,7 @@ from langchain_core.callbacks import ( manager, BaseCallbackHandler, + BaseCallbackManager, Callbacks, ) from langchain_core.agents import AgentAction, AgentFinish @@ -434,12 +435,20 @@ def new_configure( **kwargs, ) - callbacks_list = local_callbacks or [] - - if isinstance(callbacks_list, BaseCallbackHandler): - callbacks_list = [callbacks_list] - elif not isinstance(callbacks_list, list): - logger.debug("Unknown callback type: %s", callbacks_list) + local_callbacks = local_callbacks or [] + + # Handle each possible type of local_callbacks. For each type, we + # extract the list of callbacks to check for SentryLangchainCallback, + # and define a function that would add the SentryLangchainCallback + # to the existing callbacks list. + if isinstance(local_callbacks, BaseCallbackManager): + callbacks_list = local_callbacks.handlers + elif isinstance(local_callbacks, BaseCallbackHandler): + callbacks_list = [local_callbacks] + elif isinstance(local_callbacks, list): + callbacks_list = local_callbacks + else: + logger.debug("Unknown callback type: %s", local_callbacks) # Just proceed with original function call return f( callback_manager_cls, @@ -449,28 +458,38 @@ def new_configure( **kwargs, ) - inheritable_callbacks_list = ( - inheritable_callbacks if isinstance(inheritable_callbacks, list) else [] - ) + # Handle each possible type of inheritable_callbacks. + if isinstance(inheritable_callbacks, BaseCallbackManager): + inheritable_callbacks_list = inheritable_callbacks.handlers + elif isinstance(inheritable_callbacks, list): + inheritable_callbacks_list = inheritable_callbacks + else: + inheritable_callbacks_list = [] if not any( isinstance(cb, SentryLangchainCallback) for cb in itertools.chain(callbacks_list, inheritable_callbacks_list) ): - # Avoid mutating the existing callbacks list - callbacks_list = [ - *callbacks_list, - SentryLangchainCallback( - integration.max_spans, - integration.include_prompts, - integration.tiktoken_encoding_name, - ), - ] + sentry_handler = SentryLangchainCallback( + integration.max_spans, + integration.include_prompts, + integration.tiktoken_encoding_name, + ) + if isinstance(local_callbacks, BaseCallbackManager): + local_callbacks = local_callbacks.copy() + local_callbacks.handlers = [ + *local_callbacks.handlers, + sentry_handler, + ] + elif isinstance(local_callbacks, BaseCallbackHandler): + local_callbacks = [local_callbacks, sentry_handler] + else: # local_callbacks is a list + local_callbacks = [*local_callbacks, sentry_handler] return f( callback_manager_cls, inheritable_callbacks, - callbacks_list, + local_callbacks, *args, **kwargs, ) diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index a50a2849c3..ee9fb241b1 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -1,4 +1,5 @@ from typing import List, Optional, Any, Iterator +from unittest import mock from unittest.mock import Mock import pytest @@ -12,7 +13,7 @@ # Langchain < 0.2 from langchain_community.chat_models import ChatOpenAI -from langchain_core.callbacks import CallbackManagerForLLMRun +from langchain_core.callbacks import BaseCallbackManager, CallbackManagerForLLMRun from langchain_core.messages import BaseMessage, AIMessageChunk from langchain_core.outputs import ChatGenerationChunk, ChatResult from langchain_core.runnables import RunnableConfig @@ -428,3 +429,131 @@ def test_span_map_is_instance_variable(): assert ( callback1.span_map is not callback2.span_map ), "span_map should be an instance variable, not shared between instances" + + +def test_langchain_callback_manager(sentry_init): + sentry_init( + integrations=[LangchainIntegration()], + traces_sample_rate=1.0, + ) + local_manager = BaseCallbackManager(handlers=[]) + + with mock.patch("sentry_sdk.integrations.langchain.manager") as mock_manager_module: + mock_configure = mock_manager_module._configure + + # Explicitly re-run setup_once, so that mock_manager_module._configure gets patched + LangchainIntegration.setup_once() + + callback_manager_cls = Mock() + + mock_manager_module._configure( + callback_manager_cls, local_callbacks=local_manager + ) + + assert mock_configure.call_count == 1 + + call_args = mock_configure.call_args + assert call_args.args[0] is callback_manager_cls + + passed_manager = call_args.args[2] + assert passed_manager is not local_manager + assert local_manager.handlers == [] + + [handler] = passed_manager.handlers + assert isinstance(handler, SentryLangchainCallback) + + +def test_langchain_callback_manager_with_sentry_callback(sentry_init): + sentry_init( + integrations=[LangchainIntegration()], + traces_sample_rate=1.0, + ) + sentry_callback = SentryLangchainCallback(0, False) + local_manager = BaseCallbackManager(handlers=[sentry_callback]) + + with mock.patch("sentry_sdk.integrations.langchain.manager") as mock_manager_module: + mock_configure = mock_manager_module._configure + + # Explicitly re-run setup_once, so that mock_manager_module._configure gets patched + LangchainIntegration.setup_once() + + callback_manager_cls = Mock() + + mock_manager_module._configure( + callback_manager_cls, local_callbacks=local_manager + ) + + assert mock_configure.call_count == 1 + + call_args = mock_configure.call_args + assert call_args.args[0] is callback_manager_cls + + passed_manager = call_args.args[2] + assert passed_manager is local_manager + + [handler] = passed_manager.handlers + assert handler is sentry_callback + + +def test_langchain_callback_list(sentry_init): + sentry_init( + integrations=[LangchainIntegration()], + traces_sample_rate=1.0, + ) + local_callbacks = [] + + with mock.patch("sentry_sdk.integrations.langchain.manager") as mock_manager_module: + mock_configure = mock_manager_module._configure + + # Explicitly re-run setup_once, so that mock_manager_module._configure gets patched + LangchainIntegration.setup_once() + + callback_manager_cls = Mock() + + mock_manager_module._configure( + callback_manager_cls, local_callbacks=local_callbacks + ) + + assert mock_configure.call_count == 1 + + call_args = mock_configure.call_args + assert call_args.args[0] is callback_manager_cls + + passed_callbacks = call_args.args[2] + assert passed_callbacks is not local_callbacks + assert local_callbacks == [] + + [handler] = passed_callbacks + assert isinstance(handler, SentryLangchainCallback) + + +def test_langchain_callback_list_existing_callback(sentry_init): + sentry_init( + integrations=[LangchainIntegration()], + traces_sample_rate=1.0, + ) + sentry_callback = SentryLangchainCallback(0, False) + local_callbacks = [sentry_callback] + + with mock.patch("sentry_sdk.integrations.langchain.manager") as mock_manager_module: + mock_configure = mock_manager_module._configure + + # Explicitly re-run setup_once, so that mock_manager_module._configure gets patched + LangchainIntegration.setup_once() + + callback_manager_cls = Mock() + + mock_manager_module._configure( + callback_manager_cls, local_callbacks=local_callbacks + ) + + assert mock_configure.call_count == 1 + + call_args = mock_configure.call_args + assert call_args.args[0] is callback_manager_cls + + passed_callbacks = call_args.args[2] + assert passed_callbacks is local_callbacks + + [handler] = passed_callbacks + assert handler is sentry_callback From 220a235bdc9c9c14ed4aa8629f0768016d959a78 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 15 Jul 2025 12:00:46 +0000 Subject: [PATCH 354/591] release: 2.33.0 --- CHANGELOG.md | 18 ++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 21 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index fd4a98e717..7b84742017 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,23 @@ # Changelog +## 2.33.0 + +### Various fixes & improvements + +- feat(langchain): Support `BaseCallbackManager` (#4486) by @szokeasaurusrex +- tests: Regenerate tox.ini (#4583) by @sentrivana +- Remove all forked markers in test_api (#4576) by @sl0thentr0py +- Remove forked marker in client uwsgi test (#4575) by @sl0thentr0py +- Fix pytest collection warning (#4574) by @sl0thentr0py +- Remove print statements from excepthook test (#4573) by @sl0thentr0py +- Use `span.data` instead of `measurements` for token usage (#4567) by @antonpirker +- Fix custom model name (#4569) by @antonpirker +- fix: shut down "session flusher" more promptly (#4561) by @bukzor +- test(loguru): Remove hardcoded line number in test_just_log (#4552) by @srothh +- toxgen: Detect correct sentry-sdk (#4558) by @sentrivana +- tests: Tox update (#4555) by @sentrivana +- chore: Remove Lambda urllib3 pin on Python 3.10+ (#4549) by @sentrivana + ## 2.32.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index ea5995ee36..cc5131636b 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.32.0" +release = "2.33.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 01f72e2887..9dc1de9bb7 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1181,4 +1181,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.32.0" +VERSION = "2.33.0" diff --git a/setup.py b/setup.py index ae86cab158..213fcdb597 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.32.0", + version="2.33.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 98b107fd2162f367da7002ccef36714976878fe3 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Tue, 15 Jul 2025 14:02:18 +0200 Subject: [PATCH 355/591] meta: Update CHANGELOG.md Remove non-user-facing changes --- CHANGELOG.md | 8 -------- 1 file changed, 8 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 7b84742017..b9e8bb6046 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,17 +5,9 @@ ### Various fixes & improvements - feat(langchain): Support `BaseCallbackManager` (#4486) by @szokeasaurusrex -- tests: Regenerate tox.ini (#4583) by @sentrivana -- Remove all forked markers in test_api (#4576) by @sl0thentr0py -- Remove forked marker in client uwsgi test (#4575) by @sl0thentr0py -- Fix pytest collection warning (#4574) by @sl0thentr0py -- Remove print statements from excepthook test (#4573) by @sl0thentr0py - Use `span.data` instead of `measurements` for token usage (#4567) by @antonpirker - Fix custom model name (#4569) by @antonpirker - fix: shut down "session flusher" more promptly (#4561) by @bukzor -- test(loguru): Remove hardcoded line number in test_just_log (#4552) by @srothh -- toxgen: Detect correct sentry-sdk (#4558) by @sentrivana -- tests: Tox update (#4555) by @sentrivana - chore: Remove Lambda urllib3 pin on Python 3.10+ (#4549) by @sentrivana ## 2.32.0 From 2ccab61dfd5dee9135d9b7e5fb25d184f5dfd52c Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 15 Jul 2025 16:54:02 +0200 Subject: [PATCH 356/591] Improve token usage recording (#4566) Update token usage recording to work if the LLM is calling them `prompt_tokens` or `input_tokens`. Same for `completion_tokens` and `output_tokens`. Records also cached and reasoning tokens usage. Because the signature of a helper function was changed, other AI integrations also have changes. This PR does not change behavior, just prepare the ground for future changes to the AI integrations. --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/ai/monitoring.py | 39 +++++--- sentry_sdk/integrations/anthropic.py | 15 ++- sentry_sdk/integrations/cohere.py | 10 +- sentry_sdk/integrations/huggingface_hub.py | 10 +- sentry_sdk/integrations/langchain.py | 10 +- sentry_sdk/integrations/openai.py | 103 +++++++++++++-------- tests/integrations/openai/test_openai.py | 67 ++++++++++---- 7 files changed, 174 insertions(+), 80 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index d3154f0631..461fd6af85 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -96,25 +96,40 @@ async def async_wrapped(*args, **kwargs): def record_token_usage( - span, prompt_tokens=None, completion_tokens=None, total_tokens=None + span, + input_tokens=None, + input_tokens_cached=None, + output_tokens=None, + output_tokens_reasoning=None, + total_tokens=None, ): - # type: (Span, Optional[int], Optional[int], Optional[int]) -> None + # type: (Span, Optional[int], Optional[int], Optional[int], Optional[int], Optional[int]) -> None + + # TODO: move pipeline name elsewhere ai_pipeline_name = get_ai_pipeline_name() if ai_pipeline_name: span.set_data(SPANDATA.AI_PIPELINE_NAME, ai_pipeline_name) - if prompt_tokens is not None: - span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, prompt_tokens) + if input_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, input_tokens) + + if input_tokens_cached is not None: + span.set_data( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED, + input_tokens_cached, + ) + + if output_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, output_tokens) - if completion_tokens is not None: - span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, completion_tokens) + if output_tokens_reasoning is not None: + span.set_data( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING, + output_tokens_reasoning, + ) - if ( - total_tokens is None - and prompt_tokens is not None - and completion_tokens is not None - ): - total_tokens = prompt_tokens + completion_tokens + if total_tokens is None and input_tokens is not None and output_tokens is not None: + total_tokens = input_tokens + output_tokens if total_tokens is not None: span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, total_tokens) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 76a3bb9f13..1e1f9112a1 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -65,7 +65,13 @@ def _calculate_token_usage(result, span): output_tokens = usage.output_tokens total_tokens = input_tokens + output_tokens - record_token_usage(span, input_tokens, output_tokens, total_tokens) + + record_token_usage( + span, + input_tokens=input_tokens, + output_tokens=output_tokens, + total_tokens=total_tokens, + ) def _get_responses(content): @@ -126,7 +132,12 @@ def _add_ai_data_to_span( [{"type": "text", "text": complete_message}], ) total_tokens = input_tokens + output_tokens - record_token_usage(span, input_tokens, output_tokens, total_tokens) + record_token_usage( + span, + input_tokens=input_tokens, + output_tokens=output_tokens, + total_tokens=total_tokens, + ) span.set_data(SPANDATA.AI_STREAMING, True) diff --git a/sentry_sdk/integrations/cohere.py b/sentry_sdk/integrations/cohere.py index 433b285bf0..57ffdb908a 100644 --- a/sentry_sdk/integrations/cohere.py +++ b/sentry_sdk/integrations/cohere.py @@ -116,14 +116,14 @@ def collect_chat_response_fields(span, res, include_pii): if hasattr(res.meta, "billed_units"): record_token_usage( span, - prompt_tokens=res.meta.billed_units.input_tokens, - completion_tokens=res.meta.billed_units.output_tokens, + input_tokens=res.meta.billed_units.input_tokens, + output_tokens=res.meta.billed_units.output_tokens, ) elif hasattr(res.meta, "tokens"): record_token_usage( span, - prompt_tokens=res.meta.tokens.input_tokens, - completion_tokens=res.meta.tokens.output_tokens, + input_tokens=res.meta.tokens.input_tokens, + output_tokens=res.meta.tokens.output_tokens, ) if hasattr(res.meta, "warnings"): @@ -262,7 +262,7 @@ def new_embed(*args, **kwargs): ): record_token_usage( span, - prompt_tokens=res.meta.billed_units.input_tokens, + input_tokens=res.meta.billed_units.input_tokens, total_tokens=res.meta.billed_units.input_tokens, ) return res diff --git a/sentry_sdk/integrations/huggingface_hub.py b/sentry_sdk/integrations/huggingface_hub.py index dfac77e996..2dfcb5925a 100644 --- a/sentry_sdk/integrations/huggingface_hub.py +++ b/sentry_sdk/integrations/huggingface_hub.py @@ -111,7 +111,10 @@ def new_text_generation(*args, **kwargs): [res.generated_text], ) if res.details is not None and res.details.generated_tokens > 0: - record_token_usage(span, total_tokens=res.details.generated_tokens) + record_token_usage( + span, + total_tokens=res.details.generated_tokens, + ) span.__exit__(None, None, None) return res @@ -145,7 +148,10 @@ def new_details_iterator(): span, SPANDATA.AI_RESPONSES, "".join(data_buf) ) if tokens_used > 0: - record_token_usage(span, total_tokens=tokens_used) + record_token_usage( + span, + total_tokens=tokens_used, + ) span.__exit__(None, None, None) return new_details_iterator() diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 4fcc6a571d..8b67c4c994 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -279,15 +279,15 @@ def on_llm_end(self, response, *, run_id, **kwargs): if token_usage: record_token_usage( span_data.span, - token_usage.get("prompt_tokens"), - token_usage.get("completion_tokens"), - token_usage.get("total_tokens"), + input_tokens=token_usage.get("prompt_tokens"), + output_tokens=token_usage.get("completion_tokens"), + total_tokens=token_usage.get("total_tokens"), ) else: record_token_usage( span_data.span, - span_data.num_prompt_tokens, - span_data.num_completion_tokens, + input_tokens=span_data.num_prompt_tokens, + output_tokens=span_data.num_completion_tokens, ) self._exit_span(span_data, run_id) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index e95753f6e1..d906a8e0b2 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -70,48 +70,73 @@ def _capture_exception(exc): sentry_sdk.capture_event(event, hint=hint) -def _calculate_chat_completion_usage( +def _get_usage(usage, names): + # type: (Any, List[str]) -> int + for name in names: + if hasattr(usage, name) and isinstance(getattr(usage, name), int): + return getattr(usage, name) + return 0 + + +def _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ): # type: (Iterable[ChatCompletionMessageParam], Any, Span, Optional[List[str]], Callable[..., Any]) -> None - completion_tokens = 0 # type: Optional[int] - prompt_tokens = 0 # type: Optional[int] + input_tokens = 0 # type: Optional[int] + input_tokens_cached = 0 # type: Optional[int] + output_tokens = 0 # type: Optional[int] + output_tokens_reasoning = 0 # type: Optional[int] total_tokens = 0 # type: Optional[int] + if hasattr(response, "usage"): - if hasattr(response.usage, "completion_tokens") and isinstance( - response.usage.completion_tokens, int - ): - completion_tokens = response.usage.completion_tokens - if hasattr(response.usage, "prompt_tokens") and isinstance( - response.usage.prompt_tokens, int - ): - prompt_tokens = response.usage.prompt_tokens - if hasattr(response.usage, "total_tokens") and isinstance( - response.usage.total_tokens, int - ): - total_tokens = response.usage.total_tokens + input_tokens = _get_usage(response.usage, ["input_tokens", "prompt_tokens"]) + if hasattr(response.usage, "input_tokens_details"): + input_tokens_cached = _get_usage( + response.usage.input_tokens_details, ["cached_tokens"] + ) - if prompt_tokens == 0: + output_tokens = _get_usage( + response.usage, ["output_tokens", "completion_tokens"] + ) + if hasattr(response.usage, "output_tokens_details"): + output_tokens_reasoning = _get_usage( + response.usage.output_tokens_details, ["reasoning_tokens"] + ) + + total_tokens = _get_usage(response.usage, ["total_tokens"]) + + # Manually count tokens + # TODO: when implementing responses API, check for responses API + if input_tokens == 0: for message in messages: if "content" in message: - prompt_tokens += count_tokens(message["content"]) + input_tokens += count_tokens(message["content"]) - if completion_tokens == 0: + # TODO: when implementing responses API, check for responses API + if output_tokens == 0: if streaming_message_responses is not None: for message in streaming_message_responses: - completion_tokens += count_tokens(message) + output_tokens += count_tokens(message) elif hasattr(response, "choices"): for choice in response.choices: if hasattr(choice, "message"): - completion_tokens += count_tokens(choice.message) - - if prompt_tokens == 0: - prompt_tokens = None - if completion_tokens == 0: - completion_tokens = None - if total_tokens == 0: - total_tokens = None - record_token_usage(span, prompt_tokens, completion_tokens, total_tokens) + output_tokens += count_tokens(choice.message) + + # Do not set token data if it is 0 + input_tokens = input_tokens or None + input_tokens_cached = input_tokens_cached or None + output_tokens = output_tokens or None + output_tokens_reasoning = output_tokens_reasoning or None + total_tokens = total_tokens or None + + record_token_usage( + span, + input_tokens=input_tokens, + input_tokens_cached=input_tokens_cached, + output_tokens=output_tokens, + output_tokens_reasoning=output_tokens_reasoning, + total_tokens=total_tokens, + ) def _new_chat_completion_common(f, *args, **kwargs): @@ -158,9 +183,7 @@ def _new_chat_completion_common(f, *args, **kwargs): SPANDATA.AI_RESPONSES, list(map(lambda x: x.message, res.choices)), ) - _calculate_chat_completion_usage( - messages, res, span, None, integration.count_tokens - ) + _calculate_token_usage(messages, res, span, None, integration.count_tokens) span.__exit__(None, None, None) elif hasattr(res, "_iterator"): data_buf: list[list[str]] = [] # one for each choice @@ -191,7 +214,7 @@ def new_iterator(): set_data_normalized( span, SPANDATA.AI_RESPONSES, all_responses ) - _calculate_chat_completion_usage( + _calculate_token_usage( messages, res, span, @@ -224,7 +247,7 @@ async def new_iterator_async(): set_data_normalized( span, SPANDATA.AI_RESPONSES, all_responses ) - _calculate_chat_completion_usage( + _calculate_token_usage( messages, res, span, @@ -341,22 +364,26 @@ def _new_embeddings_create_common(f, *args, **kwargs): response = yield f, args, kwargs - prompt_tokens = 0 + input_tokens = 0 total_tokens = 0 if hasattr(response, "usage"): if hasattr(response.usage, "prompt_tokens") and isinstance( response.usage.prompt_tokens, int ): - prompt_tokens = response.usage.prompt_tokens + input_tokens = response.usage.prompt_tokens if hasattr(response.usage, "total_tokens") and isinstance( response.usage.total_tokens, int ): total_tokens = response.usage.total_tokens - if prompt_tokens == 0: - prompt_tokens = integration.count_tokens(kwargs["input"] or "") + if input_tokens == 0: + input_tokens = integration.count_tokens(kwargs["input"] or "") - record_token_usage(span, prompt_tokens, None, total_tokens or prompt_tokens) + record_token_usage( + span, + input_tokens=input_tokens, + total_tokens=total_tokens or input_tokens, + ) return response diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 39195de277..ac6d9f4c29 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -10,7 +10,7 @@ from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.openai import ( OpenAIIntegration, - _calculate_chat_completion_usage, + _calculate_token_usage, ) from unittest import mock # python 3.3 and above @@ -743,7 +743,7 @@ async def test_span_origin_embeddings_async(sentry_init, capture_events): assert event["spans"][0]["origin"] == "auto.ai.openai" -def test_calculate_chat_completion_usage_a(): +def test_calculate_token_usage_a(): span = mock.MagicMock() def count_tokens(msg): @@ -760,13 +760,20 @@ def count_tokens(msg): with mock.patch( "sentry_sdk.integrations.openai.record_token_usage" ) as mock_record_token_usage: - _calculate_chat_completion_usage( + _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ) - mock_record_token_usage.assert_called_once_with(span, 20, 10, 30) + mock_record_token_usage.assert_called_once_with( + span, + input_tokens=20, + input_tokens_cached=None, + output_tokens=10, + output_tokens_reasoning=None, + total_tokens=30, + ) -def test_calculate_chat_completion_usage_b(): +def test_calculate_token_usage_b(): span = mock.MagicMock() def count_tokens(msg): @@ -786,13 +793,20 @@ def count_tokens(msg): with mock.patch( "sentry_sdk.integrations.openai.record_token_usage" ) as mock_record_token_usage: - _calculate_chat_completion_usage( + _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ) - mock_record_token_usage.assert_called_once_with(span, 11, 10, 10) + mock_record_token_usage.assert_called_once_with( + span, + input_tokens=11, + input_tokens_cached=None, + output_tokens=10, + output_tokens_reasoning=None, + total_tokens=10, + ) -def test_calculate_chat_completion_usage_c(): +def test_calculate_token_usage_c(): span = mock.MagicMock() def count_tokens(msg): @@ -812,13 +826,20 @@ def count_tokens(msg): with mock.patch( "sentry_sdk.integrations.openai.record_token_usage" ) as mock_record_token_usage: - _calculate_chat_completion_usage( + _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ) - mock_record_token_usage.assert_called_once_with(span, 20, 11, 20) + mock_record_token_usage.assert_called_once_with( + span, + input_tokens=20, + input_tokens_cached=None, + output_tokens=11, + output_tokens_reasoning=None, + total_tokens=20, + ) -def test_calculate_chat_completion_usage_d(): +def test_calculate_token_usage_d(): span = mock.MagicMock() def count_tokens(msg): @@ -839,13 +860,20 @@ def count_tokens(msg): with mock.patch( "sentry_sdk.integrations.openai.record_token_usage" ) as mock_record_token_usage: - _calculate_chat_completion_usage( + _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ) - mock_record_token_usage.assert_called_once_with(span, 20, None, 20) + mock_record_token_usage.assert_called_once_with( + span, + input_tokens=20, + input_tokens_cached=None, + output_tokens=None, + output_tokens_reasoning=None, + total_tokens=20, + ) -def test_calculate_chat_completion_usage_e(): +def test_calculate_token_usage_e(): span = mock.MagicMock() def count_tokens(msg): @@ -858,7 +886,14 @@ def count_tokens(msg): with mock.patch( "sentry_sdk.integrations.openai.record_token_usage" ) as mock_record_token_usage: - _calculate_chat_completion_usage( + _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ) - mock_record_token_usage.assert_called_once_with(span, None, None, None) + mock_record_token_usage.assert_called_once_with( + span, + input_tokens=None, + input_tokens_cached=None, + output_tokens=None, + output_tokens_reasoning=None, + total_tokens=None, + ) From 9b66f3b51502ca600554c711bc3f599c18f8f18b Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 17 Jul 2025 13:53:37 +0200 Subject: [PATCH 357/591] Remove forked from test_transport, separate gevent tests and generalize capturing_server to be module level (#4577) * Move `CapturingServer` to `conftest` for reuse * Make `capturing_server` in original non-forked transport tests to be module level for faster tests * Move `gevent` based transport tests to separate file since they still need to be forked because of the global monkeypatch As a result, `test_transport` now takes 13 seconds instead of almost 3 minutes. BEFORE ``` 235 passed, 384 skipped in 173.78s (0:02:53) ``` AFTER ``` 235 passed in 12.69s ``` part of #4538 --- tests/conftest.py | 62 ++++++++++++++++++++- tests/test_gevent.py | 118 ++++++++++++++++++++++++++++++++++++++++ tests/test_transport.py | 84 +++++----------------------- 3 files changed, 192 insertions(+), 72 deletions(-) create mode 100644 tests/test_gevent.py diff --git a/tests/conftest.py b/tests/conftest.py index 6a33029d11..01b1e9a81f 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -2,12 +2,18 @@ import os import socket import warnings +import brotli +import gzip +import io from threading import Thread from contextlib import contextmanager from http.server import BaseHTTPRequestHandler, HTTPServer from unittest import mock +from collections import namedtuple import pytest +from pytest_localserver.http import WSGIServer +from werkzeug.wrappers import Request, Response import jsonschema @@ -23,7 +29,7 @@ import sentry_sdk import sentry_sdk.utils -from sentry_sdk.envelope import Envelope +from sentry_sdk.envelope import Envelope, parse_json from sentry_sdk.integrations import ( # noqa: F401 _DEFAULT_INTEGRATIONS, _installed_integrations, @@ -663,3 +669,57 @@ def __eq__(self, other): def __ne__(self, other): return not self.__eq__(other) + + +CapturedData = namedtuple("CapturedData", ["path", "event", "envelope", "compressed"]) + + +class CapturingServer(WSGIServer): + def __init__(self, host="127.0.0.1", port=0, ssl_context=None): + WSGIServer.__init__(self, host, port, self, ssl_context=ssl_context) + self.code = 204 + self.headers = {} + self.captured = [] + + def respond_with(self, code=200, headers=None): + self.code = code + if headers: + self.headers = headers + + def clear_captured(self): + del self.captured[:] + + def __call__(self, environ, start_response): + """ + This is the WSGI application. + """ + request = Request(environ) + event = envelope = None + content_encoding = request.headers.get("content-encoding") + if content_encoding == "gzip": + rdr = gzip.GzipFile(fileobj=io.BytesIO(request.data)) + compressed = True + elif content_encoding == "br": + rdr = io.BytesIO(brotli.decompress(request.data)) + compressed = True + else: + rdr = io.BytesIO(request.data) + compressed = False + + if request.mimetype == "application/json": + event = parse_json(rdr.read()) + else: + envelope = Envelope.deserialize_from(rdr) + + self.captured.append( + CapturedData( + path=request.path, + event=event, + envelope=envelope, + compressed=compressed, + ) + ) + + response = Response(status=self.code) + response.headers.extend(self.headers) + return response(environ, start_response) diff --git a/tests/test_gevent.py b/tests/test_gevent.py new file mode 100644 index 0000000000..d330760adf --- /dev/null +++ b/tests/test_gevent.py @@ -0,0 +1,118 @@ +import logging +import pickle +from datetime import datetime, timezone + +import sentry_sdk +from sentry_sdk._compat import PY37, PY38 + +import pytest +from tests.conftest import CapturingServer + +pytest.importorskip("gevent") + + +@pytest.fixture(scope="module") +def monkeypatched_gevent(): + try: + import gevent + + gevent.monkey.patch_all() + except Exception as e: + if "_RLock__owner" in str(e): + pytest.skip("https://github.com/gevent/gevent/issues/1380") + else: + raise + + +@pytest.fixture +def capturing_server(request): + server = CapturingServer() + server.start() + request.addfinalizer(server.stop) + return server + + +@pytest.fixture +def make_client(request, capturing_server): + def inner(**kwargs): + return sentry_sdk.Client( + "http://foobar@{}/132".format(capturing_server.url[len("http://") :]), + **kwargs, + ) + + return inner + + +@pytest.mark.forked +@pytest.mark.parametrize("debug", (True, False)) +@pytest.mark.parametrize("client_flush_method", ["close", "flush"]) +@pytest.mark.parametrize("use_pickle", (True, False)) +@pytest.mark.parametrize("compression_level", (0, 9, None)) +@pytest.mark.parametrize( + "compression_algo", + (("gzip", "br", "", None) if PY37 else ("gzip", "", None)), +) +@pytest.mark.parametrize("http2", [True, False] if PY38 else [False]) +def test_transport_works_gevent( + capturing_server, + request, + capsys, + caplog, + debug, + make_client, + client_flush_method, + use_pickle, + compression_level, + compression_algo, + http2, +): + caplog.set_level(logging.DEBUG) + + experiments = {} + if compression_level is not None: + experiments["transport_compression_level"] = compression_level + + if compression_algo is not None: + experiments["transport_compression_algo"] = compression_algo + + if http2: + experiments["transport_http2"] = True + + client = make_client( + debug=debug, + _experiments=experiments, + ) + + if use_pickle: + client = pickle.loads(pickle.dumps(client)) + + sentry_sdk.get_global_scope().set_client(client) + request.addfinalizer(lambda: sentry_sdk.get_global_scope().set_client(None)) + + sentry_sdk.add_breadcrumb( + level="info", message="i like bread", timestamp=datetime.now(timezone.utc) + ) + sentry_sdk.capture_message("löl") + + getattr(client, client_flush_method)() + + out, err = capsys.readouterr() + assert not err and not out + assert capturing_server.captured + should_compress = ( + # default is to compress with brotli if available, gzip otherwise + (compression_level is None) + or ( + # setting compression level to 0 means don't compress + compression_level + > 0 + ) + ) and ( + # if we couldn't resolve to a known algo, we don't compress + compression_algo + != "" + ) + + assert capturing_server.captured[0].compressed == should_compress + + assert any("Sending envelope" in record.msg for record in caplog.records) == debug diff --git a/tests/test_transport.py b/tests/test_transport.py index 6eb7cdf829..c6a1a0a7a7 100644 --- a/tests/test_transport.py +++ b/tests/test_transport.py @@ -1,29 +1,20 @@ import logging import pickle -import gzip -import io import os import socket import sys -from collections import defaultdict, namedtuple +from collections import defaultdict from datetime import datetime, timedelta, timezone from unittest import mock -import brotli import pytest -from pytest_localserver.http import WSGIServer -from werkzeug.wrappers import Request, Response +from tests.conftest import CapturingServer try: import httpcore except (ImportError, ModuleNotFoundError): httpcore = None -try: - import gevent -except ImportError: - gevent = None - import sentry_sdk from sentry_sdk import ( Client, @@ -42,65 +33,22 @@ ) from sentry_sdk.integrations.logging import LoggingIntegration, ignore_logger -CapturedData = namedtuple("CapturedData", ["path", "event", "envelope", "compressed"]) - - -class CapturingServer(WSGIServer): - def __init__(self, host="127.0.0.1", port=0, ssl_context=None): - WSGIServer.__init__(self, host, port, self, ssl_context=ssl_context) - self.code = 204 - self.headers = {} - self.captured = [] - - def respond_with(self, code=200, headers=None): - self.code = code - if headers: - self.headers = headers - - def clear_captured(self): - del self.captured[:] - - def __call__(self, environ, start_response): - """ - This is the WSGI application. - """ - request = Request(environ) - event = envelope = None - content_encoding = request.headers.get("content-encoding") - if content_encoding == "gzip": - rdr = gzip.GzipFile(fileobj=io.BytesIO(request.data)) - compressed = True - elif content_encoding == "br": - rdr = io.BytesIO(brotli.decompress(request.data)) - compressed = True - else: - rdr = io.BytesIO(request.data) - compressed = False - - if request.mimetype == "application/json": - event = parse_json(rdr.read()) - else: - envelope = Envelope.deserialize_from(rdr) - - self.captured.append( - CapturedData( - path=request.path, - event=event, - envelope=envelope, - compressed=compressed, - ) - ) - response = Response(status=self.code) - response.headers.extend(self.headers) - return response(environ, start_response) +server = None -@pytest.fixture -def capturing_server(request): +@pytest.fixture(scope="module", autouse=True) +def make_capturing_server(request): + global server server = CapturingServer() server.start() request.addfinalizer(server.stop) + + +@pytest.fixture +def capturing_server(): + global server + server.clear_captured() return server @@ -129,18 +77,13 @@ def mock_transaction_envelope(span_count): return envelope -@pytest.mark.forked @pytest.mark.parametrize("debug", (True, False)) @pytest.mark.parametrize("client_flush_method", ["close", "flush"]) @pytest.mark.parametrize("use_pickle", (True, False)) @pytest.mark.parametrize("compression_level", (0, 9, None)) @pytest.mark.parametrize( "compression_algo", - ( - ("gzip", "br", "", None) - if PY37 or gevent is None - else ("gzip", "", None) - ), + (("gzip", "br", "", None) if PY37 else ("gzip", "", None)), ) @pytest.mark.parametrize("http2", [True, False] if PY38 else [False]) def test_transport_works( @@ -155,7 +98,6 @@ def test_transport_works( compression_level, compression_algo, http2, - maybe_monkeypatched_threading, ): caplog.set_level(logging.DEBUG) From 89e624a1b1da268954e55989c6b37f6e2f2d923b Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 17 Jul 2025 16:13:38 +0200 Subject: [PATCH 358/591] test: Remove `test_installed_modules` (#4593) The test `test_installed_modules` appears to not be all that useful. The test exists to verify the behavior of the [`_generate_installed_modules` function](https://github.com/getsentry/sentry-python/blob/9b66f3b51502ca600554c711bc3f599c18f8f18b/sentry_sdk/utils.py#L1689). However, all the test does is essentially check the output of `_generate_installed_modules` against a refactored version of the function call itself. In short, in its current form, the test appears to not make too much sense. As the test recently started failing, let's just delete it. --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- tests/test_utils.py | 42 ------------------------------------------ 1 file changed, 42 deletions(-) diff --git a/tests/test_utils.py b/tests/test_utils.py index efa2e7c068..b268fbd57b 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -30,7 +30,6 @@ serialize_frame, is_sentry_url, _get_installed_modules, - _generate_installed_modules, ensure_integration_enabled, to_string, exc_info_from_error, @@ -667,47 +666,6 @@ def __str__(self): assert result == repr(obj) -def test_installed_modules(): - try: - from importlib.metadata import distributions, version - - importlib_available = True - except ImportError: - importlib_available = False - - try: - import pkg_resources - - pkg_resources_available = True - except ImportError: - pkg_resources_available = False - - installed_distributions = { - _normalize_distribution_name(dist): version - for dist, version in _generate_installed_modules() - } - - if importlib_available: - importlib_distributions = { - _normalize_distribution_name(dist.metadata.get("Name", None)): version( - dist.metadata.get("Name", None) - ) - for dist in distributions() - if dist.metadata.get("Name", None) is not None - and version(dist.metadata.get("Name", None)) is not None - } - assert installed_distributions == importlib_distributions - - elif pkg_resources_available: - pkg_resources_distributions = { - _normalize_distribution_name(dist.key): dist.version - for dist in pkg_resources.working_set - } - assert installed_distributions == pkg_resources_distributions - else: - pytest.fail("Neither importlib nor pkg_resources is available") - - def test_installed_modules_caching(): mock_generate_installed_modules = mock.Mock() mock_generate_installed_modules.return_value = {"package": "1.0.0"} From da8332a414c177f5b470c8a7ddd7eaebf44cadd4 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 17 Jul 2025 16:23:35 +0200 Subject: [PATCH 359/591] ci: Check strictly for success (#4589) Previously, some statuses other than success (e.g. cancelled) were considered passing for the `check_required_tests` job. See, for example, [this job run](https://github.com/getsentry/sentry-python/actions/runs/16342793741/job/46172510569?pr=4572), where several tests timed out, ending with cancelled status, but the final check passed. We can instead check directly against the `success` status, to strictly enforce that all tests must pass. --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 2 +- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 2 +- .github/workflows/test-integrations-tasks.yml | 2 +- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/split_tox_gh_actions/templates/check_required.jinja | 2 +- 13 files changed, 13 insertions(+), 13 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index e81d507d27..2777810a8f 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -188,6 +188,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-ai-pinned.result, 'failure') || contains(needs.test-ai-pinned.result, 'skipped') + if: needs.test-ai-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 114e904d4b..6a9b9df0de 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -188,6 +188,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-cloud-pinned.result, 'failure') || contains(needs.test-cloud-pinned.result, 'skipped') + if: needs.test-cloud-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 2ac8d827fa..2ceb23b79c 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -89,6 +89,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-common-pinned.result, 'failure') || contains(needs.test-common-pinned.result, 'skipped') + if: needs.test-common-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 460ffe1ad5..1ad39421d6 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -228,6 +228,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-dbs-pinned.result, 'failure') || contains(needs.test-dbs-pinned.result, 'skipped') + if: needs.test-dbs-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 0e2c9ef166..d6da6c8acd 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -101,6 +101,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-flags-pinned.result, 'failure') || contains(needs.test-flags-pinned.result, 'skipped') + if: needs.test-flags-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 3e0903e2c5..c0bd099e45 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -89,6 +89,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-gevent-pinned.result, 'failure') || contains(needs.test-gevent-pinned.result, 'skipped') + if: needs.test-gevent-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 51ae8a8a81..e851dfc9bb 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -101,6 +101,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-graphql-pinned.result, 'failure') || contains(needs.test-graphql-pinned.result, 'skipped') + if: needs.test-graphql-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 05a8aaeda1..8a2e87c9ca 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -109,6 +109,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-misc-pinned.result, 'failure') || contains(needs.test-misc-pinned.result, 'skipped') + if: needs.test-misc-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 769a95c08a..47ae674934 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -164,6 +164,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-network-pinned.result, 'failure') || contains(needs.test-network-pinned.result, 'skipped') + if: needs.test-network-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 91b47f90c6..6b3fcab41f 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -218,6 +218,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-tasks-pinned.result, 'failure') || contains(needs.test-tasks-pinned.result, 'skipped') + if: needs.test-tasks-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 67669c729b..3b48472d5e 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -119,6 +119,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-web_1-pinned.result, 'failure') || contains(needs.test-web_1-pinned.result, 'skipped') + if: needs.test-web_1-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index c0438dc924..b98e5f02fc 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -220,6 +220,6 @@ jobs: runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-web_2-pinned.result, 'failure') || contains(needs.test-web_2-pinned.result, 'skipped') + if: needs.test-web_2-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/scripts/split_tox_gh_actions/templates/check_required.jinja b/scripts/split_tox_gh_actions/templates/check_required.jinja index a2ca2db26e..9a2bbed830 100644 --- a/scripts/split_tox_gh_actions/templates/check_required.jinja +++ b/scripts/split_tox_gh_actions/templates/check_required.jinja @@ -8,6 +8,6 @@ runs-on: ubuntu-22.04 steps: - name: Check for failures - if: contains(needs.test-{{ lowercase_group }}-pinned.result, 'failure') || contains(needs.test-{{ lowercase_group }}-pinned.result, 'skipped') + if: needs.test-{{ lowercase_group }}-pinned.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 From 34dcba4acca5d1b33656e4fa5fe87d9c7816de34 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 17 Jul 2025 16:41:11 +0200 Subject: [PATCH 360/591] Remove explicit __del__'s in threaded classes (#4590) The changes in #4577 introduced a bit of flakiness on pre-3.10 due to a weird interaction of `capsys`, `stderr` logging and our object lifecycles. In this PR, I'm removing all the explicit `__del__` `kill`s since we [call them all explicitly in `client.close`](https://github.com/getsentry/sentry-python/blob/09c2e32cc7a618e49f5d8ae59e22d8b12f253687/sentry_sdk/client.py#L1001-L1021) anyway and that's already cleaner. Having logic in `__del__` causes non-deterministic GC behavior, especially with threaded code. Stacktrace is linked in a comment below where you can see the `transport.__del__` method is causing the weird `reentrant` logging bug to `stderr`. --- sentry_sdk/monitor.py | 4 ---- sentry_sdk/sessions.py | 4 ---- sentry_sdk/transport.py | 7 ------- 3 files changed, 15 deletions(-) diff --git a/sentry_sdk/monitor.py b/sentry_sdk/monitor.py index 68d9017bf9..b82a528851 100644 --- a/sentry_sdk/monitor.py +++ b/sentry_sdk/monitor.py @@ -118,7 +118,3 @@ def downsample_factor(self): def kill(self): # type: () -> None self._running = False - - def __del__(self): - # type: () -> None - self.kill() diff --git a/sentry_sdk/sessions.py b/sentry_sdk/sessions.py index a5dd589ee9..00fda23200 100644 --- a/sentry_sdk/sessions.py +++ b/sentry_sdk/sessions.py @@ -271,7 +271,3 @@ def add_session( def kill(self): # type: (...) -> None self.__shutdown_requested.set() - - def __del__(self): - # type: (...) -> None - self.kill() diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index f9a5262903..e904081959 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -158,13 +158,6 @@ def is_healthy(self): # type: (Self) -> bool return True - def __del__(self): - # type: (Self) -> None - try: - self.kill() - except Exception: - pass - def _parse_rate_limits(header, now=None): # type: (str, Optional[datetime]) -> Iterable[Tuple[Optional[EventDataCategory], datetime]] From b065719ddd91aa27098f3875e48d2d3004349b0c Mon Sep 17 00:00:00 2001 From: Ben Beasley Date: Mon, 21 Jul 2025 05:08:23 -0400 Subject: [PATCH 361/591] Remove pyrsistent from test dependencies (#4588) The `pyrsistent` dependency in `requirements-testing.txt` does not appear to be used anywhere, so I removed it. In Fedora, I maintain the [`python-pyrsistent` package](https://src.fedoraproject.org/rpms/python-pyrsistent), and [`python-sentry-sdk`](https://src.fedoraproject.org/rpms/python-sentry-sdk) is the only thing that still depends on it. Since this dependency appears to be vestigial, my goal is to remove it both upstream and downstream and then orphan `python-pyrsistent` in Fedora as an unused leaf package. --- requirements-testing.txt | 1 - 1 file changed, 1 deletion(-) diff --git a/requirements-testing.txt b/requirements-testing.txt index 221863f4ab..5cd669af9a 100644 --- a/requirements-testing.txt +++ b/requirements-testing.txt @@ -6,7 +6,6 @@ pytest-forked pytest-localserver pytest-watch jsonschema -pyrsistent executing asttokens responses From d32e2eed6e2db041301ba43d212ca34342efd701 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 21 Jul 2025 13:00:27 +0200 Subject: [PATCH 362/591] fix: Fix `abs_path` bug in `serialize_frame` (#4599) Fixes #4587 --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- sentry_sdk/utils.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 595bbe0cf3..3b0ab8d746 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -591,9 +591,14 @@ def serialize_frame( if tb_lineno is None: tb_lineno = frame.f_lineno + try: + os_abs_path = os.path.abspath(abs_path) if abs_path else None + except Exception: + os_abs_path = None + rv = { "filename": filename_for_module(module, abs_path) or None, - "abs_path": os.path.abspath(abs_path) if abs_path else None, + "abs_path": os_abs_path, "function": function or "", "module": module, "lineno": tb_lineno, From 7b028b6c83c4d8ad9191864f25ef7d34ad95b111 Mon Sep 17 00:00:00 2001 From: Mikhail <6589665+mshavliuk@users.noreply.github.com> Date: Mon, 21 Jul 2025 14:00:50 +0300 Subject: [PATCH 363/591] fix(integrations): allow explicit op parameter in `ai_track` (#4597) fixes #4596 Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/ai/monitoring.py | 4 ++-- tests/test_ai_monitoring.py | 41 +++++++++++++++++++++++++++++++++++++ 2 files changed, 43 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index 461fd6af85..7a687736d0 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -32,7 +32,7 @@ def decorator(f): def sync_wrapped(*args, **kwargs): # type: (Any, Any) -> Any curr_pipeline = _ai_pipeline_name.get() - op = span_kwargs.get("op", "ai.run" if curr_pipeline else "ai.pipeline") + op = span_kwargs.pop("op", "ai.run" if curr_pipeline else "ai.pipeline") with start_span(name=description, op=op, **span_kwargs) as span: for k, v in kwargs.pop("sentry_tags", {}).items(): @@ -61,7 +61,7 @@ def sync_wrapped(*args, **kwargs): async def async_wrapped(*args, **kwargs): # type: (Any, Any) -> Any curr_pipeline = _ai_pipeline_name.get() - op = span_kwargs.get("op", "ai.run" if curr_pipeline else "ai.pipeline") + op = span_kwargs.pop("op", "ai.run" if curr_pipeline else "ai.pipeline") with start_span(name=description, op=op, **span_kwargs) as span: for k, v in kwargs.pop("sentry_tags", {}).items(): diff --git a/tests/test_ai_monitoring.py b/tests/test_ai_monitoring.py index 5e7c7432fa..ee757f82cd 100644 --- a/tests/test_ai_monitoring.py +++ b/tests/test_ai_monitoring.py @@ -119,3 +119,44 @@ async def async_pipeline(): assert ai_pipeline_span["tags"]["user"] == "czyber" assert ai_pipeline_span["data"]["some_data"] == "value" assert ai_run_span["description"] == "my async tool" + + +def test_ai_track_with_explicit_op(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + @ai_track("my tool", op="custom.operation") + def tool(**kwargs): + pass + + with sentry_sdk.start_transaction(): + tool() + + transaction = events[0] + assert transaction["type"] == "transaction" + assert len(transaction["spans"]) == 1 + span = transaction["spans"][0] + + assert span["description"] == "my tool" + assert span["op"] == "custom.operation" + + +@pytest.mark.asyncio +async def test_ai_track_async_with_explicit_op(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + @ai_track("my async tool", op="custom.async.operation") + async def async_tool(**kwargs): + pass + + with sentry_sdk.start_transaction(): + await async_tool() + + transaction = events[0] + assert transaction["type"] == "transaction" + assert len(transaction["spans"]) == 1 + span = transaction["spans"][0] + + assert span["description"] == "my async tool" + assert span["op"] == "custom.async.operation" From 38c27dd99a7759ec02937d92738a3fecf81cfb88 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 21 Jul 2025 12:41:29 +0000 Subject: [PATCH 364/591] release: 2.33.1 --- CHANGELOG.md | 13 +++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 16 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index b9e8bb6046..5b6b9a36cb 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Changelog +## 2.33.1 + +### Various fixes & improvements + +- fix(integrations): allow explicit op parameter in `ai_track` (#4597) by @mshavliuk +- fix: Fix `abs_path` bug in `serialize_frame` (#4599) by @szokeasaurusrex +- Remove pyrsistent from test dependencies (#4588) by @musicinmybrain +- Remove explicit __del__'s in threaded classes (#4590) by @sl0thentr0py +- ci: Check strictly for success (#4589) by @szokeasaurusrex +- test: Remove `test_installed_modules` (#4593) by @szokeasaurusrex +- Remove forked from test_transport, separate gevent tests and generalize capturing_server to be module level (#4577) by @sl0thentr0py +- Improve token usage recording (#4566) by @antonpirker + ## 2.33.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index cc5131636b..21045e31b4 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.33.0" +release = "2.33.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 9dc1de9bb7..30ea983e83 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1181,4 +1181,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.33.0" +VERSION = "2.33.1" diff --git a/setup.py b/setup.py index 213fcdb597..cd3b656e4a 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.33.0", + version="2.33.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 5cd43be59611576e5bdfad28a9df50bc9652ac4c Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Mon, 21 Jul 2025 14:42:45 +0200 Subject: [PATCH 365/591] meta: Update CHANGELOG.md --- CHANGELOG.md | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5b6b9a36cb..861168815a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -7,9 +7,7 @@ - fix(integrations): allow explicit op parameter in `ai_track` (#4597) by @mshavliuk - fix: Fix `abs_path` bug in `serialize_frame` (#4599) by @szokeasaurusrex - Remove pyrsistent from test dependencies (#4588) by @musicinmybrain -- Remove explicit __del__'s in threaded classes (#4590) by @sl0thentr0py -- ci: Check strictly for success (#4589) by @szokeasaurusrex -- test: Remove `test_installed_modules` (#4593) by @szokeasaurusrex +- Remove explicit `__del__`'s in threaded classes (#4590) by @sl0thentr0py - Remove forked from test_transport, separate gevent tests and generalize capturing_server to be module level (#4577) by @sl0thentr0py - Improve token usage recording (#4566) by @antonpirker From 0e7304989312affda1127ede95b5b4eda243ba8e Mon Sep 17 00:00:00 2001 From: MeredithAnya Date: Mon, 21 Jul 2025 23:21:52 -0700 Subject: [PATCH 366/591] ref(gnu-integration): update clickhouse stacktrace parsing (#4598) This integration was added back in 2019 https://github.com/getsentry/sentry-python/pull/288 but I think since then the stack trace from clickhouse driver has changed, anyway I tried to update it so that it parses the stack trace again --- sentry_sdk/integrations/gnu_backtrace.py | 17 +--- tests/integrations/test_gnu_backtrace.py | 99 +++++++----------------- 2 files changed, 29 insertions(+), 87 deletions(-) diff --git a/sentry_sdk/integrations/gnu_backtrace.py b/sentry_sdk/integrations/gnu_backtrace.py index dc3dc80fe0..21d8ea9b38 100644 --- a/sentry_sdk/integrations/gnu_backtrace.py +++ b/sentry_sdk/integrations/gnu_backtrace.py @@ -12,23 +12,12 @@ from sentry_sdk._types import Event -MODULE_RE = r"[a-zA-Z0-9/._:\\-]+" -TYPE_RE = r"[a-zA-Z0-9._:<>,-]+" -HEXVAL_RE = r"[A-Fa-f0-9]+" - +FUNCTION_RE = r"[^@]+?)\s+@\s+0x[0-9a-fA-F]+" FRAME_RE = r""" -^(?P\d+)\.\s -(?P{MODULE_RE})\( - (?P{TYPE_RE}\ )? - ((?P{TYPE_RE}) - (?P\(.*\))? - )? - ((?P\ const)?\+0x(?P{HEXVAL_RE}))? -\)\s -\[0x(?P{HEXVAL_RE})\]$ +^(?P\d+)\.\s+(?P{FUNCTION_RE}\s+in\s+(?P.+)$ """.format( - MODULE_RE=MODULE_RE, HEXVAL_RE=HEXVAL_RE, TYPE_RE=TYPE_RE + FUNCTION_RE=FUNCTION_RE, ) FRAME_RE = re.compile(FRAME_RE, re.MULTILINE | re.VERBOSE) diff --git a/tests/integrations/test_gnu_backtrace.py b/tests/integrations/test_gnu_backtrace.py index b91359dfa8..63930f850d 100644 --- a/tests/integrations/test_gnu_backtrace.py +++ b/tests/integrations/test_gnu_backtrace.py @@ -4,74 +4,31 @@ from sentry_sdk.integrations.gnu_backtrace import GnuBacktraceIntegration LINES = r""" -0. clickhouse-server(StackTrace::StackTrace()+0x16) [0x99d31a6] -1. clickhouse-server(DB::Exception::Exception(std::__cxx11::basic_string, std::allocator > const&, int)+0x22) [0x372c822] -10. clickhouse-server(DB::ActionsVisitor::visit(std::shared_ptr const&)+0x1a12) [0x6ae45d2] -10. clickhouse-server(DB::InterpreterSelectQuery::executeImpl(DB::InterpreterSelectQuery::Pipeline&, std::shared_ptr const&, bool)+0x11af) [0x75c68ff] -10. clickhouse-server(ThreadPoolImpl::worker(std::_List_iterator)+0x1ab) [0x6f90c1b] -11. clickhouse-server() [0xae06ddf] -11. clickhouse-server(DB::ExpressionAnalyzer::getRootActions(std::shared_ptr const&, bool, std::shared_ptr&, bool)+0xdb) [0x6a0a63b] -11. clickhouse-server(DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, DB::Context const&, std::shared_ptr const&, std::shared_ptr const&, std::vector, std::allocator >, std::allocator, std::allocator > > > const&, DB::QueryProcessingStage::Enum, unsigned long, bool)+0x5e6) [0x75c7516] -12. /lib/x86_64-linux-gnu/libpthread.so.0(+0x8184) [0x7f3bbc568184] -12. clickhouse-server(DB::ExpressionAnalyzer::getConstActions()+0xc9) [0x6a0b059] -12. clickhouse-server(DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, DB::Context const&, std::vector, std::allocator >, std::allocator, std::allocator > > > const&, DB::QueryProcessingStage::Enum, unsigned long, bool)+0x56) [0x75c8276] -13. /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7f3bbbb8303d] -13. clickhouse-server(DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::shared_ptr const&, DB::Context const&, std::vector, std::allocator >, std::allocator, std::allocator > > > const&, DB::QueryProcessingStage::Enum, unsigned long, bool)+0x7e7) [0x75d4067] -13. clickhouse-server(DB::evaluateConstantExpression(std::shared_ptr const&, DB::Context const&)+0x3ed) [0x656bfdd] -14. clickhouse-server(DB::InterpreterFactory::get(std::shared_ptr&, DB::Context&, DB::QueryProcessingStage::Enum)+0x3a8) [0x75b0298] -14. clickhouse-server(DB::makeExplicitSet(DB::ASTFunction const*, DB::Block const&, bool, DB::Context const&, DB::SizeLimits const&, std::unordered_map, DB::PreparedSetKey::Hash, std::equal_to, std::allocator > > >&)+0x382) [0x6adf692] -15. clickhouse-server() [0x7664c79] -15. clickhouse-server(DB::ActionsVisitor::makeSet(DB::ASTFunction const*, DB::Block const&)+0x2a7) [0x6ae2227] -16. clickhouse-server(DB::ActionsVisitor::visit(std::shared_ptr const&)+0x1973) [0x6ae4533] -16. clickhouse-server(DB::executeQuery(std::__cxx11::basic_string, std::allocator > const&, DB::Context&, bool, DB::QueryProcessingStage::Enum)+0x8a) [0x76669fa] -17. clickhouse-server(DB::ActionsVisitor::visit(std::shared_ptr const&)+0x1324) [0x6ae3ee4] -17. clickhouse-server(DB::TCPHandler::runImpl()+0x4b9) [0x30973c9] -18. clickhouse-server(DB::ExpressionAnalyzer::getRootActions(std::shared_ptr const&, bool, std::shared_ptr&, bool)+0xdb) [0x6a0a63b] -18. clickhouse-server(DB::TCPHandler::run()+0x2b) [0x30985ab] -19. clickhouse-server(DB::ExpressionAnalyzer::appendGroupBy(DB::ExpressionActionsChain&, bool)+0x100) [0x6a0b4f0] -19. clickhouse-server(Poco::Net::TCPServerConnection::start()+0xf) [0x9b53e4f] -2. clickhouse-server(DB::FunctionTuple::getReturnTypeImpl(std::vector, std::allocator > > const&) const+0x122) [0x3a2a0f2] -2. clickhouse-server(DB::readException(DB::Exception&, DB::ReadBuffer&, std::__cxx11::basic_string, std::allocator > const&)+0x21f) [0x6fb253f] -2. clickhouse-server(void DB::readDateTimeTextFallback(long&, DB::ReadBuffer&, DateLUTImpl const&)+0x318) [0x99ffed8] -20. clickhouse-server(DB::InterpreterSelectQuery::analyzeExpressions(DB::QueryProcessingStage::Enum, bool)+0x364) [0x6437fa4] -20. clickhouse-server(Poco::Net::TCPServerDispatcher::run()+0x16a) [0x9b5422a] -21. clickhouse-server(DB::InterpreterSelectQuery::executeImpl(DB::InterpreterSelectQuery::Pipeline&, std::shared_ptr const&, bool)+0x36d) [0x643c28d] -21. clickhouse-server(Poco::PooledThread::run()+0x77) [0x9c70f37] -22. clickhouse-server(DB::InterpreterSelectQuery::executeWithMultipleStreams()+0x50) [0x643ecd0] -22. clickhouse-server(Poco::ThreadImpl::runnableEntry(void*)+0x38) [0x9c6caa8] -23. clickhouse-server() [0xa3c68cf] -23. clickhouse-server(DB::InterpreterSelectWithUnionQuery::executeWithMultipleStreams()+0x6c) [0x644805c] -24. /lib/x86_64-linux-gnu/libpthread.so.0(+0x8184) [0x7fe839d2d184] -24. clickhouse-server(DB::InterpreterSelectWithUnionQuery::execute()+0x38) [0x6448658] -25. /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7fe83934803d] -25. clickhouse-server() [0x65744ef] -26. clickhouse-server(DB::executeQuery(std::__cxx11::basic_string, std::allocator > const&, DB::Context&, bool, DB::QueryProcessingStage::Enum, bool)+0x81) [0x6576141] -27. clickhouse-server(DB::TCPHandler::runImpl()+0x752) [0x3739f82] -28. clickhouse-server(DB::TCPHandler::run()+0x2b) [0x373a5cb] -29. clickhouse-server(Poco::Net::TCPServerConnection::start()+0xf) [0x708e63f] -3. clickhouse-server(DB::Connection::receiveException()+0x81) [0x67d3ad1] -3. clickhouse-server(DB::DefaultFunctionBuilder::getReturnTypeImpl(std::vector > const&) const+0x223) [0x38ac3b3] -3. clickhouse-server(DB::FunctionComparison::executeDateOrDateTimeOrEnumOrUUIDWithConstString(DB::Block&, unsigned long, DB::IColumn const*, DB::IColumn const*, std::shared_ptr const&, std::shared_ptr const&, bool, unsigned long)+0xbb3) [0x411dee3] -30. clickhouse-server(Poco::Net::TCPServerDispatcher::run()+0xe9) [0x708ed79] -31. clickhouse-server(Poco::PooledThread::run()+0x81) [0x7142011] -4. clickhouse-server(DB::Connection::receivePacket()+0x767) [0x67d9cd7] -4. clickhouse-server(DB::FunctionBuilderImpl::getReturnTypeWithoutLowCardinality(std::vector > const&) const+0x75) [0x6869635] -4. clickhouse-server(DB::FunctionComparison::executeImpl(DB::Block&, std::vector > const&, unsigned long, unsigned long)+0x576) [0x41ab006] -5. clickhouse-server(DB::FunctionBuilderImpl::getReturnType(std::vector > const&) const+0x350) [0x6869f10] -5. clickhouse-server(DB::MultiplexedConnections::receivePacket()+0x7e) [0x67e7ede] -5. clickhouse-server(DB::PreparedFunctionImpl::execute(DB::Block&, std::vector > const&, unsigned long, unsigned long)+0x3e2) [0x7933492] -6. clickhouse-server(DB::ExpressionAction::execute(DB::Block&, std::unordered_map, std::allocator >, unsigned long, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator > const, unsigned long> > >&) const+0x61a) [0x7ae093a] -6. clickhouse-server(DB::FunctionBuilderImpl::build(std::vector > const&) const+0x3c) [0x38accfc] -6. clickhouse-server(DB::RemoteBlockInputStream::readImpl()+0x87) [0x631da97] -7. clickhouse-server(DB::ExpressionActions::addImpl(DB::ExpressionAction, std::vector, std::allocator >, std::allocator, std::allocator > > >&)+0x552) [0x6a00052] -7. clickhouse-server(DB::ExpressionActions::execute(DB::Block&) const+0xe6) [0x7ae1e06] -7. clickhouse-server(DB::IBlockInputStream::read()+0x178) [0x63075e8] -8. clickhouse-server(DB::ExpressionActions::add(DB::ExpressionAction const&, std::vector, std::allocator >, std::allocator, std::allocator > > >&)+0x42) [0x6a00422] -8. clickhouse-server(DB::FilterBlockInputStream::FilterBlockInputStream(std::shared_ptr const&, std::shared_ptr const&, std::__cxx11::basic_string, std::allocator > const&, bool)+0x711) [0x79970d1] -8. clickhouse-server(DB::ParallelInputsProcessor::thread(std::shared_ptr, unsigned long)+0x2f1) [0x64467c1] -9. clickhouse-server() [0x75bd5a3] -9. clickhouse-server(DB::ScopeStack::addAction(DB::ExpressionAction const&)+0xd2) [0x6ae04d2] -9. clickhouse-server(ThreadFromGlobalPool::ThreadFromGlobalPool::process()::{lambda()#1}>(DB::ParallelInputsProcessor::process()::{lambda()#1}&&)::{lambda()#1}::operator()() const+0x6d) [0x644722d] +0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000bfc38a4 in /usr/bin/clickhouse +1. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String&&, String&&) @ 0x00000000075d242c in /usr/bin/clickhouse +2. DB::ActionsMatcher::visit(DB::ASTIdentifier const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1c648 in /usr/bin/clickhouse +3. DB::ActionsMatcher::visit(DB::ASTFunction const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1f58c in /usr/bin/clickhouse +4. DB::ActionsMatcher::visit(DB::ASTFunction const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1f58c in /usr/bin/clickhouse +5. DB::ActionsMatcher::visit(std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1c394 in /usr/bin/clickhouse +6. DB::InDepthNodeVisitor const>::doVisit(std::shared_ptr const&) @ 0x0000000010b154a0 in /usr/bin/clickhouse +7. DB::ExpressionAnalyzer::getRootActions(std::shared_ptr const&, bool, std::shared_ptr&, bool) @ 0x0000000010af83b4 in /usr/bin/clickhouse +8. DB::SelectQueryExpressionAnalyzer::appendSelect(DB::ExpressionActionsChain&, bool) @ 0x0000000010aff168 in /usr/bin/clickhouse +9. DB::ExpressionAnalysisResult::ExpressionAnalysisResult(DB::SelectQueryExpressionAnalyzer&, std::shared_ptr const&, bool, bool, bool, std::shared_ptr const&, std::shared_ptr const&, DB::Block const&) @ 0x0000000010b05b74 in /usr/bin/clickhouse +10. DB::InterpreterSelectQuery::getSampleBlockImpl() @ 0x00000000111559fc in /usr/bin/clickhouse +11. DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, std::shared_ptr const&, std::optional, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&, std::shared_ptr const&, std::shared_ptr)::$_0::operator()(bool) const @ 0x0000000011148254 in /usr/bin/clickhouse +12. DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, std::shared_ptr const&, std::optional, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&, std::shared_ptr const&, std::shared_ptr) @ 0x00000000111413e8 in /usr/bin/clickhouse +13. DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::shared_ptr const&, std::shared_ptr, DB::SelectQueryOptions const&, std::vector> const&) @ 0x00000000111d3708 in /usr/bin/clickhouse +14. DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x0000000011100b64 in /usr/bin/clickhouse +15. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, bool, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x00000000114c3f3c in /usr/bin/clickhouse +16. DB::executeQuery(String const&, std::shared_ptr, bool, DB::QueryProcessingStage::Enum) @ 0x00000000114c0ec8 in /usr/bin/clickhouse +17. DB::TCPHandler::runImpl() @ 0x00000000121bb5d8 in /usr/bin/clickhouse +18. DB::TCPHandler::run() @ 0x00000000121cb728 in /usr/bin/clickhouse +19. Poco::Net::TCPServerConnection::start() @ 0x00000000146d9404 in /usr/bin/clickhouse +20. Poco::Net::TCPServerDispatcher::run() @ 0x00000000146da900 in /usr/bin/clickhouse +21. Poco::PooledThread::run() @ 0x000000001484da7c in /usr/bin/clickhouse +22. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001484bc24 in /usr/bin/clickhouse +23. start_thread @ 0x0000000000007624 in /usr/lib/aarch64-linux-gnu/libpthread-2.31.so +24. ? @ 0x00000000000d162c in /usr/lib/aarch64-linux-gnu/libc-2.31.so """ @@ -94,8 +51,4 @@ def test_basic(sentry_init, capture_events, input): ) (frame,) = exception["stacktrace"]["frames"][1:] - if frame.get("function") is None: - assert "clickhouse-server()" in input or "pthread" in input - else: - assert ")" not in frame["function"] and "(" not in frame["function"] - assert frame["function"] in input + assert frame["function"] From 24790ebb2728e03a2044b0a877220b4823cc6418 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 22 Jul 2025 11:07:15 +0200 Subject: [PATCH 367/591] ref(spotlight): Do not import `sentry_sdk.spotlight` unless enabled (#4607) Closes https://github.com/getsentry/sentry-python/issues/4605 --- sentry_sdk/client.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 979ea92906..dca39beab8 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -47,7 +47,6 @@ ) from sentry_sdk.scrubber import EventScrubber from sentry_sdk.monitor import Monitor -from sentry_sdk.spotlight import setup_spotlight if TYPE_CHECKING: from typing import Any @@ -429,6 +428,10 @@ def _capture_envelope(envelope): ) if self.options.get("spotlight"): + # This is intentionally here to prevent setting up spotlight + # stuff we don't need unless spotlight is explicitly enabled + from sentry_sdk.spotlight import setup_spotlight + self.spotlight = setup_spotlight(self.options) if not self.options["dsn"]: sample_all = lambda *_args, **_kwargs: 1.0 From f4907a9bbf8586954b4c2d651126fb5534344942 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 22 Jul 2025 09:42:02 +0000 Subject: [PATCH 368/591] release: 2.33.2 --- CHANGELOG.md | 7 +++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 10 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 861168815a..a2ac6e09f8 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,12 @@ # Changelog +## 2.33.2 + +### Various fixes & improvements + +- ref(spotlight): Do not import `sentry_sdk.spotlight` unless enabled (#4607) by @sentrivana +- ref(gnu-integration): update clickhouse stacktrace parsing (#4598) by @MeredithAnya + ## 2.33.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 21045e31b4..faf861c518 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.33.1" +release = "2.33.2" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 30ea983e83..a7e713dc0b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1181,4 +1181,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.33.1" +VERSION = "2.33.2" diff --git a/setup.py b/setup.py index cd3b656e4a..9e75720390 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.33.1", + version="2.33.2", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 1b4f8d3b1ab070b7b279d72c70e8f28488efe89c Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Tue, 22 Jul 2025 16:09:34 +0200 Subject: [PATCH 369/591] Fix threading run patch (#4610) Since we're using `current_thread` for `self` explicitly, we need to remove the first argument from `*a`. This actually doesn't matter since `*a, **kw` are actually both empty but since this is the way the patch is implemented (presumably for forward compat), I don't want to change the signature. This fixes the following warning in CI since what actually happens is this: * `Thread.run` is being patched at each `Thread.start` (and not just once as other patches do) * So the current thread keeps getting accumulated in `*a` giving the `TypeError` for subsequent thread runs except the first ```python TypeError: Thread.run() takes 1 positional argument but 5 were given tests/integrations/redis/cluster_asyncio/test_redis_cluster_asyncio.py::test_async_span_origin /Users/neel/sentry/sdks/sentry-python/.tox/py3.12-redis-v5/lib/python3.12/site-packages/_pytest/threadexception.py:58: PytestUnhandledThreadExceptionWarning: Exception in thread sentry.monitor ``` closes #4361 --- sentry_sdk/integrations/threading.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/threading.py b/sentry_sdk/integrations/threading.py index 9c99a8e896..fc4f539228 100644 --- a/sentry_sdk/integrations/threading.py +++ b/sentry_sdk/integrations/threading.py @@ -120,7 +120,7 @@ def _run_old_run_func(): # type: () -> Any try: self = current_thread() - return old_run_func(self, *a, **kw) + return old_run_func(self, *a[1:], **kw) except Exception: reraise(*_capture_exception()) From e329179ac502405b645dba4e25399bc11f601150 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Wed, 23 Jul 2025 08:54:20 +0200 Subject: [PATCH 370/591] Ignore deliberate thread exception warnings (#4611) --- tests/integrations/threading/test_threading.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tests/integrations/threading/test_threading.py b/tests/integrations/threading/test_threading.py index 4395891d62..4577c846d8 100644 --- a/tests/integrations/threading/test_threading.py +++ b/tests/integrations/threading/test_threading.py @@ -13,6 +13,7 @@ original_run = Thread.run +@pytest.mark.filterwarnings("ignore:.*:pytest.PytestUnhandledThreadExceptionWarning") @pytest.mark.parametrize("integrations", [[ThreadingIntegration()], []]) def test_handles_exceptions(sentry_init, capture_events, integrations): sentry_init(default_integrations=False, integrations=integrations) @@ -36,6 +37,7 @@ def crash(): assert not events +@pytest.mark.filterwarnings("ignore:.*:pytest.PytestUnhandledThreadExceptionWarning") @pytest.mark.parametrize("propagate_hub", (True, False)) def test_propagates_hub(sentry_init, capture_events, propagate_hub): sentry_init( @@ -125,6 +127,7 @@ def run(self): assert unreachable_objects == 0 +@pytest.mark.filterwarnings("ignore:.*:pytest.PytestUnhandledThreadExceptionWarning") def test_double_patching(sentry_init, capture_events): sentry_init(default_integrations=False, integrations=[ThreadingIntegration()]) events = capture_events() From fe3ccb841c26bc01c4346ea0da1c355c3280e3ea Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 24 Jul 2025 11:26:43 +0200 Subject: [PATCH 371/591] tests: Update tox (#4609) Regular tox update Includes a small openai-agents tests fix and restricts openai-agents tests to 3.10 as in 0.2.x they're using 3.10+ typing syntax (` | `) --- scripts/populate_tox/config.py | 1 + .../openai_agents/test_openai_agents.py | 9 +++- tox.ini | 50 ++++++++++--------- 3 files changed, 35 insertions(+), 25 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 411d7fe666..f395289b4a 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -144,6 +144,7 @@ "deps": { "*": ["pytest-asyncio"], }, + "python": ">=3.10", }, "openfeature": { "package": "openfeature-sdk", diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index 37a066aeca..3f64e5c45c 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -19,6 +19,11 @@ ResponseFunctionToolCall, ) +from openai.types.responses.response_usage import ( + InputTokensDetails, + OutputTokensDetails, +) + test_run_config = agents.RunConfig(tracing_disabled=True) @@ -29,8 +34,8 @@ def mock_usage(): input_tokens=10, output_tokens=20, total_tokens=30, - input_tokens_details=MagicMock(cached_tokens=0), - output_tokens_details=MagicMock(reasoning_tokens=5), + input_tokens_details=InputTokensDetails(cached_tokens=0), + output_tokens_details=OutputTokensDetails(reasoning_tokens=5), ) diff --git a/tox.ini b/tox.ini index 8af16d640e..a0c7e5c927 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-07-15T08:21:43.713048+00:00 +# Last generated: 2025-07-23T07:24:30.467173+00:00 [tox] requires = @@ -138,15 +138,16 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.30.1 {py3.8,py3.11,py3.12}-anthropic-v0.44.0 - {py3.8,py3.11,py3.12}-anthropic-v0.57.1 + {py3.8,py3.12,py3.13}-anthropic-v0.58.2 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 {py3.9,py3.11,py3.12}-cohere-v5.13.12 {py3.9,py3.11,py3.12}-cohere-v5.16.1 - {py3.9,py3.11,py3.12}-openai_agents-v0.0.19 - {py3.9,py3.12,py3.13}-openai_agents-v0.1.0 + {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 + {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 + {py3.10,py3.12,py3.13}-openai_agents-v0.2.3 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 @@ -181,9 +182,9 @@ envlist = {py3.9,py3.12,py3.13}-openfeature-v0.8.1 {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.56.0 {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.59.0 + {py3.7,py3.12,py3.13}-statsig-v0.59.1 + {py3.7,py3.12,py3.13}-statsig-v0.60.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -205,9 +206,9 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.231.1 - {py3.8,py3.12,py3.13}-strawberry-v0.253.1 - {py3.9,py3.12,py3.13}-strawberry-v0.276.0 + {py3.8,py3.11,py3.12}-strawberry-v0.232.2 + {py3.8,py3.12,py3.13}-strawberry-v0.255.0 + {py3.9,py3.12,py3.13}-strawberry-v0.278.0 # ~~~ Network ~~~ @@ -254,7 +255,7 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.47.1 + {py3.9,py3.12,py3.13}-starlette-v0.47.2 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 @@ -275,6 +276,7 @@ envlist = {py3.6,py3.7}-falcon-v2.0.0 {py3.6,py3.11,py3.12}-falcon-v3.1.3 {py3.8,py3.11,py3.12}-falcon-v4.0.2 + {py3.8,py3.11,py3.12}-falcon-v4.1.0a3 {py3.8,py3.10,py3.11}-litestar-v2.0.1 {py3.8,py3.11,py3.12}-litestar-v2.5.5 @@ -303,8 +305,8 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.33 - {py3.9,py3.12,py3.13}-trytond-v7.6.3 + {py3.8,py3.11,py3.12}-trytond-v7.0.34 + {py3.9,py3.12,py3.13}-trytond-v7.6.4 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.0 @@ -510,7 +512,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.30.1: anthropic==0.30.1 anthropic-v0.44.0: anthropic==0.44.0 - anthropic-v0.57.1: anthropic==0.57.1 + anthropic-v0.58.2: anthropic==0.58.2 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.30.1: httpx<0.28.0 @@ -523,6 +525,7 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 + openai_agents-v0.2.3: openai-agents==0.2.3 openai_agents: pytest-asyncio huggingface_hub-v0.22.2: huggingface_hub==0.22.2 @@ -559,9 +562,9 @@ deps = openfeature-v0.8.1: openfeature-sdk==0.8.1 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.56.0: statsig==0.56.0 statsig-v0.57.3: statsig==0.57.3 - statsig-v0.59.0: statsig==0.59.0 + statsig-v0.59.1: statsig==0.59.1 + statsig-v0.60.0: statsig==0.60.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -592,13 +595,13 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.231.1: strawberry-graphql[fastapi,flask]==0.231.1 - strawberry-v0.253.1: strawberry-graphql[fastapi,flask]==0.253.1 - strawberry-v0.276.0: strawberry-graphql[fastapi,flask]==0.276.0 + strawberry-v0.232.2: strawberry-graphql[fastapi,flask]==0.232.2 + strawberry-v0.255.0: strawberry-graphql[fastapi,flask]==0.255.0 + strawberry-v0.278.0: strawberry-graphql[fastapi,flask]==0.278.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.231.1: pydantic<2.11 - strawberry-v0.253.1: pydantic<2.11 + strawberry-v0.232.2: pydantic<2.11 + strawberry-v0.255.0: pydantic<2.11 # ~~~ Network ~~~ @@ -678,7 +681,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.47.1: starlette==0.47.1 + starlette-v0.47.2: starlette==0.47.2 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -722,6 +725,7 @@ deps = falcon-v2.0.0: falcon==2.0.0 falcon-v3.1.3: falcon==3.1.3 falcon-v4.0.2: falcon==4.0.2 + falcon-v4.1.0a3: falcon==4.1.0a3 litestar-v2.0.1: litestar==2.0.1 litestar-v2.5.5: litestar==2.5.5 @@ -767,8 +771,8 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.33: trytond==7.0.33 - trytond-v7.6.3: trytond==7.6.3 + trytond-v7.0.34: trytond==7.0.34 + trytond-v7.6.4: trytond==7.6.4 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From add8b6fce794fd088347f14ccc2605e8ab650995 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 24 Jul 2025 12:40:54 +0200 Subject: [PATCH 372/591] Fix `huggingface_hub` CI tests. (#4619) HuggingFace has changed its default model used in `InferenceClient` and the new model breaks our CI. Change to use a non-existent model. --- .../huggingface_hub/test_huggingface_hub.py | 15 +++++++++------ 1 file changed, 9 insertions(+), 6 deletions(-) diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index 540fd675b9..df0c6c6d76 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -15,8 +15,10 @@ def mock_client_post(client, post_mock): # huggingface-hub==0.28.0 deprecates the `post` method # so patch `_inner_post` instead - client.post = post_mock - client._inner_post = post_mock + if hasattr(client, "post"): + client.post = post_mock + if hasattr(client, "_inner_post"): + client._inner_post = post_mock @pytest.mark.parametrize( @@ -33,7 +35,8 @@ def test_nonstreaming_chat_completion( ) events = capture_events() - client = InferenceClient() + client = InferenceClient(model="https://") + if details_arg: post_mock = mock.Mock( return_value=b"""[{ @@ -92,7 +95,7 @@ def test_streaming_chat_completion( ) events = capture_events() - client = InferenceClient() + client = InferenceClient(model="https://") post_mock = mock.Mock( return_value=[ @@ -141,7 +144,7 @@ def test_bad_chat_completion(sentry_init, capture_events): sentry_init(integrations=[HuggingfaceHubIntegration()], traces_sample_rate=1.0) events = capture_events() - client = InferenceClient() + client = InferenceClient(model="https://") post_mock = mock.Mock(side_effect=OverloadedError("The server is overloaded")) mock_client_post(client, post_mock) @@ -159,7 +162,7 @@ def test_span_origin(sentry_init, capture_events): ) events = capture_events() - client = InferenceClient() + client = InferenceClient(model="https://") post_mock = mock.Mock( return_value=[ b"""data:{ From c1c6e0b25f92a54755643d3b4bf9442901a1e800 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 24 Jul 2025 17:11:51 +0200 Subject: [PATCH 373/591] Remove remote example.com calls (#4622) --- tests/integrations/aiohttp/test_aiohttp.py | 15 +++++++++++---- tests/integrations/stdlib/test_httplib.py | 8 ++++---- 2 files changed, 15 insertions(+), 8 deletions(-) diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index 47152f254c..dbb4286370 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -6,7 +6,7 @@ import pytest -from aiohttp import web, ClientSession +from aiohttp import web from aiohttp.client import ServerDisconnectedError from aiohttp.web_request import Request from aiohttp.web_exceptions import ( @@ -636,6 +636,7 @@ async def handler(request): @pytest.mark.asyncio async def test_span_origin( sentry_init, + aiohttp_raw_server, aiohttp_client, capture_events, ): @@ -644,10 +645,16 @@ async def test_span_origin( traces_sample_rate=1.0, ) + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + async def hello(request): - async with ClientSession() as session: - async with session.get("http://example.com/"): - return web.Response(text="hello") + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") app = web.Application() app.router.add_get(r"/", hello) diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 908a22dc6c..f6735d0e74 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -123,7 +123,7 @@ def test_empty_realurl(https://github.com/sentry_init): """ sentry_init(dsn="") - HTTPConnection("example.com", port=443).putrequest("POST", None) + HTTPConnection("localhost", port=PORT).putrequest("POST", None) def test_httplib_misuse(sentry_init, capture_events, request): @@ -379,7 +379,7 @@ def test_span_origin(sentry_init, capture_events): events = capture_events() with start_transaction(name="foo"): - conn = HTTPConnection("example.com") + conn = HTTPConnection("localhost", port=PORT) conn.request("GET", "/foo") conn.getresponse() @@ -400,7 +400,7 @@ def test_http_timeout(monkeypatch, sentry_init, capture_envelopes): with pytest.raises(TimeoutError): with start_transaction(op="op", name="name"): - conn = HTTPSConnection("www.example.com") + conn = HTTPConnection("localhost", port=PORT) conn.request("GET", "/bla") conn.getresponse() @@ -410,4 +410,4 @@ def test_http_timeout(monkeypatch, sentry_init, capture_envelopes): span = transaction["spans"][0] assert span["op"] == "http.client" - assert span["description"] == "GET https://www.example.com/bla" + assert span["description"] == f"GET http://localhost:{PORT}/bla" # noqa: E231 From 69d65db6dcf79177f446413cc48909f422b77e46 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 25 Jul 2025 09:42:12 +0200 Subject: [PATCH 374/591] Treat django.template.context.BasicContext as sequence in serializer (#4621) closes #4606 --- sentry_sdk/integrations/django/__init__.py | 13 +++++++++++- sentry_sdk/serializer.py | 10 +++++++++- tests/integrations/django/test_basic.py | 23 ++++++++++++++++++++++ 3 files changed, 44 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/django/__init__.py b/sentry_sdk/integrations/django/__init__.py index ff67b3e39b..2041598fa0 100644 --- a/sentry_sdk/integrations/django/__init__.py +++ b/sentry_sdk/integrations/django/__init__.py @@ -7,7 +7,7 @@ import sentry_sdk from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.scope import add_global_event_processor, should_send_default_pii -from sentry_sdk.serializer import add_global_repr_processor +from sentry_sdk.serializer import add_global_repr_processor, add_repr_sequence_type from sentry_sdk.tracing import SOURCE_FOR_STYLE, TransactionSource from sentry_sdk.tracing_utils import add_query_source, record_sql_queries from sentry_sdk.utils import ( @@ -269,6 +269,7 @@ def _django_queryset_repr(value, hint): patch_views() patch_templates() patch_signals() + add_template_context_repr_sequence() if patch_caching is not None: patch_caching() @@ -745,3 +746,13 @@ def _set_db_data(span, cursor_or_db): server_socket_address = connection_params.get("unix_socket") if server_socket_address is not None: span.set_data(SPANDATA.SERVER_SOCKET_ADDRESS, server_socket_address) + + +def add_template_context_repr_sequence(): + # type: () -> None + try: + from django.template.context import BaseContext + + add_repr_sequence_type(BaseContext) + except Exception: + pass diff --git a/sentry_sdk/serializer.py b/sentry_sdk/serializer.py index bc8e38c631..04df9857bd 100644 --- a/sentry_sdk/serializer.py +++ b/sentry_sdk/serializer.py @@ -63,6 +63,14 @@ def add_global_repr_processor(processor): global_repr_processors.append(processor) +sequence_types = [Sequence, Set] # type: List[type] + + +def add_repr_sequence_type(ty): + # type: (type) -> None + sequence_types.append(ty) + + class Memo: __slots__ = ("_ids", "_objs") @@ -332,7 +340,7 @@ def _serialize_node_impl( return rv_dict elif not isinstance(obj, serializable_str_types) and isinstance( - obj, (Set, Sequence) + obj, tuple(sequence_types) ): rv_list = [] diff --git a/tests/integrations/django/test_basic.py b/tests/integrations/django/test_basic.py index 0e3f700105..e96cd09e4f 100644 --- a/tests/integrations/django/test_basic.py +++ b/tests/integrations/django/test_basic.py @@ -10,11 +10,13 @@ from werkzeug.test import Client from django import VERSION as DJANGO_VERSION + from django.contrib.auth.models import User from django.core.management import execute_from_command_line from django.db.utils import OperationalError, ProgrammingError, DataError from django.http.request import RawPostDataException from django.utils.functional import SimpleLazyObject +from django.template.context import make_context try: from django.urls import reverse @@ -310,6 +312,27 @@ def test_queryset_repr(sentry_init, capture_events): ) +@pytest.mark.forked +@pytest_mark_django_db_decorator() +def test_context_nested_queryset_repr(sentry_init, capture_events): + sentry_init(integrations=[DjangoIntegration()]) + events = capture_events() + User.objects.create_user("john", "lennon@thebeatles.com", "johnpassword") + + try: + context = make_context({"entries": User.objects.all()}) # noqa + 1 / 0 + except Exception: + capture_exception() + + (event,) = events + + (exception,) = event["exception"]["values"] + assert exception["type"] == "ZeroDivisionError" + (frame,) = exception["stacktrace"]["frames"] + assert " Date: Fri, 25 Jul 2025 15:06:23 +0200 Subject: [PATCH 375/591] Simplify celery double patching test (#4626) just check double patching, no need to recurse 10000 times part of #4623 --- tests/integrations/celery/test_celery.py | 29 ++++++++++++++++-------- 1 file changed, 19 insertions(+), 10 deletions(-) diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index 8c794bd5ff..ce2e693143 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -246,25 +246,34 @@ def dummy_task(x, y): ] -def test_no_stackoverflows(celery): - """We used to have a bug in the Celery integration where its monkeypatching +def test_no_double_patching(celery): + """Ensure that Celery tasks are only patched once to prevent stack overflows. + + We used to have a bug in the Celery integration where its monkeypatching was repeated for every task invocation, leading to stackoverflows. See https://github.com/getsentry/sentry-python/issues/265 """ - results = [] - @celery.task(name="dummy_task") def dummy_task(): - sentry_sdk.get_isolation_scope().set_tag("foo", "bar") - results.append(42) + return 42 - for _ in range(10000): - dummy_task.delay() + # Initially, the task should not be marked as patched + assert not hasattr(dummy_task, "_sentry_is_patched") + + # First invocation should trigger patching + result1 = dummy_task.delay() + assert result1.get() == 42 + assert getattr(dummy_task, "_sentry_is_patched", False) is True + + patched_run = dummy_task.run - assert results == [42] * 10000 - assert not sentry_sdk.get_isolation_scope()._tags + # Second invocation should not re-patch + result2 = dummy_task.delay() + assert result2.get() == 42 + assert dummy_task.run is patched_run + assert getattr(dummy_task, "_sentry_is_patched", False) is True def test_simple_no_propagation(capture_events, init_celery): From d71b9532c766fd90c9fc51d2f1f09b9f0dc6c792 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 25 Jul 2025 15:06:42 +0200 Subject: [PATCH 376/591] Fix socket tests to not use example.com (#4627) part of #4623 --- tests/integrations/socket/test_socket.py | 26 +++++++++++++----------- 1 file changed, 14 insertions(+), 12 deletions(-) diff --git a/tests/integrations/socket/test_socket.py b/tests/integrations/socket/test_socket.py index 389256de33..cc109e0968 100644 --- a/tests/integrations/socket/test_socket.py +++ b/tests/integrations/socket/test_socket.py @@ -2,7 +2,9 @@ from sentry_sdk import start_transaction from sentry_sdk.integrations.socket import SocketIntegration -from tests.conftest import ApproxDict +from tests.conftest import ApproxDict, create_mock_http_server + +PORT = create_mock_http_server() def test_getaddrinfo_trace(sentry_init, capture_events): @@ -10,17 +12,17 @@ def test_getaddrinfo_trace(sentry_init, capture_events): events = capture_events() with start_transaction(): - socket.getaddrinfo("example.com", 443) + socket.getaddrinfo("localhost", PORT) (event,) = events (span,) = event["spans"] assert span["op"] == "socket.dns" - assert span["description"] == "example.com:443" + assert span["description"] == f"localhost:{PORT}" # noqa: E231 assert span["data"] == ApproxDict( { - "host": "example.com", - "port": 443, + "host": "localhost", + "port": PORT, } ) @@ -32,28 +34,28 @@ def test_create_connection_trace(sentry_init, capture_events): events = capture_events() with start_transaction(): - socket.create_connection(("example.com", 443), timeout, None) + socket.create_connection(("localhost", PORT), timeout, None) (event,) = events (connect_span, dns_span) = event["spans"] # as getaddrinfo gets called in create_connection it should also contain a dns span assert connect_span["op"] == "socket.connection" - assert connect_span["description"] == "example.com:443" + assert connect_span["description"] == f"localhost:{PORT}" # noqa: E231 assert connect_span["data"] == ApproxDict( { - "address": ["example.com", 443], + "address": ["localhost", PORT], "timeout": timeout, "source_address": None, } ) assert dns_span["op"] == "socket.dns" - assert dns_span["description"] == "example.com:443" + assert dns_span["description"] == f"localhost:{PORT}" # noqa: E231 assert dns_span["data"] == ApproxDict( { - "host": "example.com", - "port": 443, + "host": "localhost", + "port": PORT, } ) @@ -66,7 +68,7 @@ def test_span_origin(sentry_init, capture_events): events = capture_events() with start_transaction(name="foo"): - socket.create_connection(("example.com", 443), 1, None) + socket.create_connection(("localhost", PORT), 1, None) (event,) = events From 070653a666f0b81438c82fb7945c208ffa63f39e Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Mon, 28 Jul 2025 17:19:57 +0200 Subject: [PATCH 377/591] Expose set_transaction_name (#4634) --- sentry_sdk/__init__.py | 1 + sentry_sdk/api.py | 7 +++++++ 2 files changed, 8 insertions(+) diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index e03f3b4484..7b1eda172a 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -49,6 +49,7 @@ "logger", "start_session", "end_session", + "set_transaction_name", ] # Initialize the debug support after everything is loaded diff --git a/sentry_sdk/api.py b/sentry_sdk/api.py index 698a2085ab..a4fb95e9a1 100644 --- a/sentry_sdk/api.py +++ b/sentry_sdk/api.py @@ -84,6 +84,7 @@ def overload(x): "monitor", "start_session", "end_session", + "set_transaction_name", ] @@ -466,3 +467,9 @@ def start_session( def end_session(): # type: () -> None return get_isolation_scope().end_session() + + +@scopemethod +def set_transaction_name(name, source=None): + # type: (str, Optional[str]) -> None + return get_current_scope().set_transaction_name(name, source) From 19ed1bb0ad232f53319aec02a3be66058af338eb Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Jul 2025 10:39:00 +0200 Subject: [PATCH 378/591] tests: tox.ini update (#4635) Regular tox update --- tox.ini | 40 ++++++++++++++++++++-------------------- 1 file changed, 20 insertions(+), 20 deletions(-) diff --git a/tox.ini b/tox.ini index a0c7e5c927..16067de8c7 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-07-23T07:24:30.467173+00:00 +# Last generated: 2025-07-29T06:07:22.069934+00:00 [tox] requires = @@ -136,9 +136,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.30.1 - {py3.8,py3.11,py3.12}-anthropic-v0.44.0 - {py3.8,py3.12,py3.13}-anthropic-v0.58.2 + {py3.8,py3.11,py3.12}-anthropic-v0.31.2 + {py3.8,py3.11,py3.12}-anthropic-v0.46.0 + {py3.8,py3.12,py3.13}-anthropic-v0.60.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 @@ -152,7 +152,8 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.33.4 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 # ~~~ DBs ~~~ @@ -184,7 +185,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.57.3 {py3.7,py3.12,py3.13}-statsig-v0.59.1 - {py3.7,py3.12,py3.13}-statsig-v0.60.0 + {py3.7,py3.12,py3.13}-statsig-v0.61.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -215,8 +216,7 @@ envlist = {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.46.5 {py3.7,py3.11,py3.12}-grpc-v1.60.2 - {py3.9,py3.12,py3.13}-grpc-v1.73.1 - {py3.9,py3.12,py3.13}-grpc-v1.74.0rc1 + {py3.9,py3.12,py3.13}-grpc-v1.74.0 # ~~~ Tasks ~~~ @@ -267,7 +267,7 @@ envlist = {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.12.14 + {py3.9,py3.12,py3.13}-aiohttp-v3.12.15 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -510,13 +510,13 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.30.1: anthropic==0.30.1 - anthropic-v0.44.0: anthropic==0.44.0 - anthropic-v0.58.2: anthropic==0.58.2 + anthropic-v0.31.2: anthropic==0.31.2 + anthropic-v0.46.0: anthropic==0.46.0 + anthropic-v0.60.0: anthropic==0.60.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.30.1: httpx<0.28.0 - anthropic-v0.44.0: httpx<0.28.0 + anthropic-v0.31.2: httpx<0.28.0 + anthropic-v0.46.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 @@ -531,7 +531,8 @@ deps = huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.33.4: huggingface_hub==0.33.4 + huggingface_hub-v0.34.2: huggingface_hub==0.34.2 + huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 # ~~~ DBs ~~~ @@ -564,7 +565,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.57.3: statsig==0.57.3 statsig-v0.59.1: statsig==0.59.1 - statsig-v0.60.0: statsig==0.60.0 + statsig-v0.61.0: statsig==0.61.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -608,8 +609,7 @@ deps = grpc-v1.32.0: grpcio==1.32.0 grpc-v1.46.5: grpcio==1.46.5 grpc-v1.60.2: grpcio==1.60.2 - grpc-v1.73.1: grpcio==1.73.1 - grpc-v1.74.0rc1: grpcio==1.74.0rc1 + grpc-v1.74.0: grpcio==1.74.0 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -712,10 +712,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.12.14: aiohttp==3.12.14 + aiohttp-v3.12.15: aiohttp==3.12.15 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.12.14: pytest-asyncio + aiohttp-v3.12.15: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.4: bottle==0.13.4 From fcace85734afd5fde920ab05101ae9025f7f5041 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 29 Jul 2025 13:13:37 +0200 Subject: [PATCH 379/591] OpenAI integration update (#4612) Update our OpenAI integration to support new APIs (`/responses`) and be OTel and Sentry AI Agents insights module compatible. Contains: - https://github.com/getsentry/sentry-python/pull/4563 - https://github.com/getsentry/sentry-python/pull/4564 - https://github.com/getsentry/sentry-python/pull/4628 --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/ai/monitoring.py | 6 +- sentry_sdk/consts.py | 26 +- sentry_sdk/integrations/openai.py | 492 +++++++++++----- .../integrations/openai_agents/utils.py | 50 +- sentry_sdk/utils.py | 46 ++ tests/integrations/openai/test_openai.py | 551 +++++++++++++++++- 6 files changed, 943 insertions(+), 228 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index 7a687736d0..e3f372c3ba 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -40,7 +40,7 @@ def sync_wrapped(*args, **kwargs): for k, v in kwargs.pop("sentry_data", {}).items(): span.set_data(k, v) if curr_pipeline: - span.set_data(SPANDATA.AI_PIPELINE_NAME, curr_pipeline) + span.set_data(SPANDATA.GEN_AI_PIPELINE_NAME, curr_pipeline) return f(*args, **kwargs) else: _ai_pipeline_name.set(description) @@ -69,7 +69,7 @@ async def async_wrapped(*args, **kwargs): for k, v in kwargs.pop("sentry_data", {}).items(): span.set_data(k, v) if curr_pipeline: - span.set_data(SPANDATA.AI_PIPELINE_NAME, curr_pipeline) + span.set_data(SPANDATA.GEN_AI_PIPELINE_NAME, curr_pipeline) return await f(*args, **kwargs) else: _ai_pipeline_name.set(description) @@ -108,7 +108,7 @@ def record_token_usage( # TODO: move pipeline name elsewhere ai_pipeline_name = get_ai_pipeline_name() if ai_pipeline_name: - span.set_data(SPANDATA.AI_PIPELINE_NAME, ai_pipeline_name) + span.set_data(SPANDATA.GEN_AI_PIPELINE_NAME, ai_pipeline_name) if input_tokens is not None: span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, input_tokens) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index a7e713dc0b..a82ff94c49 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -169,6 +169,7 @@ class SPANDATA: AI_PIPELINE_NAME = "ai.pipeline.name" """ Name of the AI pipeline or chain being executed. + DEPRECATED: Use GEN_AI_PIPELINE_NAME instead. Example: "qa-pipeline" """ @@ -229,6 +230,7 @@ class SPANDATA: AI_STREAMING = "ai.streaming" """ Whether or not the AI model call's response was streamed back asynchronously + DEPRECATED: Use GEN_AI_RESPONSE_STREAMING instead. Example: true """ @@ -372,6 +374,24 @@ class SPANDATA: Example: "chat" """ + GEN_AI_PIPELINE_NAME = "gen_ai.pipeline.name" + """ + Name of the AI pipeline or chain being executed. + Example: "qa-pipeline" + """ + + GEN_AI_RESPONSE_MODEL = "gen_ai.response.model" + """ + Exact model identifier used to generate the response + Example: gpt-4o-mini-2024-07-18 + """ + + GEN_AI_RESPONSE_STREAMING = "gen_ai.response.streaming" + """ + Whether or not the AI model call's response was streamed back asynchronously + Example: true + """ + GEN_AI_RESPONSE_TEXT = "gen_ai.response.text" """ The model's response text messages. @@ -411,7 +431,7 @@ class SPANDATA: GEN_AI_REQUEST_MODEL = "gen_ai.request.model" """ The model identifier being used for the request. - Example: "gpt-4-turbo-preview" + Example: "gpt-4-turbo" """ GEN_AI_REQUEST_PRESENCE_PENALTY = "gen_ai.request.presence_penalty" @@ -649,9 +669,11 @@ class OP: FUNCTION_AWS = "function.aws" FUNCTION_GCP = "function.gcp" GEN_AI_CHAT = "gen_ai.chat" + GEN_AI_EMBEDDINGS = "gen_ai.embeddings" GEN_AI_EXECUTE_TOOL = "gen_ai.execute_tool" GEN_AI_HANDOFF = "gen_ai.handoff" GEN_AI_INVOKE_AGENT = "gen_ai.invoke_agent" + GEN_AI_RESPONSES = "gen_ai.responses" GRAPHQL_EXECUTE = "graphql.execute" GRAPHQL_MUTATION = "graphql.mutation" GRAPHQL_PARSE = "graphql.parse" @@ -674,8 +696,6 @@ class OP: MIDDLEWARE_STARLITE = "middleware.starlite" MIDDLEWARE_STARLITE_RECEIVE = "middleware.starlite.receive" MIDDLEWARE_STARLITE_SEND = "middleware.starlite.send" - OPENAI_CHAT_COMPLETIONS_CREATE = "ai.chat_completions.create.openai" - OPENAI_EMBEDDINGS_CREATE = "ai.embeddings.create.openai" HUGGINGFACE_HUB_CHAT_COMPLETIONS_CREATE = ( "ai.chat_completions.create.huggingface_hub" ) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index d906a8e0b2..78fcdd49e2 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -10,6 +10,7 @@ from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, + safe_serialize, ) from typing import TYPE_CHECKING @@ -27,6 +28,14 @@ except ImportError: raise DidNotEnable("OpenAI not installed") +RESPONSES_API_ENABLED = True +try: + # responses API support was introduced in v1.66.0 + from openai.resources.responses import Responses, AsyncResponses + from openai.types.responses.response_completed_event import ResponseCompletedEvent +except ImportError: + RESPONSES_API_ENABLED = False + class OpenAIIntegration(Integration): identifier = "openai" @@ -46,13 +55,17 @@ def __init__(self, include_prompts=True, tiktoken_encoding_name=None): def setup_once(): # type: () -> None Completions.create = _wrap_chat_completion_create(Completions.create) - Embeddings.create = _wrap_embeddings_create(Embeddings.create) - AsyncCompletions.create = _wrap_async_chat_completion_create( AsyncCompletions.create ) + + Embeddings.create = _wrap_embeddings_create(Embeddings.create) AsyncEmbeddings.create = _wrap_async_embeddings_create(AsyncEmbeddings.create) + if RESPONSES_API_ENABLED: + Responses.create = _wrap_responses_create(Responses.create) + AsyncResponses.create = _wrap_async_responses_create(AsyncResponses.create) + def count_tokens(self, s): # type: (OpenAIIntegration, str) -> int if self.tiktoken_encoding is not None: @@ -62,6 +75,12 @@ def count_tokens(self, s): def _capture_exception(exc): # type: (Any) -> None + # Close an eventually open span + # We need to do this by hand because we are not using the start_span context manager + current_span = sentry_sdk.get_current_span() + if current_span is not None: + current_span.__exit__(None, None, None) + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, @@ -81,7 +100,7 @@ def _get_usage(usage, names): def _calculate_token_usage( messages, response, span, streaming_message_responses, count_tokens ): - # type: (Iterable[ChatCompletionMessageParam], Any, Span, Optional[List[str]], Callable[..., Any]) -> None + # type: (Optional[Iterable[ChatCompletionMessageParam]], Any, Span, Optional[List[str]], Callable[..., Any]) -> None input_tokens = 0 # type: Optional[int] input_tokens_cached = 0 # type: Optional[int] output_tokens = 0 # type: Optional[int] @@ -106,13 +125,13 @@ def _calculate_token_usage( total_tokens = _get_usage(response.usage, ["total_tokens"]) # Manually count tokens - # TODO: when implementing responses API, check for responses API if input_tokens == 0: - for message in messages: - if "content" in message: + for message in messages or []: + if isinstance(message, dict) and "content" in message: input_tokens += count_tokens(message["content"]) + elif isinstance(message, str): + input_tokens += count_tokens(message) - # TODO: when implementing responses API, check for responses API if output_tokens == 0: if streaming_message_responses is not None: for message in streaming_message_responses: @@ -139,138 +158,254 @@ def _calculate_token_usage( ) -def _new_chat_completion_common(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any - integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) - if integration is None: - return f(*args, **kwargs) +def _set_input_data(span, kwargs, operation, integration): + # type: (Span, dict[str, Any], str, OpenAIIntegration) -> None + # Input messages (the prompt or data sent to the model) + messages = kwargs.get("messages") + if messages is None: + messages = kwargs.get("input") + + if isinstance(messages, str): + messages = [messages] + + if ( + messages is not None + and len(messages) > 0 + and should_send_default_pii() + and integration.include_prompts + ): + set_data_normalized(span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages) + + # Input attributes: Common + set_data_normalized(span, SPANDATA.GEN_AI_SYSTEM, "openai") + set_data_normalized(span, SPANDATA.GEN_AI_OPERATION_NAME, operation) + + # Input attributes: Optional + kwargs_keys_to_attributes = { + "model": SPANDATA.GEN_AI_REQUEST_MODEL, + "stream": SPANDATA.GEN_AI_RESPONSE_STREAMING, + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, + } + for key, attribute in kwargs_keys_to_attributes.items(): + value = kwargs.get(key) + if value is not None: + set_data_normalized(span, attribute, value) + + # Input attributes: Tools + tools = kwargs.get("tools") + if tools is not None and len(tools) > 0: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) + ) - if "messages" not in kwargs: - # invalid call (in all versions of openai), let it return error - return f(*args, **kwargs) - try: - iter(kwargs["messages"]) - except TypeError: - # invalid call (in all versions), messages must be iterable - return f(*args, **kwargs) +def _set_output_data(span, response, kwargs, integration, finish_span=True): + # type: (Span, Any, dict[str, Any], OpenAIIntegration, bool) -> None + if hasattr(response, "model"): + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_MODEL, response.model) - kwargs["messages"] = list(kwargs["messages"]) - messages = kwargs["messages"] - model = kwargs.get("model") - streaming = kwargs.get("stream") - - span = sentry_sdk.start_span( - op=consts.OP.OPENAI_CHAT_COMPLETIONS_CREATE, - name="Chat Completion", - origin=OpenAIIntegration.origin, - ) - span.__enter__() + # Input messages (the prompt or data sent to the model) + # used for the token usage calculation + messages = kwargs.get("messages") + if messages is None: + messages = kwargs.get("input") - res = yield f, args, kwargs + if messages is not None and isinstance(messages, str): + messages = [messages] - with capture_internal_exceptions(): + if hasattr(response, "choices"): if should_send_default_pii() and integration.include_prompts: - set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, messages) - - set_data_normalized(span, SPANDATA.AI_MODEL_ID, model) - set_data_normalized(span, SPANDATA.AI_STREAMING, streaming) + response_text = [choice.message.dict() for choice in response.choices] + if len(response_text) > 0: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + safe_serialize(response_text), + ) + _calculate_token_usage(messages, response, span, None, integration.count_tokens) + if finish_span: + span.__exit__(None, None, None) - if hasattr(res, "choices"): - if should_send_default_pii() and integration.include_prompts: + elif hasattr(response, "output"): + if should_send_default_pii() and integration.include_prompts: + response_text = [item.to_dict() for item in response.output] + if len(response_text) > 0: set_data_normalized( span, - SPANDATA.AI_RESPONSES, - list(map(lambda x: x.message, res.choices)), + SPANDATA.GEN_AI_RESPONSE_TEXT, + safe_serialize(response_text), ) - _calculate_token_usage(messages, res, span, None, integration.count_tokens) + _calculate_token_usage(messages, response, span, None, integration.count_tokens) + if finish_span: span.__exit__(None, None, None) - elif hasattr(res, "_iterator"): - data_buf: list[list[str]] = [] # one for each choice - - old_iterator = res._iterator - - def new_iterator(): - # type: () -> Iterator[ChatCompletionChunk] - with capture_internal_exceptions(): - for x in old_iterator: - if hasattr(x, "choices"): - choice_index = 0 - for choice in x.choices: - if hasattr(choice, "delta") and hasattr( - choice.delta, "content" - ): - content = choice.delta.content - if len(data_buf) <= choice_index: - data_buf.append([]) - data_buf[choice_index].append(content or "") - choice_index += 1 - yield x - if len(data_buf) > 0: - all_responses = list( - map(lambda chunk: "".join(chunk), data_buf) + + elif hasattr(response, "_iterator"): + data_buf: list[list[str]] = [] # one for each choice + + old_iterator = response._iterator + + def new_iterator(): + # type: () -> Iterator[ChatCompletionChunk] + with capture_internal_exceptions(): + count_tokens_manually = True + for x in old_iterator: + # OpenAI chat completion API + if hasattr(x, "choices"): + choice_index = 0 + for choice in x.choices: + if hasattr(choice, "delta") and hasattr( + choice.delta, "content" + ): + content = choice.delta.content + if len(data_buf) <= choice_index: + data_buf.append([]) + data_buf[choice_index].append(content or "") + choice_index += 1 + + # OpenAI responses API + elif hasattr(x, "delta"): + if len(data_buf) == 0: + data_buf.append([]) + data_buf[0].append(x.delta or "") + + # OpenAI responses API end of streaming response + if RESPONSES_API_ENABLED and isinstance(x, ResponseCompletedEvent): + _calculate_token_usage( + messages, + x.response, + span, + None, + integration.count_tokens, ) - if should_send_default_pii() and integration.include_prompts: - set_data_normalized( - span, SPANDATA.AI_RESPONSES, all_responses - ) + count_tokens_manually = False + + yield x + + if len(data_buf) > 0: + all_responses = ["".join(chunk) for chunk in data_buf] + if should_send_default_pii() and integration.include_prompts: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, all_responses + ) + if count_tokens_manually: _calculate_token_usage( messages, - res, + response, span, all_responses, integration.count_tokens, ) + + if finish_span: span.__exit__(None, None, None) - async def new_iterator_async(): - # type: () -> AsyncIterator[ChatCompletionChunk] - with capture_internal_exceptions(): - async for x in old_iterator: - if hasattr(x, "choices"): - choice_index = 0 - for choice in x.choices: - if hasattr(choice, "delta") and hasattr( - choice.delta, "content" - ): - content = choice.delta.content - if len(data_buf) <= choice_index: - data_buf.append([]) - data_buf[choice_index].append(content or "") - choice_index += 1 - yield x - if len(data_buf) > 0: - all_responses = list( - map(lambda chunk: "".join(chunk), data_buf) + async def new_iterator_async(): + # type: () -> AsyncIterator[ChatCompletionChunk] + with capture_internal_exceptions(): + count_tokens_manually = True + async for x in old_iterator: + # OpenAI chat completion API + if hasattr(x, "choices"): + choice_index = 0 + for choice in x.choices: + if hasattr(choice, "delta") and hasattr( + choice.delta, "content" + ): + content = choice.delta.content + if len(data_buf) <= choice_index: + data_buf.append([]) + data_buf[choice_index].append(content or "") + choice_index += 1 + + # OpenAI responses API + elif hasattr(x, "delta"): + if len(data_buf) == 0: + data_buf.append([]) + data_buf[0].append(x.delta or "") + + # OpenAI responses API end of streaming response + if RESPONSES_API_ENABLED and isinstance(x, ResponseCompletedEvent): + _calculate_token_usage( + messages, + x.response, + span, + None, + integration.count_tokens, + ) + count_tokens_manually = False + + yield x + + if len(data_buf) > 0: + all_responses = ["".join(chunk) for chunk in data_buf] + if should_send_default_pii() and integration.include_prompts: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, all_responses ) - if should_send_default_pii() and integration.include_prompts: - set_data_normalized( - span, SPANDATA.AI_RESPONSES, all_responses - ) + if count_tokens_manually: _calculate_token_usage( messages, - res, + response, span, all_responses, integration.count_tokens, ) + if finish_span: span.__exit__(None, None, None) - if str(type(res._iterator)) == "": - res._iterator = new_iterator_async() - else: - res._iterator = new_iterator() - + if str(type(response._iterator)) == "": + response._iterator = new_iterator_async() else: - set_data_normalized(span, "unknown_response", True) + response._iterator = new_iterator() + else: + _calculate_token_usage(messages, response, span, None, integration.count_tokens) + if finish_span: span.__exit__(None, None, None) - return res + + +def _new_chat_completion_common(f, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) + if integration is None: + return f(*args, **kwargs) + + if "messages" not in kwargs: + # invalid call (in all versions of openai), let it return error + return f(*args, **kwargs) + + try: + iter(kwargs["messages"]) + except TypeError: + # invalid call (in all versions), messages must be iterable + return f(*args, **kwargs) + + model = kwargs.get("model") + operation = "chat" + + span = sentry_sdk.start_span( + op=consts.OP.GEN_AI_CHAT, + name=f"{operation} {model}", + origin=OpenAIIntegration.origin, + ) + span.__enter__() + + _set_input_data(span, kwargs, operation, integration) + + response = yield f, args, kwargs + + _set_output_data(span, response, kwargs, integration, finish_span=True) + + return response def _wrap_chat_completion_create(f): # type: (Callable[..., Any]) -> Callable[..., Any] def _execute_sync(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any + # type: (Any, Any, Any) -> Any gen = _new_chat_completion_common(f, *args, **kwargs) try: @@ -291,7 +426,7 @@ def _execute_sync(f, *args, **kwargs): @wraps(f) def _sentry_patched_create_sync(*args, **kwargs): - # type: (*Any, **Any) -> Any + # type: (Any, Any) -> Any integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) if integration is None or "messages" not in kwargs: # no "messages" means invalid call (in all versions of openai), let it return error @@ -305,7 +440,7 @@ def _sentry_patched_create_sync(*args, **kwargs): def _wrap_async_chat_completion_create(f): # type: (Callable[..., Any]) -> Callable[..., Any] async def _execute_async(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any + # type: (Any, Any, Any) -> Any gen = _new_chat_completion_common(f, *args, **kwargs) try: @@ -326,7 +461,7 @@ async def _execute_async(f, *args, **kwargs): @wraps(f) async def _sentry_patched_create_async(*args, **kwargs): - # type: (*Any, **Any) -> Any + # type: (Any, Any) -> Any integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) if integration is None or "messages" not in kwargs: # no "messages" means invalid call (in all versions of openai), let it return error @@ -338,52 +473,24 @@ async def _sentry_patched_create_async(*args, **kwargs): def _new_embeddings_create_common(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any + # type: (Any, Any, Any) -> Any integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) if integration is None: return f(*args, **kwargs) + model = kwargs.get("model") + operation = "embeddings" + with sentry_sdk.start_span( - op=consts.OP.OPENAI_EMBEDDINGS_CREATE, - description="OpenAI Embedding Creation", + op=consts.OP.GEN_AI_EMBEDDINGS, + name=f"{operation} {model}", origin=OpenAIIntegration.origin, ) as span: - if "input" in kwargs and ( - should_send_default_pii() and integration.include_prompts - ): - if isinstance(kwargs["input"], str): - set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, [kwargs["input"]]) - elif ( - isinstance(kwargs["input"], list) - and len(kwargs["input"]) > 0 - and isinstance(kwargs["input"][0], str) - ): - set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, kwargs["input"]) - if "model" in kwargs: - set_data_normalized(span, SPANDATA.AI_MODEL_ID, kwargs["model"]) + _set_input_data(span, kwargs, operation, integration) response = yield f, args, kwargs - input_tokens = 0 - total_tokens = 0 - if hasattr(response, "usage"): - if hasattr(response.usage, "prompt_tokens") and isinstance( - response.usage.prompt_tokens, int - ): - input_tokens = response.usage.prompt_tokens - if hasattr(response.usage, "total_tokens") and isinstance( - response.usage.total_tokens, int - ): - total_tokens = response.usage.total_tokens - - if input_tokens == 0: - input_tokens = integration.count_tokens(kwargs["input"] or "") - - record_token_usage( - span, - input_tokens=input_tokens, - total_tokens=total_tokens or input_tokens, - ) + _set_output_data(span, response, kwargs, integration, finish_span=False) return response @@ -391,7 +498,7 @@ def _new_embeddings_create_common(f, *args, **kwargs): def _wrap_embeddings_create(f): # type: (Any) -> Any def _execute_sync(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any + # type: (Any, Any, Any) -> Any gen = _new_embeddings_create_common(f, *args, **kwargs) try: @@ -412,7 +519,7 @@ def _execute_sync(f, *args, **kwargs): @wraps(f) def _sentry_patched_create_sync(*args, **kwargs): - # type: (*Any, **Any) -> Any + # type: (Any, Any) -> Any integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) if integration is None: return f(*args, **kwargs) @@ -425,7 +532,7 @@ def _sentry_patched_create_sync(*args, **kwargs): def _wrap_async_embeddings_create(f): # type: (Any) -> Any async def _execute_async(f, *args, **kwargs): - # type: (Any, *Any, **Any) -> Any + # type: (Any, Any, Any) -> Any gen = _new_embeddings_create_common(f, *args, **kwargs) try: @@ -446,7 +553,7 @@ async def _execute_async(f, *args, **kwargs): @wraps(f) async def _sentry_patched_create_async(*args, **kwargs): - # type: (*Any, **Any) -> Any + # type: (Any, Any) -> Any integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) if integration is None: return await f(*args, **kwargs) @@ -454,3 +561,96 @@ async def _sentry_patched_create_async(*args, **kwargs): return await _execute_async(f, *args, **kwargs) return _sentry_patched_create_async + + +def _new_responses_create_common(f, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) + if integration is None: + return f(*args, **kwargs) + + model = kwargs.get("model") + operation = "responses" + + span = sentry_sdk.start_span( + op=consts.OP.GEN_AI_RESPONSES, + name=f"{operation} {model}", + origin=OpenAIIntegration.origin, + ) + span.__enter__() + + _set_input_data(span, kwargs, operation, integration) + + response = yield f, args, kwargs + + _set_output_data(span, response, kwargs, integration, finish_span=True) + + return response + + +def _wrap_responses_create(f): + # type: (Any) -> Any + def _execute_sync(f, *args, **kwargs): + # type: (Any, Any, Any) -> Any + gen = _new_responses_create_common(f, *args, **kwargs) + + try: + f, args, kwargs = next(gen) + except StopIteration as e: + return e.value + + try: + try: + result = f(*args, **kwargs) + except Exception as e: + _capture_exception(e) + raise e from None + + return gen.send(result) + except StopIteration as e: + return e.value + + @wraps(f) + def _sentry_patched_create_sync(*args, **kwargs): + # type: (Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) + if integration is None: + return f(*args, **kwargs) + + return _execute_sync(f, *args, **kwargs) + + return _sentry_patched_create_sync + + +def _wrap_async_responses_create(f): + # type: (Any) -> Any + async def _execute_async(f, *args, **kwargs): + # type: (Any, Any, Any) -> Any + gen = _new_responses_create_common(f, *args, **kwargs) + + try: + f, args, kwargs = next(gen) + except StopIteration as e: + return await e.value + + try: + try: + result = await f(*args, **kwargs) + except Exception as e: + _capture_exception(e) + raise e from None + + return gen.send(result) + except StopIteration as e: + return e.value + + @wraps(f) + async def _sentry_patched_responses_async(*args, **kwargs): + # type: (Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(OpenAIIntegration) + if integration is None: + return await f(*args, **kwargs) + + return await _execute_async(f, *args, **kwargs) + + return _sentry_patched_responses_async diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index dc66521c83..1525346726 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -1,16 +1,14 @@ -import json import sentry_sdk from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.utils import event_from_exception +from sentry_sdk.utils import event_from_exception, safe_serialize from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Any from typing import Callable - from typing import Union from agents import Usage try: @@ -162,49 +160,3 @@ def _set_output_data(span, result): span.set_data( SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(output_messages["response"]) ) - - -def safe_serialize(data): - # type: (Any) -> str - """Safely serialize to a readable string.""" - - def serialize_item(item): - # type: (Any) -> Union[str, dict[Any, Any], list[Any], tuple[Any, ...]] - if callable(item): - try: - module = getattr(item, "__module__", None) - qualname = getattr(item, "__qualname__", None) - name = getattr(item, "__name__", "anonymous") - - if module and qualname: - full_path = f"{module}.{qualname}" - elif module and name: - full_path = f"{module}.{name}" - else: - full_path = name - - return f"" - except Exception: - return f"" - elif isinstance(item, dict): - return {k: serialize_item(v) for k, v in item.items()} - elif isinstance(item, (list, tuple)): - return [serialize_item(x) for x in item] - elif hasattr(item, "__dict__"): - try: - attrs = { - k: serialize_item(v) - for k, v in vars(item).items() - if not k.startswith("_") - } - return f"<{type(item).__name__} {attrs}>" - except Exception: - return repr(item) - else: - return item - - try: - serialized = serialize_item(data) - return json.dumps(serialized, default=str) - except Exception: - return str(data) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 3b0ab8d746..9c6f2cfc3b 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1938,3 +1938,49 @@ def try_convert(convert_func, value): return convert_func(value) except Exception: return None + + +def safe_serialize(data): + # type: (Any) -> str + """Safely serialize to a readable string.""" + + def serialize_item(item): + # type: (Any) -> Union[str, dict[Any, Any], list[Any], tuple[Any, ...]] + if callable(item): + try: + module = getattr(item, "__module__", None) + qualname = getattr(item, "__qualname__", None) + name = getattr(item, "__name__", "anonymous") + + if module and qualname: + full_path = f"{module}.{qualname}" + elif module and name: + full_path = f"{module}.{name}" + else: + full_path = name + + return f"" + except Exception: + return f"" + elif isinstance(item, dict): + return {k: serialize_item(v) for k, v in item.items()} + elif isinstance(item, (list, tuple)): + return [serialize_item(x) for x in item] + elif hasattr(item, "__dict__"): + try: + attrs = { + k: serialize_item(v) + for k, v in vars(item).items() + if not k.startswith("_") + } + return f"<{type(item).__name__} {attrs}>" + except Exception: + return repr(item) + else: + return item + + try: + serialized = serialize_item(data) + return json.dumps(serialized, default=str) + except Exception: + return str(data) diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index ac6d9f4c29..dfac08d762 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1,3 +1,4 @@ +import json import pytest from openai import AsyncOpenAI, OpenAI, AsyncStream, Stream, OpenAIError from openai.types import CompletionUsage, CreateEmbeddingResponse, Embedding @@ -6,6 +7,25 @@ from openai.types.chat.chat_completion_chunk import ChoiceDelta, Choice as DeltaChoice from openai.types.create_embedding_response import Usage as EmbeddingTokenUsage +SKIP_RESPONSES_TESTS = False + +try: + from openai.types.responses.response_completed_event import ResponseCompletedEvent + from openai.types.responses.response_created_event import ResponseCreatedEvent + from openai.types.responses.response_text_delta_event import ResponseTextDeltaEvent + from openai.types.responses.response_usage import ( + InputTokensDetails, + OutputTokensDetails, + ) + from openai.types.responses import ( + Response, + ResponseUsage, + ResponseOutputMessage, + ResponseOutputText, + ) +except ImportError: + SKIP_RESPONSES_TESTS = True + from sentry_sdk import start_transaction from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations.openai import ( @@ -36,7 +56,7 @@ async def __call__(self, *args, **kwargs): ) ], created=10000000, - model="model-id", + model="response-model-id", object="chat.completion", usage=CompletionUsage( completion_tokens=10, @@ -46,6 +66,46 @@ async def __call__(self, *args, **kwargs): ) +if SKIP_RESPONSES_TESTS: + EXAMPLE_RESPONSE = None +else: + EXAMPLE_RESPONSE = Response( + id="chat-id", + output=[ + ResponseOutputMessage( + id="message-id", + content=[ + ResponseOutputText( + annotations=[], + text="the model response", + type="output_text", + ), + ], + role="assistant", + status="completed", + type="message", + ), + ], + parallel_tool_calls=False, + tool_choice="none", + tools=[], + created_at=10000000, + model="response-model-id", + object="response", + usage=ResponseUsage( + input_tokens=20, + input_tokens_details=InputTokensDetails( + cached_tokens=5, + ), + output_tokens=10, + output_tokens_details=OutputTokensDetails( + reasoning_tokens=8, + ), + total_tokens=30, + ), + ) + + async def async_iterator(values): for value in values: yield value @@ -81,14 +141,17 @@ def test_nonstreaming_chat_completion( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.openai" + assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] - assert "the model response" in span["data"][SPANDATA.AI_RESPONSES]["content"] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert ( + "the model response" + in json.loads(span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT])[0]["content"] + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] assert span["data"]["gen_ai.usage.output_tokens"] == 10 assert span["data"]["gen_ai.usage.input_tokens"] == 20 @@ -123,14 +186,17 @@ async def test_nonstreaming_chat_completion_async( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.openai" + assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] - assert "the model response" in span["data"][SPANDATA.AI_RESPONSES]["content"] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert ( + "the model response" + in json.loads(span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT])[0]["content"] + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] assert span["data"]["gen_ai.usage.output_tokens"] == 10 assert span["data"]["gen_ai.usage.input_tokens"] == 20 @@ -216,14 +282,14 @@ def test_streaming_chat_completion( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.openai" + assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] - assert "hello world" in span["data"][SPANDATA.AI_RESPONSES] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert "hello world" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import @@ -312,14 +378,14 @@ async def test_streaming_chat_completion_async( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.openai" + assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES]["content"] - assert "hello world" in span["data"][SPANDATA.AI_RESPONSES] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert "hello world" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] try: import tiktoken # type: ignore # noqa # pylint: disable=unused-import @@ -403,11 +469,11 @@ def test_embeddings_create( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.embeddings.create.openai" + assert span["op"] == "gen_ai.embeddings" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert span["data"]["gen_ai.usage.input_tokens"] == 20 assert span["data"]["gen_ai.usage.total_tokens"] == 30 @@ -451,11 +517,11 @@ async def test_embeddings_create_async( tx = events[0] assert tx["type"] == "transaction" span = tx["spans"][0] - assert span["op"] == "ai.embeddings.create.openai" + assert span["op"] == "gen_ai.embeddings" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert span["data"]["gen_ai.usage.input_tokens"] == 20 assert span["data"]["gen_ai.usage.total_tokens"] == 30 @@ -897,3 +963,434 @@ def count_tokens(msg): output_tokens_reasoning=None, total_tokens=None, ) + + +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +def test_ai_client_span_responses_api_no_pii(sentry_init, capture_events): + sentry_init( + integrations=[OpenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.responses._post = mock.Mock(return_value=EXAMPLE_RESPONSE) + + with start_transaction(name="openai tx"): + client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + ) + + (transaction,) = events + spans = transaction["spans"] + + assert len(spans) == 1 + assert spans[0]["op"] == "gen_ai.responses" + assert spans[0]["origin"] == "auto.ai.openai" + assert spans[0]["data"] == { + "gen_ai.operation.name": "responses", + "gen_ai.request.model": "gpt-4o", + "gen_ai.response.model": "response-model-id", + "gen_ai.system": "openai", + "gen_ai.usage.input_tokens": 20, + "gen_ai.usage.input_tokens.cached": 5, + "gen_ai.usage.output_tokens": 10, + "gen_ai.usage.output_tokens.reasoning": 8, + "gen_ai.usage.total_tokens": 30, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + assert "gen_ai.request.messages" not in spans[0]["data"] + assert "gen_ai.response.text" not in spans[0]["data"] + + +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +def test_ai_client_span_responses_api(sentry_init, capture_events): + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.responses._post = mock.Mock(return_value=EXAMPLE_RESPONSE) + + with start_transaction(name="openai tx"): + client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + ) + + (transaction,) = events + spans = transaction["spans"] + + assert len(spans) == 1 + assert spans[0]["op"] == "gen_ai.responses" + assert spans[0]["origin"] == "auto.ai.openai" + assert spans[0]["data"] == { + "gen_ai.operation.name": "responses", + "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.model": "gpt-4o", + "gen_ai.system": "openai", + "gen_ai.response.model": "response-model-id", + "gen_ai.usage.input_tokens": 20, + "gen_ai.usage.input_tokens.cached": 5, + "gen_ai.usage.output_tokens": 10, + "gen_ai.usage.output_tokens.reasoning": 8, + "gen_ai.usage.total_tokens": 30, + "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +def test_error_in_responses_api(sentry_init, capture_events): + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.responses._post = mock.Mock( + side_effect=OpenAIError("API rate limit reached") + ) + + with start_transaction(name="openai tx"): + with pytest.raises(OpenAIError): + client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + ) + + (error_event, transaction_event) = events + + assert transaction_event["type"] == "transaction" + # make sure the span where the error occurred is captured + assert transaction_event["spans"][0]["op"] == "gen_ai.responses" + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "OpenAIError" + + assert ( + error_event["contexts"]["trace"]["trace_id"] + == transaction_event["contexts"]["trace"]["trace_id"] + ) + + +@pytest.mark.asyncio +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +async def test_ai_client_span_responses_async_api(sentry_init, capture_events): + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = AsyncOpenAI(api_key="z") + client.responses._post = AsyncMock(return_value=EXAMPLE_RESPONSE) + + with start_transaction(name="openai tx"): + await client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + ) + + (transaction,) = events + spans = transaction["spans"] + + assert len(spans) == 1 + assert spans[0]["op"] == "gen_ai.responses" + assert spans[0]["origin"] == "auto.ai.openai" + assert spans[0]["data"] == { + "gen_ai.operation.name": "responses", + "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.model": "gpt-4o", + "gen_ai.response.model": "response-model-id", + "gen_ai.system": "openai", + "gen_ai.usage.input_tokens": 20, + "gen_ai.usage.input_tokens.cached": 5, + "gen_ai.usage.output_tokens": 10, + "gen_ai.usage.output_tokens.reasoning": 8, + "gen_ai.usage.total_tokens": 30, + "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + +@pytest.mark.asyncio +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +async def test_ai_client_span_streaming_responses_async_api( + sentry_init, capture_events +): + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = AsyncOpenAI(api_key="z") + client.responses._post = AsyncMock(return_value=EXAMPLE_RESPONSE) + + with start_transaction(name="openai tx"): + await client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + stream=True, + ) + + (transaction,) = events + spans = transaction["spans"] + + assert len(spans) == 1 + assert spans[0]["op"] == "gen_ai.responses" + assert spans[0]["origin"] == "auto.ai.openai" + assert spans[0]["data"] == { + "gen_ai.operation.name": "responses", + "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.model": "gpt-4o", + "gen_ai.response.model": "response-model-id", + "gen_ai.response.streaming": True, + "gen_ai.system": "openai", + "gen_ai.usage.input_tokens": 20, + "gen_ai.usage.input_tokens.cached": 5, + "gen_ai.usage.output_tokens": 10, + "gen_ai.usage.output_tokens.reasoning": 8, + "gen_ai.usage.total_tokens": 30, + "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + +@pytest.mark.asyncio +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +async def test_error_in_responses_async_api(sentry_init, capture_events): + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = AsyncOpenAI(api_key="z") + client.responses._post = AsyncMock( + side_effect=OpenAIError("API rate limit reached") + ) + + with start_transaction(name="openai tx"): + with pytest.raises(OpenAIError): + await client.responses.create( + model="gpt-4o", + instructions="You are a coding assistant that talks like a pirate.", + input="How do I check if a Python object is an instance of a class?", + ) + + (error_event, transaction_event) = events + + assert transaction_event["type"] == "transaction" + # make sure the span where the error occurred is captured + assert transaction_event["spans"][0]["op"] == "gen_ai.responses" + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "OpenAIError" + + assert ( + error_event["contexts"]["trace"]["trace_id"] + == transaction_event["contexts"]["trace"]["trace_id"] + ) + + +if SKIP_RESPONSES_TESTS: + EXAMPLE_RESPONSES_STREAM = [] +else: + EXAMPLE_RESPONSES_STREAM = [ + ResponseCreatedEvent( + sequence_number=1, + type="response.created", + response=Response( + id="chat-id", + created_at=10000000, + model="response-model-id", + object="response", + output=[], + parallel_tool_calls=False, + tool_choice="none", + tools=[], + ), + ), + ResponseTextDeltaEvent( + item_id="msg_1", + sequence_number=2, + type="response.output_text.delta", + logprobs=[], + content_index=0, + output_index=0, + delta="hel", + ), + ResponseTextDeltaEvent( + item_id="msg_1", + sequence_number=3, + type="response.output_text.delta", + logprobs=[], + content_index=0, + output_index=0, + delta="lo ", + ), + ResponseTextDeltaEvent( + item_id="msg_1", + sequence_number=4, + type="response.output_text.delta", + logprobs=[], + content_index=0, + output_index=0, + delta="world", + ), + ResponseCompletedEvent( + sequence_number=5, + type="response.completed", + response=Response( + id="chat-id", + created_at=10000000, + model="response-model-id", + object="response", + output=[], + parallel_tool_calls=False, + tool_choice="none", + tools=[], + usage=ResponseUsage( + input_tokens=20, + input_tokens_details=InputTokensDetails( + cached_tokens=5, + ), + output_tokens=10, + output_tokens_details=OutputTokensDetails( + reasoning_tokens=8, + ), + total_tokens=30, + ), + ), + ), + ] + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +def test_streaming_responses_api( + sentry_init, capture_events, send_default_pii, include_prompts +): + sentry_init( + integrations=[ + OpenAIIntegration( + include_prompts=include_prompts, + ) + ], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + client = OpenAI(api_key="z") + returned_stream = Stream(cast_to=None, response=None, client=client) + returned_stream._iterator = EXAMPLE_RESPONSES_STREAM + client.responses._post = mock.Mock(return_value=returned_stream) + + with start_transaction(name="openai tx"): + response_stream = client.responses.create( + model="some-model", + input="hello", + stream=True, + ) + + response_string = "" + for item in response_stream: + if hasattr(item, "delta"): + response_string += item.delta + + assert response_string == "hello world" + + (transaction,) = events + (span,) = transaction["spans"] + assert span["op"] == "gen_ai.responses" + + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == "hello" + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "hello world" + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] + + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 + + +@pytest.mark.asyncio +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +@pytest.mark.skipif(SKIP_RESPONSES_TESTS, reason="Responses API not available") +async def test_streaming_responses_api_async( + sentry_init, capture_events, send_default_pii, include_prompts +): + sentry_init( + integrations=[ + OpenAIIntegration( + include_prompts=include_prompts, + ) + ], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + client = AsyncOpenAI(api_key="z") + returned_stream = AsyncStream(cast_to=None, response=None, client=client) + returned_stream._iterator = async_iterator(EXAMPLE_RESPONSES_STREAM) + client.responses._post = AsyncMock(return_value=returned_stream) + + with start_transaction(name="openai tx"): + response_stream = await client.responses.create( + model="some-model", + input="hello", + stream=True, + ) + + response_string = "" + async for item in response_stream: + if hasattr(item, "delta"): + response_string += item.delta + + assert response_string == "hello world" + + (transaction,) = events + (span,) = transaction["spans"] + assert span["op"] == "gen_ai.responses" + + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == "hello" + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "hello world" + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] + + assert span["data"]["gen_ai.usage.input_tokens"] == 20 + assert span["data"]["gen_ai.usage.output_tokens"] == 10 + assert span["data"]["gen_ai.usage.total_tokens"] == 30 From fd7dca446da0815e1491a5ec2acdb67e630646fd Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Jul 2025 13:16:59 +0200 Subject: [PATCH 380/591] fix(celery): Latency should be in milliseconds, not seconds (#4637) Fixes https://github.com/getsentry/sentry-python/issues/4636 --- sentry_sdk/integrations/celery/__init__.py | 1 + 1 file changed, 1 insertion(+) diff --git a/sentry_sdk/integrations/celery/__init__.py b/sentry_sdk/integrations/celery/__init__.py index d8d89217ca..b5601fc0f9 100644 --- a/sentry_sdk/integrations/celery/__init__.py +++ b/sentry_sdk/integrations/celery/__init__.py @@ -391,6 +391,7 @@ def _inner(*args, **kwargs): ) if latency is not None: + latency *= 1000 # milliseconds span.set_data(SPANDATA.MESSAGING_MESSAGE_RECEIVE_LATENCY, latency) with capture_internal_exceptions(): From 4f9d326c86477052aa30230393497a20edb17da4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Jul 2025 13:41:42 +0200 Subject: [PATCH 381/591] Considerably raise `DEFAULT_MAX_VALUE_LENGTH` (#4632) AI prompts/messages are potentially huge. * raise `DEFAULT_MAX_VALUE_LENGTH` (responsible for string trimming) from 1024 to 100 000 * adapt tests (and make them more generic, without hardcoded parts, where possible) --- sentry_sdk/consts.py | 5 ++- tests/integrations/bottle/test_bottle.py | 41 ++++++++++++----- tests/integrations/falcon/test_falcon.py | 14 ++++-- tests/integrations/flask/test_flask.py | 45 ++++++++++++++----- tests/integrations/pyramid/test_pyramid.py | 28 +++++++++--- .../sqlalchemy/test_sqlalchemy.py | 7 ++- tests/test_client.py | 11 +++-- tests/test_serializer.py | 5 ++- 8 files changed, 115 insertions(+), 41 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index a82ff94c49..ae8afecf57 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -3,7 +3,10 @@ from typing import TYPE_CHECKING # up top to prevent circular import due to integration import -DEFAULT_MAX_VALUE_LENGTH = 1024 +# This is more or less an arbitrary large-ish value for now, so that we allow +# pretty long strings (like LLM prompts), but still have *some* upper limit +# until we verify that removing the trimming completely is safe. +DEFAULT_MAX_VALUE_LENGTH = 100_000 DEFAULT_MAX_STACK_FRAMES = 100 DEFAULT_ADD_FULL_STACK = False diff --git a/tests/integrations/bottle/test_bottle.py b/tests/integrations/bottle/test_bottle.py index 363a9167e6..1965691d6c 100644 --- a/tests/integrations/bottle/test_bottle.py +++ b/tests/integrations/bottle/test_bottle.py @@ -5,6 +5,7 @@ from io import BytesIO from bottle import Bottle, debug as set_debug, abort, redirect, HTTPResponse from sentry_sdk import capture_message +from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH from sentry_sdk.integrations.bottle import BottleIntegration from sentry_sdk.serializer import MAX_DATABAG_BREADTH @@ -121,9 +122,9 @@ def index(): def test_large_json_request(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()], max_request_body_size="always") - data = {"foo": {"bar": "a" * 2000}} + data = {"foo": {"bar": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)}} @app.route("/", method="POST") def index(): @@ -144,9 +145,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"]["bar"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]["bar"]) == 1024 + assert len(event["request"]["data"]["foo"]["bar"]) == DEFAULT_MAX_VALUE_LENGTH @pytest.mark.parametrize("data", [{}, []], ids=["empty-dict", "empty-list"]) @@ -174,9 +180,9 @@ def index(): def test_medium_formdata_request(sentry_init, capture_events, app, get_client): - sentry_init(integrations=[BottleIntegration()]) + sentry_init(integrations=[BottleIntegration()], max_request_body_size="always") - data = {"foo": "a" * 2000} + data = {"foo": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)} @app.route("/", method="POST") def index(): @@ -194,9 +200,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]) == 1024 + assert len(event["request"]["data"]["foo"]) == DEFAULT_MAX_VALUE_LENGTH @pytest.mark.parametrize("input_char", ["a", b"a"]) @@ -233,7 +244,10 @@ def index(): def test_files_and_form(sentry_init, capture_events, app, get_client): sentry_init(integrations=[BottleIntegration()], max_request_body_size="always") - data = {"foo": "a" * 2000, "file": (BytesIO(b"hello"), "hello.txt")} + data = { + "foo": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10), + "file": (BytesIO(b"hello"), "hello.txt"), + } @app.route("/", method="POST") def index(): @@ -253,9 +267,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]) == 1024 + assert len(event["request"]["data"]["foo"]) == DEFAULT_MAX_VALUE_LENGTH assert event["_meta"]["request"]["data"]["file"] == { "": { diff --git a/tests/integrations/falcon/test_falcon.py b/tests/integrations/falcon/test_falcon.py index 51a1d94334..f972419092 100644 --- a/tests/integrations/falcon/test_falcon.py +++ b/tests/integrations/falcon/test_falcon.py @@ -5,6 +5,7 @@ import falcon import falcon.testing import sentry_sdk +from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH from sentry_sdk.integrations.falcon import FalconIntegration from sentry_sdk.integrations.logging import LoggingIntegration from sentry_sdk.utils import parse_version @@ -207,9 +208,9 @@ def on_get(self, req, resp): def test_falcon_large_json_request(sentry_init, capture_events): - sentry_init(integrations=[FalconIntegration()]) + sentry_init(integrations=[FalconIntegration()], max_request_body_size="always") - data = {"foo": {"bar": "a" * 2000}} + data = {"foo": {"bar": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)}} class Resource: def on_post(self, req, resp): @@ -228,9 +229,14 @@ def on_post(self, req, resp): (event,) = events assert event["_meta"]["request"]["data"]["foo"]["bar"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]["bar"]) == 1024 + assert len(event["request"]["data"]["foo"]["bar"]) == DEFAULT_MAX_VALUE_LENGTH @pytest.mark.parametrize("data", [{}, []], ids=["empty-dict", "empty-list"]) diff --git a/tests/integrations/flask/test_flask.py b/tests/integrations/flask/test_flask.py index 6febb12b8b..49ee684797 100644 --- a/tests/integrations/flask/test_flask.py +++ b/tests/integrations/flask/test_flask.py @@ -27,6 +27,7 @@ capture_message, capture_exception, ) +from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH from sentry_sdk.integrations.logging import LoggingIntegration from sentry_sdk.serializer import MAX_DATABAG_BREADTH @@ -248,9 +249,11 @@ def login(): def test_flask_large_json_request(sentry_init, capture_events, app): - sentry_init(integrations=[flask_sentry.FlaskIntegration()]) + sentry_init( + integrations=[flask_sentry.FlaskIntegration()], max_request_body_size="always" + ) - data = {"foo": {"bar": "a" * 2000}} + data = {"foo": {"bar": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)}} @app.route("/", methods=["POST"]) def index(): @@ -268,9 +271,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"]["bar"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]["bar"]) == 1024 + assert len(event["request"]["data"]["foo"]["bar"]) == DEFAULT_MAX_VALUE_LENGTH def test_flask_session_tracking(sentry_init, capture_envelopes, app): @@ -336,9 +344,11 @@ def index(): def test_flask_medium_formdata_request(sentry_init, capture_events, app): - sentry_init(integrations=[flask_sentry.FlaskIntegration()]) + sentry_init( + integrations=[flask_sentry.FlaskIntegration()], max_request_body_size="always" + ) - data = {"foo": "a" * 2000} + data = {"foo": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)} @app.route("/", methods=["POST"]) def index(): @@ -360,9 +370,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]) == 1024 + assert len(event["request"]["data"]["foo"]) == DEFAULT_MAX_VALUE_LENGTH def test_flask_formdata_request_appear_transaction_body( @@ -441,7 +456,10 @@ def test_flask_files_and_form(sentry_init, capture_events, app): integrations=[flask_sentry.FlaskIntegration()], max_request_body_size="always" ) - data = {"foo": "a" * 2000, "file": (BytesIO(b"hello"), "hello.txt")} + data = { + "foo": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10), + "file": (BytesIO(b"hello"), "hello.txt"), + } @app.route("/", methods=["POST"]) def index(): @@ -463,9 +481,14 @@ def index(): (event,) = events assert event["_meta"]["request"]["data"]["foo"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]) == 1024 + assert len(event["request"]["data"]["foo"]) == DEFAULT_MAX_VALUE_LENGTH assert event["_meta"]["request"]["data"]["file"] == {"": {"rem": [["!raw", "x"]]}} assert not event["request"]["data"]["file"] diff --git a/tests/integrations/pyramid/test_pyramid.py b/tests/integrations/pyramid/test_pyramid.py index d42d7887c4..cd200f7f7b 100644 --- a/tests/integrations/pyramid/test_pyramid.py +++ b/tests/integrations/pyramid/test_pyramid.py @@ -9,6 +9,7 @@ from werkzeug.test import Client from sentry_sdk import capture_message, add_breadcrumb +from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH from sentry_sdk.integrations.pyramid import PyramidIntegration from sentry_sdk.serializer import MAX_DATABAG_BREADTH from tests.conftest import unpack_werkzeug_response @@ -156,9 +157,9 @@ def test_transaction_style( def test_large_json_request(sentry_init, capture_events, route, get_client): - sentry_init(integrations=[PyramidIntegration()]) + sentry_init(integrations=[PyramidIntegration()], max_request_body_size="always") - data = {"foo": {"bar": "a" * 2000}} + data = {"foo": {"bar": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10)}} @route("/") def index(request): @@ -175,9 +176,14 @@ def index(request): (event,) = events assert event["_meta"]["request"]["data"]["foo"]["bar"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]["bar"]) == 1024 + assert len(event["request"]["data"]["foo"]["bar"]) == DEFAULT_MAX_VALUE_LENGTH @pytest.mark.parametrize("data", [{}, []], ids=["empty-dict", "empty-list"]) @@ -230,7 +236,10 @@ def index(request): def test_files_and_form(sentry_init, capture_events, route, get_client): sentry_init(integrations=[PyramidIntegration()], max_request_body_size="always") - data = {"foo": "a" * 2000, "file": (BytesIO(b"hello"), "hello.txt")} + data = { + "foo": "a" * (DEFAULT_MAX_VALUE_LENGTH + 10), + "file": (BytesIO(b"hello"), "hello.txt"), + } @route("/") def index(request): @@ -244,9 +253,14 @@ def index(request): (event,) = events assert event["_meta"]["request"]["data"]["foo"] == { - "": {"len": 2000, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } - assert len(event["request"]["data"]["foo"]) == 1024 + assert len(event["request"]["data"]["foo"]) == DEFAULT_MAX_VALUE_LENGTH assert event["_meta"]["request"]["data"]["file"] == {"": {"rem": [["!raw", "x"]]}} assert not event["request"]["data"]["file"] diff --git a/tests/integrations/sqlalchemy/test_sqlalchemy.py b/tests/integrations/sqlalchemy/test_sqlalchemy.py index 2b95fe02d4..d2a31a55d5 100644 --- a/tests/integrations/sqlalchemy/test_sqlalchemy.py +++ b/tests/integrations/sqlalchemy/test_sqlalchemy.py @@ -275,7 +275,12 @@ def processor(event, hint): # The _meta for other truncated fields should be there as well. assert event["_meta"]["message"] == { - "": {"len": 1034, "rem": [["!limit", "x", 1021, 1024]]} + "": { + "len": DEFAULT_MAX_VALUE_LENGTH + 10, + "rem": [ + ["!limit", "x", DEFAULT_MAX_VALUE_LENGTH - 3, DEFAULT_MAX_VALUE_LENGTH] + ], + } } diff --git a/tests/test_client.py b/tests/test_client.py index 9c6dbfe740..0468fcbb7b 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -773,14 +773,14 @@ def test_databag_string_stripping(sentry_init, capture_events, benchmark): def inner(): del events[:] try: - a = "A" * 1000000 # noqa + a = "A" * DEFAULT_MAX_VALUE_LENGTH * 10 # noqa 1 / 0 except Exception: capture_exception() (event,) = events - assert len(json.dumps(event)) < 10000 + assert len(json.dumps(event)) < DEFAULT_MAX_VALUE_LENGTH * 10 def test_databag_breadth_stripping(sentry_init, capture_events, benchmark): @@ -1073,7 +1073,10 @@ def test_multiple_positional_args(sentry_init): "sdk_options, expected_data_length", [ ({}, DEFAULT_MAX_VALUE_LENGTH), - ({"max_value_length": 1800}, 1800), + ( + {"max_value_length": DEFAULT_MAX_VALUE_LENGTH + 1000}, + DEFAULT_MAX_VALUE_LENGTH + 1000, + ), ], ) def test_max_value_length_option( @@ -1082,7 +1085,7 @@ def test_max_value_length_option( sentry_init(sdk_options) events = capture_events() - capture_message("a" * 2000) + capture_message("a" * (DEFAULT_MAX_VALUE_LENGTH + 2000)) assert len(events[0]["message"]) == expected_data_length diff --git a/tests/test_serializer.py b/tests/test_serializer.py index 2f158097bd..2f44ba8a08 100644 --- a/tests/test_serializer.py +++ b/tests/test_serializer.py @@ -2,6 +2,7 @@ import pytest +from sentry_sdk.consts import DEFAULT_MAX_VALUE_LENGTH from sentry_sdk.serializer import MAX_DATABAG_BREADTH, MAX_DATABAG_DEPTH, serialize try: @@ -166,11 +167,11 @@ def test_no_trimming_if_max_request_body_size_is_always(body_normalizer): def test_max_value_length_default(body_normalizer): - data = {"key": "a" * 2000} + data = {"key": "a" * (DEFAULT_MAX_VALUE_LENGTH * 10)} result = body_normalizer(data) - assert len(result["key"]) == 1024 # fallback max length + assert len(result["key"]) == DEFAULT_MAX_VALUE_LENGTH # fallback max length def test_max_value_length(body_normalizer): From e84f6f30682e0b14e5a2ab575d96c686894c5aaa Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 29 Jul 2025 11:52:15 +0000 Subject: [PATCH 382/591] release: 2.34.0 --- CHANGELOG.md | 18 ++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 21 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a2ac6e09f8..8fc40148b5 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,23 @@ # Changelog +## 2.34.0 + +### Various fixes & improvements + +- Considerably raise `DEFAULT_MAX_VALUE_LENGTH` (#4632) by @sentrivana +- fix(celery): Latency should be in milliseconds, not seconds (#4637) by @sentrivana +- OpenAI integration update (#4612) by @antonpirker +- tests: tox.ini update (#4635) by @sentrivana +- Expose set_transaction_name (#4634) by @sl0thentr0py +- Fix socket tests to not use example.com (#4627) by @sl0thentr0py +- Simplify celery double patching test (#4626) by @sl0thentr0py +- Treat django.template.context.BasicContext as sequence in serializer (#4621) by @sl0thentr0py +- Remove remote example.com calls (#4622) by @sl0thentr0py +- Fix `huggingface_hub` CI tests. (#4619) by @antonpirker +- tests: Update tox (#4609) by @sentrivana +- Ignore deliberate thread exception warnings (#4611) by @sl0thentr0py +- Fix threading run patch (#4610) by @sl0thentr0py + ## 2.33.2 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index faf861c518..c8debb897e 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.33.2" +release = "2.34.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index ae8afecf57..dd9055b869 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1204,4 +1204,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.33.2" +VERSION = "2.34.0" diff --git a/setup.py b/setup.py index 9e75720390..5f1640ac97 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.33.2", + version="2.34.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From e1848d4f33039f04b77caf43d3d2444a18ac2dac Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 29 Jul 2025 13:55:17 +0200 Subject: [PATCH 383/591] Update CHANGELOG.md --- CHANGELOG.md | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8fc40148b5..599fd87fd7 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,6 +5,16 @@ ### Various fixes & improvements - Considerably raise `DEFAULT_MAX_VALUE_LENGTH` (#4632) by @sentrivana + + We have increased the string trimming limit considerably, allowing you to see more data + without it being truncated. Note that this might, in rare cases, result in issue regrouping, + for example if you're capturing message events with very long messages (longer than the + default 1024 characters/bytes). + + If you want to adjust the limit, you can set a + [`max_value_limit`](https://docs.sentry.io/platforms/python/configuration/options/#max_value_length) + in your `sentry_sdk.init()`. + - fix(celery): Latency should be in milliseconds, not seconds (#4637) by @sentrivana - OpenAI integration update (#4612) by @antonpirker - tests: tox.ini update (#4635) by @sentrivana From 72766a79acf6df132f62584bd6ef4ac47904c155 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 29 Jul 2025 14:10:15 +0200 Subject: [PATCH 384/591] Update changelog --- CHANGELOG.md | 31 +++++++++++++++++++------------ 1 file changed, 19 insertions(+), 12 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 599fd87fd7..a1d046b4a0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -15,18 +15,25 @@ [`max_value_limit`](https://docs.sentry.io/platforms/python/configuration/options/#max_value_length) in your `sentry_sdk.init()`. -- fix(celery): Latency should be in milliseconds, not seconds (#4637) by @sentrivana -- OpenAI integration update (#4612) by @antonpirker -- tests: tox.ini update (#4635) by @sentrivana -- Expose set_transaction_name (#4634) by @sl0thentr0py -- Fix socket tests to not use example.com (#4627) by @sl0thentr0py -- Simplify celery double patching test (#4626) by @sl0thentr0py -- Treat django.template.context.BasicContext as sequence in serializer (#4621) by @sl0thentr0py -- Remove remote example.com calls (#4622) by @sl0thentr0py -- Fix `huggingface_hub` CI tests. (#4619) by @antonpirker -- tests: Update tox (#4609) by @sentrivana -- Ignore deliberate thread exception warnings (#4611) by @sl0thentr0py -- Fix threading run patch (#4610) by @sl0thentr0py +- `OpenAI` integration update (#4612) by @antonpirker + + The `OpenAIIntegration` now supports [OpenAI Responses API](https://platform.openai.com/docs/api-reference/responses). + + The data captured will also show up in the new [AI Agents Dashboard](https://docs.sentry.io/product/insights/agents/dashboard/). + + This works out of the box, nothing to do on your side. + +- Expose `set_transaction_name` (#4634) by @sl0thentr0py +- Fix(Celery): Latency should be in milliseconds, not seconds (#4637) by @sentrivana +- Fix(Django): Treat `django.template.context.BasicContext` as sequence in serializer (#4621) by @sl0thentr0py +- Fix(Huggingface): Fix `huggingface_hub` CI tests. (#4619) by @antonpirker +- Fix: Ignore deliberate thread exception warnings (#4611) by @sl0thentr0py +- Fix: Socket tests to not use example.com (#4627) by @sl0thentr0py +- Fix: Threading run patch (#4610) by @sl0thentr0py +- Tests: Simplify celery double patching test (#4626) by @sl0thentr0py +- Tests: Remove remote example.com calls (#4622) by @sl0thentr0py +- Tests: tox.ini update (#4635) by @sentrivana +- Tests: Update tox (#4609) by @sentrivana ## 2.33.2 From bab6215b1f8b6992c0343992601896adc1963897 Mon Sep 17 00:00:00 2001 From: James Gillard Date: Wed, 30 Jul 2025 09:44:59 +0100 Subject: [PATCH 385/591] Fix typo in CHANGELOG.md (#4640) Just a typo in the changelog that confused me for a minute while I searched for that variable. --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a1d046b4a0..b3111eeee8 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -12,7 +12,7 @@ default 1024 characters/bytes). If you want to adjust the limit, you can set a - [`max_value_limit`](https://docs.sentry.io/platforms/python/configuration/options/#max_value_length) + [`max_value_length`](https://docs.sentry.io/platforms/python/configuration/options/#max_value_length) in your `sentry_sdk.init()`. - `OpenAI` integration update (#4612) by @antonpirker From 38b570a467c9633c8c28c8486433038c1a19fdda Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 30 Jul 2025 12:35:12 +0200 Subject: [PATCH 386/591] Span data is always be a primitive data type (#4643) The AI Agent insights module excepts the data not to be lists, tuples, or dicts. Make sure that we always send a string in this case. --- sentry_sdk/ai/utils.py | 6 +++++- tests/integrations/cohere/test_cohere.py | 20 +++++++++++++++---- .../integrations/langchain/test_langchain.py | 6 ++---- tests/integrations/openai/test_openai.py | 19 ++++++------------ 4 files changed, 29 insertions(+), 22 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index ed3494f679..a3c62600c0 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -23,10 +23,14 @@ def _normalize_data(data): return list(_normalize_data(x) for x in data) if isinstance(data, dict): return {k: _normalize_data(v) for (k, v) in data.items()} + return data def set_data_normalized(span, key, value): # type: (Span, str, Any) -> None normalized = _normalize_data(value) - span.set_data(key, normalized) + if isinstance(normalized, (int, float, bool, str)): + span.set_data(key, normalized) + else: + span.set_data(key, str(normalized)) diff --git a/tests/integrations/cohere/test_cohere.py b/tests/integrations/cohere/test_cohere.py index f13a77ae90..b8b6067625 100644 --- a/tests/integrations/cohere/test_cohere.py +++ b/tests/integrations/cohere/test_cohere.py @@ -57,8 +57,14 @@ def test_nonstreaming_chat( assert span["data"][SPANDATA.AI_MODEL_ID] == "some-model" if send_default_pii and include_prompts: - assert "some context" in span["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES][1]["content"] + assert ( + "{'role': 'system', 'content': 'some context'}" + in span["data"][SPANDATA.AI_INPUT_MESSAGES] + ) + assert ( + "{'role': 'user', 'content': 'hello'}" + in span["data"][SPANDATA.AI_INPUT_MESSAGES] + ) assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] @@ -128,8 +134,14 @@ def test_streaming_chat(sentry_init, capture_events, send_default_pii, include_p assert span["data"][SPANDATA.AI_MODEL_ID] == "some-model" if send_default_pii and include_prompts: - assert "some context" in span["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES][1]["content"] + assert ( + "{'role': 'system', 'content': 'some context'}" + in span["data"][SPANDATA.AI_INPUT_MESSAGES] + ) + assert ( + "{'role': 'user', 'content': 'hello'}" + in span["data"][SPANDATA.AI_INPUT_MESSAGES] + ) assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] else: assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index ee9fb241b1..9d55a49f82 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -196,15 +196,13 @@ def test_langchain_agent( if send_default_pii and include_prompts: assert ( - "You are very powerful" - in chat_spans[0]["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] + "You are very powerful" in chat_spans[0]["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert "5" in chat_spans[0]["data"][SPANDATA.AI_RESPONSES] assert "word" in tool_exec_span["data"][SPANDATA.AI_INPUT_MESSAGES] assert 5 == int(tool_exec_span["data"][SPANDATA.AI_RESPONSES]) assert ( - "You are very powerful" - in chat_spans[1]["data"][SPANDATA.AI_INPUT_MESSAGES][0]["content"] + "You are very powerful" in chat_spans[1]["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert "5" in chat_spans[1]["data"][SPANDATA.AI_RESPONSES] else: diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index dfac08d762..5767f84d04 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1,4 +1,3 @@ -import json import pytest from openai import AsyncOpenAI, OpenAI, AsyncStream, Stream, OpenAIError from openai.types import CompletionUsage, CreateEmbeddingResponse, Embedding @@ -144,11 +143,8 @@ def test_nonstreaming_chat_completion( assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] - assert ( - "the model response" - in json.loads(span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT])[0]["content"] - ) + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert "the model response" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] @@ -189,11 +185,8 @@ async def test_nonstreaming_chat_completion_async( assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] - assert ( - "the model response" - in json.loads(span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT])[0]["content"] - ) + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert "the model response" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] @@ -285,7 +278,7 @@ def test_streaming_chat_completion( assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] assert "hello world" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] @@ -381,7 +374,7 @@ async def test_streaming_chat_completion_async( assert span["op"] == "gen_ai.chat" if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]["content"] + assert "hello" in span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] assert "hello world" in span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] From a71ef66d37aa77316713c9e312891009727d55fe Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Wed, 30 Jul 2025 11:05:25 +0000 Subject: [PATCH 387/591] release: 2.34.1 --- CHANGELOG.md | 7 +++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 10 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index b3111eeee8..a447850e24 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,12 @@ # Changelog +## 2.34.1 + +### Various fixes & improvements + +- Span data is always be a primitive data type (#4643) by @antonpirker +- Fix typo in CHANGELOG.md (#4640) by @jgillard + ## 2.34.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index c8debb897e..f5d0b9e121 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.34.0" +release = "2.34.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index dd9055b869..3ae33b6a94 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1204,4 +1204,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.34.0" +VERSION = "2.34.1" diff --git a/setup.py b/setup.py index 5f1640ac97..11b02cbca8 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.34.0", + version="2.34.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 9276f2a150c1d0f831d54959b8dc7b138cd50bb6 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 30 Jul 2025 13:06:53 +0200 Subject: [PATCH 388/591] update changelog --- CHANGELOG.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a447850e24..21b1d5fec9 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,8 +4,8 @@ ### Various fixes & improvements -- Span data is always be a primitive data type (#4643) by @antonpirker -- Fix typo in CHANGELOG.md (#4640) by @jgillard +- Fix: Make sure Span data in AI instrumentations is always a primitive data type (#4643) by @antonpirker +- Fix: Typo in CHANGELOG.md (#4640) by @jgillard ## 2.34.0 From 493ac4bb088954f69c19174c23832e11a5a7dcb6 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 30 Jul 2025 18:02:51 +0200 Subject: [PATCH 389/591] Better checking for empty tools list (#4647) Fixes #4646 --- sentry_sdk/integrations/openai.py | 10 ++++-- tests/integrations/openai/test_openai.py | 40 ++++++++++++++++++++++++ 2 files changed, 48 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 78fcdd49e2..187f795807 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -20,6 +20,11 @@ from sentry_sdk.tracing import Span try: + try: + from openai import NOT_GIVEN + except ImportError: + NOT_GIVEN = None + from openai.resources.chat.completions import Completions, AsyncCompletions from openai.resources import Embeddings, AsyncEmbeddings @@ -192,12 +197,13 @@ def _set_input_data(span, kwargs, operation, integration): } for key, attribute in kwargs_keys_to_attributes.items(): value = kwargs.get(key) - if value is not None: + + if value is not NOT_GIVEN and value is not None: set_data_normalized(span, attribute, value) # Input attributes: Tools tools = kwargs.get("tools") - if tools is not None and len(tools) > 0: + if tools is not NOT_GIVEN and tools is not None and len(tools) > 0: set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) ) diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 5767f84d04..a3c7bdd9d9 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1,4 +1,12 @@ import pytest + +from sentry_sdk.utils import package_version + +try: + from openai import NOT_GIVEN +except ImportError: + NOT_GIVEN = None + from openai import AsyncOpenAI, OpenAI, AsyncStream, Stream, OpenAIError from openai.types import CompletionUsage, CreateEmbeddingResponse, Embedding from openai.types.chat import ChatCompletion, ChatCompletionMessage, ChatCompletionChunk @@ -43,6 +51,7 @@ async def __call__(self, *args, **kwargs): return super(AsyncMock, self).__call__(*args, **kwargs) +OPENAI_VERSION = package_version("openai") EXAMPLE_CHAT_COMPLETION = ChatCompletion( id="chat-id", choices=[ @@ -1387,3 +1396,34 @@ async def test_streaming_responses_api_async( assert span["data"]["gen_ai.usage.input_tokens"] == 20 assert span["data"]["gen_ai.usage.output_tokens"] == 10 assert span["data"]["gen_ai.usage.total_tokens"] == 30 + + +@pytest.mark.skipif( + OPENAI_VERSION <= (1, 1, 0), + reason="OpenAI versions <=1.1.0 do not support the tools parameter.", +) +@pytest.mark.parametrize( + "tools", + [[], None, NOT_GIVEN], +) +def test_empty_tools_in_chat_completion(sentry_init, capture_events, tools): + sentry_init( + integrations=[OpenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.chat.completions._post = mock.Mock(return_value=EXAMPLE_CHAT_COMPLETION) + + with start_transaction(name="openai tx"): + client.chat.completions.create( + model="some-model", + messages=[{"role": "system", "content": "hello"}], + tools=tools, + ) + + (event,) = events + span = event["spans"][0] + + assert "gen_ai.request.available_tools" not in span["data"] From c1861a3ca963512f1609cc125dd2648dc029b64b Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 31 Jul 2025 12:01:59 +0200 Subject: [PATCH 390/591] Fix mypy (#4649) --- sentry_sdk/integrations/openfeature.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/openfeature.py b/sentry_sdk/integrations/openfeature.py index e2b33d83f2..3ac73edd93 100644 --- a/sentry_sdk/integrations/openfeature.py +++ b/sentry_sdk/integrations/openfeature.py @@ -1,4 +1,4 @@ -from typing import TYPE_CHECKING +from typing import TYPE_CHECKING, Any from sentry_sdk.feature_flags import add_feature_flag from sentry_sdk.integrations import DidNotEnable, Integration @@ -8,7 +8,6 @@ from openfeature.hook import Hook if TYPE_CHECKING: - from openfeature.flag_evaluation import FlagEvaluationDetails from openfeature.hook import HookContext, HookHints except ImportError: raise DidNotEnable("OpenFeature is not installed") @@ -25,9 +24,8 @@ def setup_once(): class OpenFeatureHook(Hook): - def after(self, hook_context, details, hints): - # type: (HookContext, FlagEvaluationDetails[bool], HookHints) -> None + # type: (Any, Any, Any) -> None if isinstance(details.value, bool): add_feature_flag(details.flag_key, details.value) From 3425d4c936420d223a2c015bd6325fe96b12719c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 31 Jul 2025 14:20:06 +0200 Subject: [PATCH 391/591] Add `enable_logs`, `before_send_log` as top-level options (#4644) Promote `enable_logs` and `before_send_log` to regular (non-experimental) SDK options. Keep supporting the experimental versions too, for backwards compat. Closes https://github.com/getsentry/sentry-python/issues/4641 --- sentry_sdk/client.py | 13 ++-- sentry_sdk/consts.py | 11 +++- sentry_sdk/integrations/logging.py | 3 +- sentry_sdk/integrations/loguru.py | 3 +- sentry_sdk/utils.py | 23 ++++++- tests/integrations/logging/test_logging.py | 16 ++--- tests/integrations/loguru/test_loguru.py | 18 ++--- tests/test_logs.py | 76 +++++++++++++++++----- 8 files changed, 122 insertions(+), 41 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index dca39beab8..5d584a5537 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -23,6 +23,8 @@ handle_in_app, is_gevent, logger, + get_before_send_log, + has_logs_enabled, ) from sentry_sdk.serializer import serialize from sentry_sdk.tracing import trace @@ -382,7 +384,8 @@ def _capture_envelope(envelope): ) self.log_batcher = None - if experiments.get("enable_logs", False): + + if has_logs_enabled(self.options): from sentry_sdk._log_batcher import LogBatcher self.log_batcher = LogBatcher(capture_func=_capture_envelope) @@ -898,9 +901,8 @@ def capture_event( return return_value def _capture_experimental_log(self, log): - # type: (Log) -> None - logs_enabled = self.options["_experiments"].get("enable_logs", False) - if not logs_enabled: + # type: (Optional[Log]) -> None + if not has_logs_enabled(self.options) or log is None: return current_scope = sentry_sdk.get_current_scope() @@ -955,9 +957,10 @@ def _capture_experimental_log(self, log): f'[Sentry Logs] [{log.get("severity_text")}] {log.get("body")}' ) - before_send_log = self.options["_experiments"].get("before_send_log") + before_send_log = get_before_send_log(self.options) if before_send_log is not None: log = before_send_log(log, {}) + if log is None: return diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 3ae33b6a94..b56c0ba2dd 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -798,6 +798,8 @@ def __init__( custom_repr=None, # type: Optional[Callable[..., Optional[str]]] add_full_stack=DEFAULT_ADD_FULL_STACK, # type: bool max_stack_frames=DEFAULT_MAX_STACK_FRAMES, # type: Optional[int] + enable_logs=False, # type: bool + before_send_log=None, # type: Optional[Callable[[Log, Hint], Optional[Log]]] ): # type: (...) -> None """Initialize the Sentry SDK with the given parameters. All parameters described here can be used in a call to `sentry_sdk.init()`. @@ -1168,7 +1170,6 @@ def __init__( :param profile_session_sample_rate: - :param enable_tracing: :param propagate_traces: @@ -1179,6 +1180,14 @@ def __init__( :param instrumenter: + :param enable_logs: Set `enable_logs` to True to enable the SDK to emit + Sentry logs. Defaults to False. + + :param before_send_log: An optional function to modify or filter out logs + before they're sent to Sentry. Any modifications to the log in this + function will be retained. If the function returns None, the log will + not be sent to Sentry. + :param _experiments: """ pass diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index a50512f622..15ff2ed233 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -12,6 +12,7 @@ event_from_exception, current_stacktrace, capture_internal_exceptions, + has_logs_enabled, ) from sentry_sdk.integrations import Integration @@ -344,7 +345,7 @@ def emit(self, record): if not client.is_active(): return - if not client.options["_experiments"].get("enable_logs", False): + if not has_logs_enabled(client.options): return self._capture_log_from_record(client, record) diff --git a/sentry_sdk/integrations/loguru.py b/sentry_sdk/integrations/loguru.py index df3ecf161a..b910b9a407 100644 --- a/sentry_sdk/integrations/loguru.py +++ b/sentry_sdk/integrations/loguru.py @@ -8,6 +8,7 @@ _BaseHandler, ) from sentry_sdk.logger import _log_level_to_otel +from sentry_sdk.utils import has_logs_enabled from typing import TYPE_CHECKING @@ -151,7 +152,7 @@ def loguru_sentry_logs_handler(message): if not client.is_active(): return - if not client.options["_experiments"].get("enable_logs", False): + if not has_logs_enabled(client.options): return record = message.record diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 9c6f2cfc3b..b0f3fa4a4c 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -59,7 +59,7 @@ from gevent.hub import Hub - from sentry_sdk._types import Event, ExcInfo + from sentry_sdk._types import Event, ExcInfo, Log, Hint P = ParamSpec("P") R = TypeVar("R") @@ -1984,3 +1984,24 @@ def serialize_item(item): return json.dumps(serialized, default=str) except Exception: return str(data) + + +def has_logs_enabled(options): + # type: (Optional[dict[str, Any]]) -> bool + if options is None: + return False + + return bool( + options.get("enable_logs", False) + or options["_experiments"].get("enable_logs", False) + ) + + +def get_before_send_log(options): + # type: (Optional[dict[str, Any]]) -> Optional[Callable[[Log, Hint], Optional[Log]]] + if options is None: + return None + + return options.get("before_send_log") or options["_experiments"].get( + "before_send_log" + ) diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 6ef4ae371b..7ecdf42500 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -304,7 +304,7 @@ def test_sentry_logs_warning(sentry_init, capture_envelopes): """ The python logger module should create 'warn' sentry logs if the flag is on. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -329,7 +329,7 @@ def test_sentry_logs_debug(sentry_init, capture_envelopes): """ The python logger module should not create 'debug' sentry logs if the flag is on by default """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -344,7 +344,7 @@ def test_no_log_infinite_loop(sentry_init, capture_envelopes): If 'debug' mode is true, and you set a low log level in the logging integration, there should be no infinite loops. """ sentry_init( - _experiments={"enable_logs": True}, + enable_logs=True, integrations=[LoggingIntegration(sentry_logs_level=logging.DEBUG)], debug=True, ) @@ -361,7 +361,7 @@ def test_logging_errors(sentry_init, capture_envelopes): """ The python logger module should be able to log errors without erroring """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -396,7 +396,7 @@ def test_log_strips_project_root(sentry_init, capture_envelopes): The python logger should strip project roots from the log record path """ sentry_init( - _experiments={"enable_logs": True}, + enable_logs=True, project_root="/custom/test", ) envelopes = capture_envelopes() @@ -425,7 +425,7 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): """ The python logger should be able to log all attributes, including extra data. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -498,7 +498,7 @@ def test_sentry_logs_named_parameters(sentry_init, capture_envelopes): """ The python logger module should capture named parameters from dictionary arguments in Sentry logs. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -543,7 +543,7 @@ def test_sentry_logs_named_parameters_complex_values(sentry_init, capture_envelo """ The python logger module should handle complex values in named parameters using safe_repr. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index c120d1d7e2..38093d24cb 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -141,7 +141,7 @@ def test_sentry_logs_warning( uninstall_integration("loguru") request.addfinalizer(logger.remove) - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() logger.warning("this is {} a {}", "just", "template") @@ -165,7 +165,7 @@ def test_sentry_logs_debug( uninstall_integration("loguru") request.addfinalizer(logger.remove) - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() logger.debug("this is %s a template %s", "1", "2") @@ -182,7 +182,7 @@ def test_sentry_log_levels( sentry_init( integrations=[LoguruIntegration(sentry_logs_level=LoggingLevels.SUCCESS)], - _experiments={"enable_logs": True}, + enable_logs=True, ) envelopes = capture_envelopes() @@ -216,7 +216,7 @@ def test_disable_loguru_logs( sentry_init( integrations=[LoguruIntegration(sentry_logs_level=None)], - _experiments={"enable_logs": True}, + enable_logs=True, ) envelopes = capture_envelopes() @@ -267,7 +267,7 @@ def test_no_log_infinite_loop( request.addfinalizer(logger.remove) sentry_init( - _experiments={"enable_logs": True}, + enable_logs=True, integrations=[LoguruIntegration(sentry_logs_level=LoggingLevels.DEBUG)], debug=True, ) @@ -284,7 +284,7 @@ def test_logging_errors(sentry_init, capture_envelopes, uninstall_integration, r uninstall_integration("loguru") request.addfinalizer(logger.remove) - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() logger.error(Exception("test exc 1")) @@ -313,7 +313,7 @@ def test_log_strips_project_root( request.addfinalizer(logger.remove) sentry_init( - _experiments={"enable_logs": True}, + enable_logs=True, project_root="/custom/test", ) envelopes = capture_envelopes() @@ -362,7 +362,7 @@ def test_log_keeps_full_path_if_not_in_project_root( request.addfinalizer(logger.remove) sentry_init( - _experiments={"enable_logs": True}, + enable_logs=True, project_root="/custom/test", ) envelopes = capture_envelopes() @@ -410,7 +410,7 @@ def test_logger_with_all_attributes( uninstall_integration("loguru") request.addfinalizer(logger.remove) - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() logger.warning("log #{}", 1) diff --git a/tests/test_logs.py b/tests/test_logs.py index a2f412dcb0..b2578d83d5 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -80,7 +80,7 @@ def test_logs_disabled_by_default(sentry_init, capture_envelopes): @minimum_python_37 def test_logs_basics(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() sentry_sdk.logger.trace("This is a 'trace' log...") @@ -111,11 +111,29 @@ def test_logs_basics(sentry_init, capture_envelopes): assert logs[5].get("severity_number") == 21 +@minimum_python_37 +def test_logs_experimental_option_still_works(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_logs": True}) + envelopes = capture_envelopes() + + sentry_sdk.logger.error("This is an error log...") + + get_client().flush() + + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + + assert logs[0].get("severity_text") == "error" + assert logs[0].get("severity_number") == 17 + + @minimum_python_37 def test_logs_before_send_log(sentry_init, capture_envelopes): - before_log_called = [False] + before_log_called = False def _before_log(record, hint): + nonlocal before_log_called + assert set(record.keys()) == { "severity_text", "severity_number", @@ -128,15 +146,13 @@ def _before_log(record, hint): if record["severity_text"] in ["fatal", "error"]: return None - before_log_called[0] = True + before_log_called = True return record sentry_init( - _experiments={ - "enable_logs": True, - "before_send_log": _before_log, - } + enable_logs=True, + before_send_log=_before_log, ) envelopes = capture_envelopes() @@ -155,7 +171,37 @@ def _before_log(record, hint): assert logs[1]["severity_text"] == "debug" assert logs[2]["severity_text"] == "info" assert logs[3]["severity_text"] == "warn" - assert before_log_called[0] + assert before_log_called is True + + +@minimum_python_37 +def test_logs_before_send_log_experimental_option_still_works( + sentry_init, capture_envelopes +): + before_log_called = False + + def _before_log(record, hint): + nonlocal before_log_called + before_log_called = True + + return record + + sentry_init( + enable_logs=True, + _experiments={ + "before_send_log": _before_log, + }, + ) + envelopes = capture_envelopes() + + sentry_sdk.logger.error("This is an error log...") + + get_client().flush() + logs = envelopes_to_logs(envelopes) + assert len(logs) == 1 + + assert logs[0]["severity_text"] == "error" + assert before_log_called is True @minimum_python_37 @@ -163,7 +209,7 @@ def test_logs_attributes(sentry_init, capture_envelopes): """ Passing arbitrary attributes to log messages. """ - sentry_init(_experiments={"enable_logs": True}, server_name="test-server") + sentry_init(enable_logs=True, server_name="test-server") envelopes = capture_envelopes() attrs = { @@ -196,7 +242,7 @@ def test_logs_message_params(sentry_init, capture_envelopes): """ This is the official way of how to pass vars to log messages. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() sentry_sdk.logger.warning("The recorded value was '{int_var}'", int_var=1) @@ -239,7 +285,7 @@ def test_logs_tied_to_transactions(sentry_init, capture_envelopes): """ Log messages are also tied to transactions. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction") as trx: @@ -255,7 +301,7 @@ def test_logs_tied_to_spans(sentry_init, capture_envelopes): """ Log messages are also tied to spans. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction"): @@ -271,7 +317,7 @@ def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): """ If you log >100 logs, it should automatically trigger a flush. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") @@ -288,7 +334,7 @@ def test_auto_flush_logs_after_100(sentry_init, capture_envelopes): def test_log_user_attributes(sentry_init, capture_envelopes): """User attributes are sent if enable_logs is True.""" - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) sentry_sdk.set_user({"id": "1", "email": "test@example.com", "username": "test"}) envelopes = capture_envelopes() @@ -314,7 +360,7 @@ def test_auto_flush_logs_after_5s(sentry_init, capture_envelopes): """ If you log a single log, it should automatically flush after 5 seconds, at most 10 seconds. """ - sentry_init(_experiments={"enable_logs": True}) + sentry_init(enable_logs=True) envelopes = capture_envelopes() python_logger = logging.Logger("test-logger") From 84adbb74e27b7716cbaaddbb299a44fe3fbcd6a6 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 31 Jul 2025 17:05:24 +0200 Subject: [PATCH 392/591] Fix plugins key codecov (#4655) was deprecated https://github.com/codecov/codecov-action?tab=readme-ov-file#migration-guide --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 13 files changed, 19 insertions(+), 19 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 2777810a8f..dd57f5909b 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -92,7 +92,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -171,7 +171,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 6a9b9df0de..e79c9513ef 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -92,7 +92,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -171,7 +171,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 2ceb23b79c..c7e356420c 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -72,7 +72,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 1ad39421d6..6c203379fe 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -112,7 +112,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -211,7 +211,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index d6da6c8acd..926465990d 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -84,7 +84,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index c0bd099e45..a08e91c909 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -72,7 +72,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index e851dfc9bb..9bbeee6c6a 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -84,7 +84,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 8a2e87c9ca..3595640ce1 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -92,7 +92,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 47ae674934..3ac5508dab 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -80,7 +80,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -147,7 +147,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 6b3fcab41f..13c34224be 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -107,7 +107,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -201,7 +201,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 3b48472d5e..e52a903208 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -102,7 +102,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index b98e5f02fc..c703cfafce 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -108,7 +108,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} @@ -203,7 +203,7 @@ jobs: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov if: ${{ !cancelled() }} diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 44f51f473e..96faefc54e 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -105,7 +105,7 @@ token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml # make sure no plugins alter our coverage reports - plugin: noop + plugins: noop verbose: true - name: Upload test results to Codecov From 0d569d29fa80f8efc4b992e7cb1bfc8266e7909b Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 5 Aug 2025 10:55:03 +0200 Subject: [PATCH 393/591] Add `update_data` to `Span`. (#4666) This allows users to set multiple `span.data` attributes at once. In 3.x this then will become `set_attributes` (plural). --- sentry_sdk/tracing.py | 8 ++++++++ tests/tracing/test_misc.py | 31 +++++++++++++++++++++++++++++++ 2 files changed, 39 insertions(+) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index fc40221b9f..dd1392d150 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -602,6 +602,10 @@ def set_data(self, key, value): # type: (str, Any) -> None self._data[key] = value + def update_data(self, data): + # type: (Dict[str, Any]) -> None + self._data.update(data) + def set_flag(self, flag, result): # type: (str, bool) -> None if len(self._flags) < self._flags_capacity: @@ -1275,6 +1279,10 @@ def set_data(self, key, value): # type: (str, Any) -> None pass + def update_data(self, data): + # type: (Dict[str, Any]) -> None + pass + def set_status(self, value): # type: (str) -> None pass diff --git a/tests/tracing/test_misc.py b/tests/tracing/test_misc.py index b954d36e1a..651228b45e 100644 --- a/tests/tracing/test_misc.py +++ b/tests/tracing/test_misc.py @@ -509,3 +509,34 @@ def test_transaction_not_started_warning(sentry_init): "The transaction will not be sent to Sentry. To fix, start the transaction by" "passing it to sentry_sdk.start_transaction." ) + + +def test_span_set_data_update_data(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + + events = capture_events() + + with sentry_sdk.start_transaction(name="test-transaction"): + with start_span(op="test-span") as span: + span.set_data("key0", "value0") + span.set_data("key1", "value1") + + span.update_data( + { + "key1": "updated-value1", + "key2": "value2", + "key3": "value3", + } + ) + + (event,) = events + span = event["spans"][0] + + assert span["data"] == { + "key0": "value0", + "key1": "updated-value1", + "key2": "value2", + "key3": "value3", + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } From 63e0c670deb28a0c027736a8dd8ef7e8be8f8ef7 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 5 Aug 2025 11:05:47 +0200 Subject: [PATCH 394/591] Update `gen_ai.*` and `ai.*` attributes (#4665) All `ai.*` attributes are deprecated. I also added some missing `gen_ai.*` attributes from [Sentry Conventions](https://getsentry.github.io/sentry-conventions/generated/attributes/gen_ai.html). --- sentry_sdk/consts.py | 115 +++++++++++++++++++++++++++++++++++++++++-- 1 file changed, 112 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index b56c0ba2dd..d402467e5e 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -113,71 +113,106 @@ class SPANDATA: AI_CITATIONS = "ai.citations" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + References or sources cited by the AI model in its response. Example: ["Smith et al. 2020", "Jones 2019"] """ AI_DOCUMENTS = "ai.documents" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Documents or content chunks used as context for the AI model. Example: ["doc1.txt", "doc2.pdf"] """ AI_FINISH_REASON = "ai.finish_reason" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_FINISH_REASONS instead. + The reason why the model stopped generating. Example: "length" """ AI_FREQUENCY_PENALTY = "ai.frequency_penalty" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_FREQUENCY_PENALTY instead. + Used to reduce repetitiveness of generated tokens. Example: 0.5 """ AI_FUNCTION_CALL = "ai.function_call" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_TOOL_CALLS instead. + For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls """ AI_GENERATION_ID = "ai.generation_id" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_ID instead. + Unique identifier for the completion. Example: "gen_123abc" """ AI_INPUT_MESSAGES = "ai.input_messages" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_MESSAGES instead. + The input messages to an LLM call. Example: [{"role": "user", "message": "hello"}] """ AI_LOGIT_BIAS = "ai.logit_bias" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + For an AI model call, the logit bias """ AI_METADATA = "ai.metadata" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Extra metadata passed to an AI pipeline step. Example: {"executed_function": "add_integers"} """ AI_MODEL_ID = "ai.model_id" """ - The unique descriptor of the model being execugted + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_MODEL or GEN_AI_RESPONSE_MODEL instead. + + The unique descriptor of the model being executed. Example: gpt-4 """ AI_PIPELINE_NAME = "ai.pipeline.name" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_PIPELINE_NAME instead. + Name of the AI pipeline or chain being executed. - DEPRECATED: Use GEN_AI_PIPELINE_NAME instead. Example: "qa-pipeline" """ AI_PREAMBLE = "ai.preamble" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + For an AI model call, the preamble parameter. Preambles are a part of the prompt used to adjust the model's overall behavior and conversation style. Example: "You are now a clown." @@ -185,100 +220,150 @@ class SPANDATA: AI_PRESENCE_PENALTY = "ai.presence_penalty" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_PRESENCE_PENALTY instead. + Used to reduce repetitiveness of generated tokens. Example: 0.5 """ AI_RAW_PROMPTING = "ai.raw_prompting" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Minimize pre-processing done to the prompt sent to the LLM. Example: true """ AI_RESPONSE_FORMAT = "ai.response_format" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + For an AI model call, the format of the response """ AI_RESPONSES = "ai.responses" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_TEXT instead. + The responses to an AI model call. Always as a list. Example: ["hello", "world"] """ AI_SEARCH_QUERIES = "ai.search_queries" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Queries used to search for relevant context or documents. Example: ["climate change effects", "renewable energy"] """ AI_SEARCH_REQUIRED = "ai.is_search_required" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Boolean indicating if the model needs to perform a search. Example: true """ AI_SEARCH_RESULTS = "ai.search_results" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Results returned from search queries for context. Example: ["Result 1", "Result 2"] """ AI_SEED = "ai.seed" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_SEED instead. + The seed, ideally models given the same seed and same other parameters will produce the exact same output. Example: 123.45 """ AI_STREAMING = "ai.streaming" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_STREAMING instead. + Whether or not the AI model call's response was streamed back asynchronously - DEPRECATED: Use GEN_AI_RESPONSE_STREAMING instead. Example: true """ AI_TAGS = "ai.tags" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Tags that describe an AI pipeline step. Example: {"executed_function": "add_integers"} """ AI_TEMPERATURE = "ai.temperature" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_TEMPERATURE instead. + For an AI model call, the temperature parameter. Temperature essentially means how random the output will be. Example: 0.5 """ AI_TEXTS = "ai.texts" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Raw text inputs provided to the model. Example: ["What is machine learning?"] """ AI_TOP_K = "ai.top_k" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_TOP_K instead. + For an AI model call, the top_k parameter. Top_k essentially controls how random the output will be. Example: 35 """ AI_TOP_P = "ai.top_p" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_TOP_P instead. + For an AI model call, the top_p parameter. Top_p essentially controls how random the output will be. Example: 0.5 """ AI_TOOL_CALLS = "ai.tool_calls" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_RESPONSE_TOOL_CALLS instead. + For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls """ AI_TOOLS = "ai.tools" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_REQUEST_AVAILABLE_TOOLS instead. + For an AI model call, the functions that are available """ AI_WARNINGS = "ai.warnings" """ + .. deprecated:: + This attribute is deprecated. Use GEN_AI_* attributes instead. + Warning messages generated during model execution. Example: ["Token limit exceeded"] """ @@ -383,6 +468,18 @@ class SPANDATA: Example: "qa-pipeline" """ + GEN_AI_RESPONSE_FINISH_REASONS = "gen_ai.response.finish_reasons" + """ + The reason why the model stopped generating. + Example: "COMPLETE" + """ + + GEN_AI_RESPONSE_ID = "gen_ai.response.id" + """ + Unique identifier for the completion. + Example: "gen_123abc" + """ + GEN_AI_RESPONSE_MODEL = "gen_ai.response.model" """ Exact model identifier used to generate the response @@ -443,12 +540,24 @@ class SPANDATA: Example: 0.1 """ + GEN_AI_REQUEST_SEED = "gen_ai.request.seed" + """ + The seed, ideally models given the same seed and same other parameters will produce the exact same output. + Example: "1234567890" + """ + GEN_AI_REQUEST_TEMPERATURE = "gen_ai.request.temperature" """ The temperature parameter used to control randomness in the output. Example: 0.7 """ + GEN_AI_REQUEST_TOP_K = "gen_ai.request.top_k" + """ + Limits the model to only consider the K most likely next tokens, where K is an integer (e.g., top_k=20 means only the 20 highest probability tokens are considered). + Example: 35 + """ + GEN_AI_REQUEST_TOP_P = "gen_ai.request.top_p" """ The top_p parameter used to control diversity via nucleus sampling. From 19914cd5f0013aa8d50f5ad8c01fac8c16d702f4 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 5 Aug 2025 12:04:54 +0200 Subject: [PATCH 395/591] feat(tracing): Add convenience function `update_current_span`. (#4673) Manually setting data on spans for our insights modules can be made easier. This PR introduces a convenience function. Right now users need to do some like this: ```python import sentry_sdk span = sentry_sdk.get_current_span() if span is not None: span.description = f"Some {dynamic} name" span.set_attribute("key1", "value1") span.set_attribute("key2", "value2") ``` With this new convenience function a user can do: ```python import sentry_sdk sentry_sdk.update_current_span( name=f"Some {dynamic} name", attributes={ "key1": "value1", "key2": "value2", }, ) --- docs/api.rst | 1 + sentry_sdk/__init__.py | 1 + sentry_sdk/api.py | 80 ++++++++++++++++++++++++++++++++++++++ tests/tracing/test_misc.py | 45 +++++++++++++++++++++ 4 files changed, 127 insertions(+) diff --git a/docs/api.rst b/docs/api.rst index a6fb49346d..7d59030033 100644 --- a/docs/api.rst +++ b/docs/api.rst @@ -41,6 +41,7 @@ Performance Monitoring .. autofunction:: sentry_sdk.api.get_current_span .. autofunction:: sentry_sdk.api.start_span .. autofunction:: sentry_sdk.api.start_transaction +.. autofunction:: sentry_sdk.api.update_current_span Distributed Tracing diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index 7b1eda172a..a37b52ff4e 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -50,6 +50,7 @@ "start_session", "end_session", "set_transaction_name", + "update_current_span", ] # Initialize the debug support after everything is loaded diff --git a/sentry_sdk/api.py b/sentry_sdk/api.py index a4fb95e9a1..43758b4d78 100644 --- a/sentry_sdk/api.py +++ b/sentry_sdk/api.py @@ -85,6 +85,7 @@ def overload(x): "start_session", "end_session", "set_transaction_name", + "update_current_span", ] @@ -473,3 +474,82 @@ def end_session(): def set_transaction_name(name, source=None): # type: (str, Optional[str]) -> None return get_current_scope().set_transaction_name(name, source) + + +def update_current_span(op=None, name=None, attributes=None, data=None): + # type: (Optional[str], Optional[str], Optional[dict[str, Union[str, int, float, bool]]], Optional[dict[str, Any]]) -> None + """ + Update the current active span with the provided parameters. + + This function allows you to modify properties of the currently active span. + If no span is currently active, this function will do nothing. + + :param op: The operation name for the span. This is a high-level description + of what the span represents (e.g., "http.client", "db.query"). + You can use predefined constants from :py:class:`sentry_sdk.consts.OP` + or provide your own string. If not provided, the span's operation will + remain unchanged. + :type op: str or None + + :param name: The human-readable name/description for the span. This provides + more specific details about what the span represents (e.g., "GET /api/users", + "SELECT * FROM users"). If not provided, the span's name will remain unchanged. + :type name: str or None + + :param data: A dictionary of key-value pairs to add as data to the span. This + data will be merged with any existing span data. If not provided, + no data will be added. + + .. deprecated:: 2.35.0 + Use ``attributes`` instead. The ``data`` parameter will be removed + in a future version. + :type data: dict[str, Union[str, int, float, bool]] or None + + :param attributes: A dictionary of key-value pairs to add as attributes to the span. + Attribute values must be strings, integers, floats, or booleans. These + attributes will be merged with any existing span data. If not provided, + no attributes will be added. + :type attributes: dict[str, Union[str, int, float, bool]] or None + + :returns: None + + .. versionadded:: 2.35.0 + + Example:: + + import sentry_sdk + from sentry_sdk.consts import OP + + sentry_sdk.update_current_span( + op=OP.FUNCTION, + name="process_user_data", + attributes={"user_id": 123, "batch_size": 50} + ) + """ + current_span = get_current_span() + + if current_span is None: + return + + if op is not None: + current_span.op = op + + if name is not None: + # internally it is still description + current_span.description = name + + if data is not None and attributes is not None: + raise ValueError( + "Cannot provide both `data` and `attributes`. Please use only `attributes`." + ) + + if data is not None: + warnings.warn( + "The `data` parameter is deprecated. Please use `attributes` instead.", + DeprecationWarning, + stacklevel=2, + ) + attributes = data + + if attributes is not None: + current_span.update_data(attributes) diff --git a/tests/tracing/test_misc.py b/tests/tracing/test_misc.py index 651228b45e..e1de847102 100644 --- a/tests/tracing/test_misc.py +++ b/tests/tracing/test_misc.py @@ -540,3 +540,48 @@ def test_span_set_data_update_data(sentry_init, capture_events): "thread.id": mock.ANY, "thread.name": mock.ANY, } + + +def test_update_current_span(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + + events = capture_events() + + with sentry_sdk.start_transaction(name="test-transaction"): + with start_span(op="test-span-op", name="test-span-name"): + sentry_sdk.update_current_span( + op="updated-span-op", + name="updated-span-name", + attributes={ + "key0": "value0", + "key1": "value1", + }, + ) + + sentry_sdk.update_current_span( + op="updated-span-op-2", + ) + + sentry_sdk.update_current_span( + name="updated-span-name-3", + ) + + sentry_sdk.update_current_span( + attributes={ + "key1": "updated-value-4", + "key2": "value2", + }, + ) + + (event,) = events + span = event["spans"][0] + + assert span["op"] == "updated-span-op-2" + assert span["description"] == "updated-span-name-3" + assert span["data"] == { + "key0": "value0", + "key1": "updated-value-4", + "key2": "value2", + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } From 7e5d40193afeb51ed0096e91ad5ba2f8b4f6291d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 7 Aug 2025 10:20:55 +0200 Subject: [PATCH 396/591] feat(tracing): Improve `@trace` decorator. (#4648) Update the `@trace` decorator and make it more powerful. It accepts now the following parameters: `op`, `name`, `attributes`. Example usage: ```python import sentry_sdk from sentry_sdk.consts import OP # Simple usage (like before) @sentry_sdk.trace def process_data(): # Function implementation pass # With custom parameters @sentry_sdk.trace( op=OP.DB_QUERY, name="Get user data", attributes={"postgres": True} ) def make_db_query(sql): # Function implementation pass ``` This creates better DX for our users. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- docs/api.rst | 1 + sentry_sdk/integrations/starlite.py | 2 +- sentry_sdk/tracing.py | 78 +++++++++++++++++++------ sentry_sdk/tracing_utils.py | 90 +++++++++++++++++------------ tests/tracing/test_decorator.py | 6 +- 5 files changed, 120 insertions(+), 57 deletions(-) diff --git a/docs/api.rst b/docs/api.rst index 7d59030033..802abee75d 100644 --- a/docs/api.rst +++ b/docs/api.rst @@ -37,6 +37,7 @@ Enriching Events Performance Monitoring ====================== +.. autofunction:: sentry_sdk.api.trace .. autofunction:: sentry_sdk.api.continue_trace .. autofunction:: sentry_sdk.api.get_current_span .. autofunction:: sentry_sdk.api.start_span diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index 24707a18b1..6ab80712e5 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -17,7 +17,7 @@ from starlite.plugins.base import get_plugin_for_value # type: ignore from starlite.routes.http import HTTPRoute # type: ignore from starlite.utils import ConnectionDataExtractor, is_async_callable, Ref # type: ignore - from pydantic import BaseModel # type: ignore + from pydantic import BaseModel except ImportError: raise DidNotEnable("Starlite is not installed") diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index dd1392d150..e9d726cc66 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1340,43 +1340,85 @@ def _set_initial_sampling_decision(self, sampling_context): if TYPE_CHECKING: @overload - def trace(func=None): - # type: (None) -> Callable[[Callable[P, R]], Callable[P, R]] + def trace(func=None, *, op=None, name=None, attributes=None): + # type: (None, Optional[str], Optional[str], Optional[dict[str, Any]]) -> Callable[[Callable[P, R]], Callable[P, R]] + # Handles: @trace() and @trace(op="custom") pass @overload def trace(func): # type: (Callable[P, R]) -> Callable[P, R] + # Handles: @trace pass -def trace(func=None): - # type: (Optional[Callable[P, R]]) -> Union[Callable[P, R], Callable[[Callable[P, R]], Callable[P, R]]] +def trace(func=None, *, op=None, name=None, attributes=None): + # type: (Optional[Callable[P, R]], Optional[str], Optional[str], Optional[dict[str, Any]]) -> Union[Callable[P, R], Callable[[Callable[P, R]], Callable[P, R]]] """ - Decorator to start a child span under the existing current transaction. - If there is no current transaction, then nothing will be traced. + Decorator to start a child span around a function call. - .. code-block:: - :caption: Usage + This decorator automatically creates a new span when the decorated function + is called, and finishes the span when the function returns or raises an exception. + + :param func: The function to trace. When used as a decorator without parentheses, + this is the function being decorated. When used with parameters (e.g., + ``@trace(op="custom")``, this should be None. + :type func: Callable or None + + :param op: The operation name for the span. This is a high-level description + of what the span represents (e.g., "http.client", "db.query"). + You can use predefined constants from :py:class:`sentry_sdk.consts.OP` + or provide your own string. If not provided, a default operation will + be assigned based on the template. + :type op: str or None + + :param name: The human-readable name/description for the span. If not provided, + defaults to the function name. This provides more specific details about + what the span represents (e.g., "GET /api/users", "process_user_data"). + :type name: str or None + + :param attributes: A dictionary of key-value pairs to add as attributes to the span. + Attribute values must be strings, integers, floats, or booleans. These + attributes provide additional context about the span's execution. + :type attributes: dict[str, Any] or None + + :returns: When used as ``@trace``, returns the decorated function. When used as + ``@trace(...)`` with parameters, returns a decorator function. + :rtype: Callable or decorator function + + Example:: import sentry_sdk + from sentry_sdk.consts import OP + # Simple usage with default values @sentry_sdk.trace - def my_function(): - ... + def process_data(): + # Function implementation + pass - @sentry_sdk.trace - async def my_async_function(): - ... + # With custom parameters + @sentry_sdk.trace( + op=OP.DB_QUERY, + name="Get user data", + attributes={"postgres": True} + ) + def make_db_query(sql): + # Function implementation + pass """ - from sentry_sdk.tracing_utils import start_child_span_decorator + from sentry_sdk.tracing_utils import create_span_decorator + + decorator = create_span_decorator( + op=op, + name=name, + attributes=attributes, + ) - # This patterns allows usage of both @sentry_traced and @sentry_traced(...) - # See https://stackoverflow.com/questions/52126071/decorator-with-arguments-avoid-parenthesis-when-no-arguments/52126278 if func: - return start_child_span_decorator(func) + return decorator(func) else: - return start_child_span_decorator + return decorator # Circular imports diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 552f4fd59a..447a708d4d 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -1,4 +1,5 @@ import contextlib +import functools import inspect import os import re @@ -6,7 +7,6 @@ from collections.abc import Mapping from datetime import timedelta from decimal import ROUND_DOWN, Decimal, DefaultContext, localcontext -from functools import wraps from random import Random from urllib.parse import quote, unquote import uuid @@ -770,70 +770,86 @@ def normalize_incoming_data(incoming_data): return data -def start_child_span_decorator(func): - # type: (Any) -> Any +def create_span_decorator(op=None, name=None, attributes=None): + # type: (Optional[str], Optional[str], Optional[dict[str, Any]]) -> Any """ - Decorator to add child spans for functions. + Create a span decorator that can wrap both sync and async functions. - See also ``sentry_sdk.tracing.trace()``. + :param op: The operation type for the span. + :param name: The name of the span. + :param attributes: Additional attributes to set on the span. """ - # Asynchronous case - if inspect.iscoroutinefunction(func): - @wraps(func) - async def func_with_tracing(*args, **kwargs): - # type: (*Any, **Any) -> Any + def span_decorator(f): + # type: (Any) -> Any + """ + Decorator to create a span for the given function. + """ - span = get_current_span() + @functools.wraps(f) + async def async_wrapper(*args, **kwargs): + # type: (*Any, **Any) -> Any + current_span = get_current_span() - if span is None: + if current_span is None: logger.debug( "Cannot create a child span for %s. " "Please start a Sentry transaction before calling this function.", - qualname_from_function(func), + qualname_from_function(f), ) - return await func(*args, **kwargs) + return await f(*args, **kwargs) + + span_op = op or OP.FUNCTION + span_name = name or qualname_from_function(f) or "" - with span.start_child( - op=OP.FUNCTION, - name=qualname_from_function(func), - ): - return await func(*args, **kwargs) + with current_span.start_child( + op=span_op, + name=span_name, + ) as span: + span.update_data(attributes or {}) + result = await f(*args, **kwargs) + return result try: - func_with_tracing.__signature__ = inspect.signature(func) # type: ignore[attr-defined] + async_wrapper.__signature__ = inspect.signature(f) # type: ignore[attr-defined] except Exception: pass - # Synchronous case - else: - - @wraps(func) - def func_with_tracing(*args, **kwargs): + @functools.wraps(f) + def sync_wrapper(*args, **kwargs): # type: (*Any, **Any) -> Any + current_span = get_current_span() - span = get_current_span() - - if span is None: + if current_span is None: logger.debug( "Cannot create a child span for %s. " "Please start a Sentry transaction before calling this function.", - qualname_from_function(func), + qualname_from_function(f), ) - return func(*args, **kwargs) + return f(*args, **kwargs) + + span_op = op or OP.FUNCTION + span_name = name or qualname_from_function(f) or "" - with span.start_child( - op=OP.FUNCTION, - name=qualname_from_function(func), - ): - return func(*args, **kwargs) + with current_span.start_child( + op=span_op, + name=span_name, + ) as span: + span.update_data(attributes or {}) + result = f(*args, **kwargs) + return result try: - func_with_tracing.__signature__ = inspect.signature(func) # type: ignore[attr-defined] + sync_wrapper.__signature__ = inspect.signature(f) # type: ignore[attr-defined] except Exception: pass - return func_with_tracing + if inspect.iscoroutinefunction(f): + return async_wrapper + else: + return sync_wrapper + + return span_decorator def get_current_span(scope=None): diff --git a/tests/tracing/test_decorator.py b/tests/tracing/test_decorator.py index 18a66bd43e..9a7074c470 100644 --- a/tests/tracing/test_decorator.py +++ b/tests/tracing/test_decorator.py @@ -4,7 +4,7 @@ import pytest from sentry_sdk.tracing import trace -from sentry_sdk.tracing_utils import start_child_span_decorator +from sentry_sdk.tracing_utils import create_span_decorator from sentry_sdk.utils import logger from tests.conftest import patch_start_tracing_child @@ -24,6 +24,7 @@ def test_trace_decorator(): fake_start_child.assert_not_called() assert result == "return_of_sync_function" + start_child_span_decorator = create_span_decorator() result2 = start_child_span_decorator(my_example_function)() fake_start_child.assert_called_once_with( op="function", name="test_decorator.my_example_function" @@ -38,6 +39,7 @@ def test_trace_decorator_no_trx(): fake_debug.assert_not_called() assert result == "return_of_sync_function" + start_child_span_decorator = create_span_decorator() result2 = start_child_span_decorator(my_example_function)() fake_debug.assert_called_once_with( "Cannot create a child span for %s. " @@ -55,6 +57,7 @@ async def test_trace_decorator_async(): fake_start_child.assert_not_called() assert result == "return_of_async_function" + start_child_span_decorator = create_span_decorator() result2 = await start_child_span_decorator(my_async_example_function)() fake_start_child.assert_called_once_with( op="function", @@ -71,6 +74,7 @@ async def test_trace_decorator_async_no_trx(): fake_debug.assert_not_called() assert result == "return_of_async_function" + start_child_span_decorator = create_span_decorator() result2 = await start_child_span_decorator(my_async_example_function)() fake_debug.assert_called_once_with( "Cannot create a child span for %s. " From b73f8763e75a3118a235a4d1da51358d09d881ee Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 7 Aug 2025 17:37:01 +0200 Subject: [PATCH 397/591] ref(clickhouse): List `send_data` parameters (#4667) Explicitly list the `send_data` parameters in the wrapped function. The parameters are coming from [here](https://github.com/mymarilyn/clickhouse-driver/blob/8a4e7c5b99b532df2b015651d893a6f36288a22c/clickhouse_driver/client.py#L634). Continue also providing `*args` and `**kwargs`, but only for forwards-compatibility. --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- sentry_sdk/integrations/clickhouse_driver.py | 24 +++++++++++--------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/sentry_sdk/integrations/clickhouse_driver.py b/sentry_sdk/integrations/clickhouse_driver.py index 2561bfad04..7a977bc278 100644 --- a/sentry_sdk/integrations/clickhouse_driver.py +++ b/sentry_sdk/integrations/clickhouse_driver.py @@ -49,9 +49,7 @@ def setup_once() -> None: ) # If the query contains parameters then the send_data function is used to send those parameters to clickhouse - clickhouse_driver.client.Client.send_data = _wrap_send_data( - clickhouse_driver.client.Client.send_data - ) + _wrap_send_data() # Every query ends either with the Client's `receive_end_of_query` (no result expected) # or its `receive_result` (result expected) @@ -128,23 +126,27 @@ def _inner_end(*args: P.args, **kwargs: P.kwargs) -> T: return _inner_end -def _wrap_send_data(f: Callable[P, T]) -> Callable[P, T]: - def _inner_send_data(*args: P.args, **kwargs: P.kwargs) -> T: - instance = args[0] # type: clickhouse_driver.client.Client - data = args[2] - span = getattr(instance.connection, "_sentry_span", None) +def _wrap_send_data() -> None: + original_send_data = clickhouse_driver.client.Client.send_data + + def _inner_send_data( # type: ignore[no-untyped-def] # clickhouse-driver does not type send_data + self, sample_block, data, types_check=False, columnar=False, *args, **kwargs + ): + span = getattr(self.connection, "_sentry_span", None) if span is not None: - _set_db_data(span, instance.connection) + _set_db_data(span, self.connection) if should_send_default_pii(): db_params = span._data.get("db.params", []) db_params.extend(data) span.set_data("db.params", db_params) - return f(*args, **kwargs) + return original_send_data( + self, sample_block, data, types_check, columnar, *args, **kwargs + ) - return _inner_send_data + clickhouse_driver.client.Client.send_data = _inner_send_data def _set_db_data( From 3ef02a11544c238ae8d2c7399f6b2d9c16419aa2 Mon Sep 17 00:00:00 2001 From: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Date: Thu, 7 Aug 2025 17:37:12 +0200 Subject: [PATCH 398/591] fix(clickhouse): Don't eat the generator data (#4669) Currently, the Clickhouse integration consumes any data passed as a generator when reading it for insertion as `db_params`. Instead, since generators cannot be cloned, we need to wrap the generator to add the params as we iterate over it. Fixes #4657 --- Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. Running the test suite on your PR might require maintainer approval. --- sentry_sdk/integrations/clickhouse_driver.py | 22 +++++++++++-- .../test_clickhouse_driver.py | 32 +++++++++++++++++++ 2 files changed, 52 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/clickhouse_driver.py b/sentry_sdk/integrations/clickhouse_driver.py index 7a977bc278..bbaaaeec8e 100644 --- a/sentry_sdk/integrations/clickhouse_driver.py +++ b/sentry_sdk/integrations/clickhouse_driver.py @@ -11,7 +11,8 @@ # without introducing a hard dependency on `typing_extensions` # from: https://stackoverflow.com/a/71944042/300572 if TYPE_CHECKING: - from typing import ParamSpec, Callable + from collections.abc import Iterator + from typing import Any, ParamSpec, Callable else: # Fake ParamSpec class ParamSpec: @@ -139,7 +140,24 @@ def _inner_send_data( # type: ignore[no-untyped-def] # clickhouse-driver does n if should_send_default_pii(): db_params = span._data.get("db.params", []) - db_params.extend(data) + + if isinstance(data, (list, tuple)): + db_params.extend(data) + + else: # data is a generic iterator + orig_data = data + + # Wrap the generator to add items to db.params as they are yielded. + # This allows us to send the params to Sentry without needing to allocate + # memory for the entire generator at once. + def wrapped_generator() -> "Iterator[Any]": + for item in orig_data: + db_params.append(item) + yield item + + # Replace the original iterator with the wrapped one. + data = wrapped_generator() + span.set_data("db.params", db_params) return original_send_data( diff --git a/tests/integrations/clickhouse_driver/test_clickhouse_driver.py b/tests/integrations/clickhouse_driver/test_clickhouse_driver.py index 0675ad9ff5..635f9334c4 100644 --- a/tests/integrations/clickhouse_driver/test_clickhouse_driver.py +++ b/tests/integrations/clickhouse_driver/test_clickhouse_driver.py @@ -342,6 +342,38 @@ def test_clickhouse_client_spans( assert event["spans"] == expected_spans +def test_clickhouse_spans_with_generator(sentry_init, capture_events): + sentry_init( + integrations=[ClickhouseDriverIntegration()], + send_default_pii=True, + traces_sample_rate=1.0, + ) + events = capture_events() + + # Use a generator to test that the integration obtains values from the generator, + # without consuming the generator. + values = ({"x": i} for i in range(3)) + + with start_transaction(name="test_clickhouse_transaction"): + client = Client("localhost") + client.execute("DROP TABLE IF EXISTS test") + client.execute("CREATE TABLE test (x Int32) ENGINE = Memory") + client.execute("INSERT INTO test (x) VALUES", values) + res = client.execute("SELECT x FROM test") + + # Verify that the integration did not consume the generator + assert res == [(0,), (1,), (2,)] + + (event,) = events + spans = event["spans"] + + [span] = [ + span for span in spans if span["description"] == "INSERT INTO test (x) VALUES" + ] + + assert span["data"]["db.params"] == [{"x": 0}, {"x": 1}, {"x": 2}] + + def test_clickhouse_client_spans_with_pii( sentry_init, capture_events, capture_envelopes ) -> None: From 378fe812127e86590abbcbeb4e8fe7a4f258a31b Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Fri, 8 Aug 2025 13:03:03 +0200 Subject: [PATCH 399/591] feat(anthropic) Update span attributes to use `gen_ai.*` namespace instead of `ai.*` (#4674) Update `AnthropicIntegration` to support Otel and Sentry AI Agents module compatible span attributes of `gen_ai.*` family. Closes https://linear.app/getsentry/issue/TET-996/improve-integration-for-anthropic-sdk --------- Co-authored-by: Anton Pirker --- sentry_sdk/ai/utils.py | 18 +- sentry_sdk/integrations/anthropic.py | 213 ++++++++++++------ sentry_sdk/integrations/starlite.py | 2 +- .../integrations/anthropic/test_anthropic.py | 212 +++++++++-------- 4 files changed, 267 insertions(+), 178 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index a3c62600c0..cf52cba6e8 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -7,8 +7,8 @@ from sentry_sdk.utils import logger -def _normalize_data(data): - # type: (Any) -> Any +def _normalize_data(data, unpack=True): + # type: (Any, bool) -> Any # convert pydantic data (e.g. OpenAI v1+) to json compatible format if hasattr(data, "model_dump"): @@ -18,18 +18,18 @@ def _normalize_data(data): logger.warning("Could not convert pydantic data to JSON: %s", e) return data if isinstance(data, list): - if len(data) == 1: - return _normalize_data(data[0]) # remove empty dimensions - return list(_normalize_data(x) for x in data) + if unpack and len(data) == 1: + return _normalize_data(data[0], unpack=unpack) # remove empty dimensions + return list(_normalize_data(x, unpack=unpack) for x in data) if isinstance(data, dict): - return {k: _normalize_data(v) for (k, v) in data.items()} + return {k: _normalize_data(v, unpack=unpack) for (k, v) in data.items()} return data -def set_data_normalized(span, key, value): - # type: (Span, str, Any) -> None - normalized = _normalize_data(value) +def set_data_normalized(span, key, value, unpack=True): + # type: (Span, str, Any, bool) -> None + normalized = _normalize_data(value, unpack=unpack) if isinstance(normalized, (int, float, bool, str)): span.set_data(key, normalized) else: diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 1e1f9112a1..05d45ef62f 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -1,8 +1,10 @@ from functools import wraps +import json from typing import TYPE_CHECKING import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage +from sentry_sdk.ai.utils import set_data_normalized from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -10,9 +12,15 @@ capture_internal_exceptions, event_from_exception, package_version, + safe_serialize, ) try: + try: + from anthropic import NOT_GIVEN + except ImportError: + NOT_GIVEN = None + from anthropic.resources import AsyncMessages, Messages if TYPE_CHECKING: @@ -53,8 +61,11 @@ def _capture_exception(exc): sentry_sdk.capture_event(event, hint=hint) -def _calculate_token_usage(result, span): - # type: (Messages, Span) -> None +def _get_token_usage(result): + # type: (Messages) -> tuple[int, int] + """ + Get token usage from the Anthropic response. + """ input_tokens = 0 output_tokens = 0 if hasattr(result, "usage"): @@ -64,37 +75,13 @@ def _calculate_token_usage(result, span): if hasattr(usage, "output_tokens") and isinstance(usage.output_tokens, int): output_tokens = usage.output_tokens - total_tokens = input_tokens + output_tokens + return input_tokens, output_tokens - record_token_usage( - span, - input_tokens=input_tokens, - output_tokens=output_tokens, - total_tokens=total_tokens, - ) - -def _get_responses(content): - # type: (list[Any]) -> list[dict[str, Any]] +def _collect_ai_data(event, model, input_tokens, output_tokens, content_blocks): + # type: (MessageStreamEvent, str | None, int, int, list[str]) -> tuple[str | None, int, int, list[str]] """ - Get JSON of a Anthropic responses. - """ - responses = [] - for item in content: - if hasattr(item, "text"): - responses.append( - { - "type": item.type, - "text": item.text, - } - ) - return responses - - -def _collect_ai_data(event, input_tokens, output_tokens, content_blocks): - # type: (MessageStreamEvent, int, int, list[str]) -> tuple[int, int, list[str]] - """ - Count token usage and collect content blocks from the AI streaming response. + Collect model information, token usage, and collect content blocks from the AI streaming response. """ with capture_internal_exceptions(): if hasattr(event, "type"): @@ -102,6 +89,7 @@ def _collect_ai_data(event, input_tokens, output_tokens, content_blocks): usage = event.message.usage input_tokens += usage.input_tokens output_tokens += usage.output_tokens + model = event.message.model or model elif event.type == "content_block_start": pass elif event.type == "content_block_delta": @@ -114,31 +102,80 @@ def _collect_ai_data(event, input_tokens, output_tokens, content_blocks): elif event.type == "message_delta": output_tokens += event.usage.output_tokens - return input_tokens, output_tokens, content_blocks + return model, input_tokens, output_tokens, content_blocks -def _add_ai_data_to_span( - span, integration, input_tokens, output_tokens, content_blocks -): - # type: (Span, AnthropicIntegration, int, int, list[str]) -> None +def _set_input_data(span, kwargs, integration): + # type: (Span, dict[str, Any], AnthropicIntegration) -> None """ - Add token usage and content blocks from the AI streaming response to the span. + Set input data for the span based on the provided keyword arguments for the anthropic message creation. """ - with capture_internal_exceptions(): - if should_send_default_pii() and integration.include_prompts: - complete_message = "".join(content_blocks) - span.set_data( - SPANDATA.AI_RESPONSES, - [{"type": "text", "text": complete_message}], - ) - total_tokens = input_tokens + output_tokens - record_token_usage( + messages = kwargs.get("messages") + if ( + messages is not None + and len(messages) > 0 + and should_send_default_pii() + and integration.include_prompts + ): + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, safe_serialize(messages) + ) + + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_STREAMING, kwargs.get("stream", False) + ) + + kwargs_keys_to_attributes = { + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "model": SPANDATA.GEN_AI_REQUEST_MODEL, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "top_k": SPANDATA.GEN_AI_REQUEST_TOP_K, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, + } + for key, attribute in kwargs_keys_to_attributes.items(): + value = kwargs.get(key) + if value is not NOT_GIVEN and value is not None: + set_data_normalized(span, attribute, value) + + # Input attributes: Tools + tools = kwargs.get("tools") + if tools is not NOT_GIVEN and tools is not None and len(tools) > 0: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) + ) + + +def _set_output_data( + span, + integration, + model, + input_tokens, + output_tokens, + content_blocks, + finish_span=False, +): + # type: (Span, AnthropicIntegration, str | None, int | None, int | None, list[Any], bool) -> None + """ + Set output data for the span based on the AI response.""" + span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, model) + if should_send_default_pii() and integration.include_prompts: + set_data_normalized( span, - input_tokens=input_tokens, - output_tokens=output_tokens, - total_tokens=total_tokens, + SPANDATA.GEN_AI_RESPONSE_TEXT, + json.dumps(content_blocks), + unpack=False, ) - span.set_data(SPANDATA.AI_STREAMING, True) + + record_token_usage( + span, + input_tokens=input_tokens, + output_tokens=output_tokens, + ) + + # TODO: GEN_AI_RESPONSE_TOOL_CALLS ? + + if finish_span: + span.__exit__(None, None, None) def _sentry_patched_create_common(f, *args, **kwargs): @@ -155,31 +192,41 @@ def _sentry_patched_create_common(f, *args, **kwargs): except TypeError: return f(*args, **kwargs) + model = kwargs.get("model", "") + span = sentry_sdk.start_span( - op=OP.ANTHROPIC_MESSAGES_CREATE, - description="Anthropic messages create", + op=OP.GEN_AI_CHAT, + name=f"chat {model}".strip(), origin=AnthropicIntegration.origin, ) span.__enter__() - result = yield f, args, kwargs + _set_input_data(span, kwargs, integration) - # add data to span and finish it - messages = list(kwargs["messages"]) - model = kwargs.get("model") + result = yield f, args, kwargs with capture_internal_exceptions(): - span.set_data(SPANDATA.AI_MODEL_ID, model) - span.set_data(SPANDATA.AI_STREAMING, False) - - if should_send_default_pii() and integration.include_prompts: - span.set_data(SPANDATA.AI_INPUT_MESSAGES, messages) - if hasattr(result, "content"): - if should_send_default_pii() and integration.include_prompts: - span.set_data(SPANDATA.AI_RESPONSES, _get_responses(result.content)) - _calculate_token_usage(result, span) - span.__exit__(None, None, None) + input_tokens, output_tokens = _get_token_usage(result) + + content_blocks = [] + for content_block in result.content: + if hasattr(content_block, "to_dict"): + content_blocks.append(content_block.to_dict()) + elif hasattr(content_block, "model_dump"): + content_blocks.append(content_block.model_dump()) + elif hasattr(content_block, "text"): + content_blocks.append({"type": "text", "text": content_block.text}) + + _set_output_data( + span=span, + integration=integration, + model=getattr(result, "model", None), + input_tokens=input_tokens, + output_tokens=output_tokens, + content_blocks=content_blocks, + finish_span=True, + ) # Streaming response elif hasattr(result, "_iterator"): @@ -187,37 +234,53 @@ def _sentry_patched_create_common(f, *args, **kwargs): def new_iterator(): # type: () -> Iterator[MessageStreamEvent] + model = None input_tokens = 0 output_tokens = 0 content_blocks = [] # type: list[str] for event in old_iterator: - input_tokens, output_tokens, content_blocks = _collect_ai_data( - event, input_tokens, output_tokens, content_blocks + model, input_tokens, output_tokens, content_blocks = ( + _collect_ai_data( + event, model, input_tokens, output_tokens, content_blocks + ) ) yield event - _add_ai_data_to_span( - span, integration, input_tokens, output_tokens, content_blocks + _set_output_data( + span=span, + integration=integration, + model=model, + input_tokens=input_tokens, + output_tokens=output_tokens, + content_blocks=[{"text": "".join(content_blocks), "type": "text"}], + finish_span=True, ) - span.__exit__(None, None, None) async def new_iterator_async(): # type: () -> AsyncIterator[MessageStreamEvent] + model = None input_tokens = 0 output_tokens = 0 content_blocks = [] # type: list[str] async for event in old_iterator: - input_tokens, output_tokens, content_blocks = _collect_ai_data( - event, input_tokens, output_tokens, content_blocks + model, input_tokens, output_tokens, content_blocks = ( + _collect_ai_data( + event, model, input_tokens, output_tokens, content_blocks + ) ) yield event - _add_ai_data_to_span( - span, integration, input_tokens, output_tokens, content_blocks + _set_output_data( + span=span, + integration=integration, + model=model, + input_tokens=input_tokens, + output_tokens=output_tokens, + content_blocks=[{"text": "".join(content_blocks), "type": "text"}], + finish_span=True, ) - span.__exit__(None, None, None) if str(type(result._iterator)) == "": result._iterator = new_iterator_async() diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index 6ab80712e5..24707a18b1 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -17,7 +17,7 @@ from starlite.plugins.base import get_plugin_for_value # type: ignore from starlite.routes.http import HTTPRoute # type: ignore from starlite.utils import ConnectionDataExtractor, is_async_callable, Ref # type: ignore - from pydantic import BaseModel + from pydantic import BaseModel # type: ignore except ImportError: raise DidNotEnable("Starlite is not installed") diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index e6e1a40aa9..eba07a1df6 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -20,7 +20,7 @@ async def __call__(self, *args, **kwargs): from anthropic.types.message_delta_event import MessageDeltaEvent from anthropic.types.message_start_event import MessageStartEvent -from sentry_sdk.integrations.anthropic import _add_ai_data_to_span, _collect_ai_data +from sentry_sdk.integrations.anthropic import _set_output_data, _collect_ai_data from sentry_sdk.utils import package_version try: @@ -112,23 +112,27 @@ def test_nonstreaming_create_message( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"type": "text", "text": "Hi, I'm Claude."} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "Hello, Claude"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "Hi, I\'m Claude.", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 10 - assert span["data"]["gen_ai.usage.output_tokens"] == 20 - assert span["data"]["gen_ai.usage.total_tokens"] == 30 - assert span["data"][SPANDATA.AI_STREAMING] is False + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 20 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 30 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is False @pytest.mark.asyncio @@ -180,23 +184,27 @@ async def test_nonstreaming_create_message_async( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"type": "text", "text": "Hi, I'm Claude."} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "Hello, Claude"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "Hi, I\'m Claude.", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 10 - assert span["data"]["gen_ai.usage.output_tokens"] == 20 - assert span["data"]["gen_ai.usage.total_tokens"] == 30 - assert span["data"][SPANDATA.AI_STREAMING] is False + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 20 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 30 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is False @pytest.mark.parametrize( @@ -279,24 +287,28 @@ def test_streaming_create_message( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"type": "text", "text": "Hi! I'm Claude!"} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "Hello, Claude"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "Hi! I\'m Claude!", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 10 - assert span["data"]["gen_ai.usage.output_tokens"] == 30 - assert span["data"]["gen_ai.usage.total_tokens"] == 40 - assert span["data"][SPANDATA.AI_STREAMING] is True + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 30 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 40 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True @pytest.mark.asyncio @@ -382,24 +394,28 @@ async def test_streaming_create_message_async( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"type": "text", "text": "Hi! I'm Claude!"} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "Hello, Claude"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "Hi! I\'m Claude!", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 10 - assert span["data"]["gen_ai.usage.output_tokens"] == 30 - assert span["data"]["gen_ai.usage.total_tokens"] == 40 - assert span["data"][SPANDATA.AI_STREAMING] is True + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 30 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 40 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True @pytest.mark.skipif( @@ -512,23 +528,27 @@ def test_streaming_create_message_with_input_json_delta( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"text": "{'location': 'San Francisco, CA'}", "type": "text"} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "What is the weather like in San Francisco?"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "{\'location\': \'San Francisco, CA\'}", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 366 - assert span["data"]["gen_ai.usage.output_tokens"] == 51 - assert span["data"]["gen_ai.usage.total_tokens"] == 417 - assert span["data"][SPANDATA.AI_STREAMING] is True + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 366 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 51 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 417 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True @pytest.mark.asyncio @@ -648,24 +668,28 @@ async def test_streaming_create_message_with_input_json_delta_async( assert len(event["spans"]) == 1 (span,) = event["spans"] - assert span["op"] == OP.ANTHROPIC_MESSAGES_CREATE - assert span["description"] == "Anthropic messages create" - assert span["data"][SPANDATA.AI_MODEL_ID] == "model" + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat model" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "model" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.AI_INPUT_MESSAGES] == messages - assert span["data"][SPANDATA.AI_RESPONSES] == [ - {"text": "{'location': 'San Francisco, CA'}", "type": "text"} - ] + assert ( + span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + == '[{"role": "user", "content": "What is the weather like in San Francisco?"}]' + ) + assert ( + span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + == '[{"text": "{\'location\': \'San Francisco, CA\'}", "type": "text"}]' + ) else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] - assert span["data"]["gen_ai.usage.input_tokens"] == 366 - assert span["data"]["gen_ai.usage.output_tokens"] == 51 - assert span["data"]["gen_ai.usage.total_tokens"] == 417 - assert span["data"][SPANDATA.AI_STREAMING] is True + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 366 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 51 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 417 + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True def test_exception_message_create(sentry_init, capture_events): @@ -770,15 +794,16 @@ def test_collect_ai_data_with_input_json_delta(): index=0, type="content_block_delta", ) - + model = None input_tokens = 10 output_tokens = 20 content_blocks = [] - new_input_tokens, new_output_tokens, new_content_blocks = _collect_ai_data( - event, input_tokens, output_tokens, content_blocks + model, new_input_tokens, new_output_tokens, new_content_blocks = _collect_ai_data( + event, model, input_tokens, output_tokens, content_blocks ) + assert model is None assert new_input_tokens == input_tokens assert new_output_tokens == output_tokens assert new_content_blocks == ["test"] @@ -788,7 +813,7 @@ def test_collect_ai_data_with_input_json_delta(): ANTHROPIC_VERSION < (0, 27), reason="Versions <0.27.0 do not include InputJSONDelta.", ) -def test_add_ai_data_to_span_with_input_json_delta(sentry_init): +def test_set_output_data_with_input_json_delta(sentry_init): sentry_init( integrations=[AnthropicIntegration(include_prompts=True)], traces_sample_rate=1.0, @@ -798,19 +823,20 @@ def test_add_ai_data_to_span_with_input_json_delta(sentry_init): with start_transaction(name="test"): span = start_span() integration = AnthropicIntegration() - - _add_ai_data_to_span( + json_deltas = ["{'test': 'data',", "'more': 'json'}"] + _set_output_data( span, integration, + model="", input_tokens=10, output_tokens=20, - content_blocks=["{'test': 'data',", "'more': 'json'}"], + content_blocks=[{"text": "".join(json_deltas), "type": "text"}], ) - assert span._data.get("ai.responses") == [ - {"type": "text", "text": "{'test': 'data','more': 'json'}"} - ] - assert span._data.get("ai.streaming") is True - assert span._data.get("gen_ai.usage.input_tokens") == 10 - assert span._data.get("gen_ai.usage.output_tokens") == 20 - assert span._data.get("gen_ai.usage.total_tokens") == 30 + assert ( + span._data.get(SPANDATA.GEN_AI_RESPONSE_TEXT) + == "[{\"text\": \"{'test': 'data','more': 'json'}\", \"type\": \"text\"}]" + ) + assert span._data.get(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS) == 10 + assert span._data.get(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS) == 20 + assert span._data.get(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS) == 30 From 2a3954631fc0efa9b87da37da11c3de66dd737ba Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 11 Aug 2025 14:26:27 +0200 Subject: [PATCH 400/591] Help for debugging Cron problems (#4686) A debug message to see what check-ins are send including the `monitor_slug` and the check-in `status`. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- sentry_sdk/crons/api.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/sentry_sdk/crons/api.py b/sentry_sdk/crons/api.py index 20e95685a7..b67e5961c8 100644 --- a/sentry_sdk/crons/api.py +++ b/sentry_sdk/crons/api.py @@ -1,6 +1,7 @@ import uuid import sentry_sdk +from sentry_sdk.utils import logger from typing import TYPE_CHECKING @@ -54,4 +55,8 @@ def capture_checkin( sentry_sdk.capture_event(check_in_event) + logger.debug( + f"[Crons] Captured check-in ({check_in_event.get('check_in_id')}): {check_in_event.get('monitor_slug')} -> {check_in_event.get('status')}" + ) + return check_in_event["check_in_id"] From 9c7f9aa30a7eb325eac36b15fe92d3072903e259 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 12 Aug 2025 10:24:08 +0200 Subject: [PATCH 401/591] Fix Redis CI (#4691) A new version of fakeredis was released which [sets a new keyword arg](https://github.com/cunla/fakeredis-py/compare/v2.30.3..v2.31.0#diff-7d354eae970f35e5aa784b88fa4d0fb98ad887adb45b5a1cba5b8df4494c9561R183) that doesn't exist in older Redis versions --- scripts/populate_tox/tox.jinja | 1 + tox.ini | 1 + 2 files changed, 2 insertions(+) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index c67f4127d5..4c3b86af81 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -300,6 +300,7 @@ deps = {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 + redis-v4: fakeredis<2.31.0 redis-v5: redis~=5.0 redis-latest: redis diff --git a/tox.ini b/tox.ini index 16067de8c7..88cf8ceddb 100644 --- a/tox.ini +++ b/tox.ini @@ -468,6 +468,7 @@ deps = {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio redis-v3: redis~=3.0 redis-v4: redis~=4.0 + redis-v4: fakeredis<2.31.0 redis-v5: redis~=5.0 redis-latest: redis From 0dc5d43d90a98c40c9ba3c4154ef422f89a845f2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 12 Aug 2025 11:52:41 +0200 Subject: [PATCH 402/591] Update tox.ini (#4689) Also fix pymongo tests complaining about the server version --- tests/integrations/pymongo/test_pymongo.py | 2 +- tox.ini | 68 +++++++++++----------- 2 files changed, 35 insertions(+), 35 deletions(-) diff --git a/tests/integrations/pymongo/test_pymongo.py b/tests/integrations/pymongo/test_pymongo.py index 10f1c9fba9..7e6556f85a 100644 --- a/tests/integrations/pymongo/test_pymongo.py +++ b/tests/integrations/pymongo/test_pymongo.py @@ -10,7 +10,7 @@ @pytest.fixture(scope="session") def mongo_server(): server = MockupDB(verbose=True) - server.autoresponds("ismaster", maxWireVersion=7) + server.autoresponds("ismaster", maxWireVersion=8) server.run() server.autoresponds( {"find": "test_collection"}, cursor={"id": 123, "firstBatch": []} diff --git a/tox.ini b/tox.ini index 88cf8ceddb..a1b1327af5 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-07-29T06:07:22.069934+00:00 +# Last generated: 2025-08-12T07:16:34.585160+00:00 [tox] requires = @@ -138,21 +138,21 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.31.2 {py3.8,py3.11,py3.12}-anthropic-v0.46.0 - {py3.8,py3.12,py3.13}-anthropic-v0.60.0 + {py3.8,py3.12,py3.13}-anthropic-v0.62.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 {py3.9,py3.11,py3.12}-cohere-v5.13.12 - {py3.9,py3.11,py3.12}-cohere-v5.16.1 + {py3.9,py3.11,py3.12}-cohere-v5.16.3 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 - {py3.10,py3.12,py3.13}-openai_agents-v0.2.3 + {py3.10,py3.12,py3.13}-openai_agents-v0.2.6 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.2 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.4 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 @@ -162,7 +162,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.13.2 + {py3.9,py3.12,py3.13}-pymongo-v4.14.0 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -170,7 +170,7 @@ envlist = {py3.6,py3.8,py3.9}-sqlalchemy-v1.3.24 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 - {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.41 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.43 # ~~~ Flags ~~~ @@ -180,7 +180,7 @@ envlist = {py3.9,py3.12,py3.13}-launchdarkly-v9.12.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 - {py3.9,py3.12,py3.13}-openfeature-v0.8.1 + {py3.9,py3.12,py3.13}-openfeature-v0.8.2 {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.57.3 @@ -209,7 +209,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.232.2 {py3.8,py3.12,py3.13}-strawberry-v0.255.0 - {py3.9,py3.12,py3.13}-strawberry-v0.278.0 + {py3.9,py3.12,py3.13}-strawberry-v0.278.1 # ~~~ Network ~~~ @@ -245,7 +245,7 @@ envlist = {py3.6,py3.9,py3.10}-django-v3.2.25 {py3.8,py3.11,py3.12}-django-v4.2.23 {py3.10,py3.11,py3.12}-django-v5.0.14 - {py3.10,py3.12,py3.13}-django-v5.2.4 + {py3.10,py3.12,py3.13}-django-v5.2.5 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -276,12 +276,12 @@ envlist = {py3.6,py3.7}-falcon-v2.0.0 {py3.6,py3.11,py3.12}-falcon-v3.1.3 {py3.8,py3.11,py3.12}-falcon-v4.0.2 - {py3.8,py3.11,py3.12}-falcon-v4.1.0a3 + {py3.8,py3.11,py3.12}-falcon-v4.1.0 {py3.8,py3.10,py3.11}-litestar-v2.0.1 - {py3.8,py3.11,py3.12}-litestar-v2.5.5 - {py3.8,py3.11,py3.12}-litestar-v2.10.0 - {py3.8,py3.12,py3.13}-litestar-v2.16.0 + {py3.8,py3.11,py3.12}-litestar-v2.6.4 + {py3.8,py3.11,py3.12}-litestar-v2.12.1 + {py3.8,py3.12,py3.13}-litestar-v2.17.0 {py3.6}-pyramid-v1.8.6 {py3.6,py3.8,py3.9}-pyramid-v1.10.8 @@ -295,7 +295,7 @@ envlist = {py3.6,py3.7,py3.8}-tornado-v6.0.4 {py3.7,py3.9,py3.10}-tornado-v6.2 {py3.8,py3.10,py3.11}-tornado-v6.4.2 - {py3.9,py3.12,py3.13}-tornado-v6.5.1 + {py3.9,py3.12,py3.13}-tornado-v6.5.2 # ~~~ Misc ~~~ @@ -306,7 +306,7 @@ envlist = {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 {py3.8,py3.11,py3.12}-trytond-v7.0.34 - {py3.9,py3.12,py3.13}-trytond-v7.6.4 + {py3.9,py3.12,py3.13}-trytond-v7.6.5 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.0 @@ -513,7 +513,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.31.2: anthropic==0.31.2 anthropic-v0.46.0: anthropic==0.46.0 - anthropic-v0.60.0: anthropic==0.60.0 + anthropic-v0.62.0: anthropic==0.62.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.31.2: httpx<0.28.0 @@ -522,17 +522,17 @@ deps = cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 cohere-v5.13.12: cohere==5.13.12 - cohere-v5.16.1: cohere==5.16.1 + cohere-v5.16.3: cohere==5.16.3 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 - openai_agents-v0.2.3: openai-agents==0.2.3 + openai_agents-v0.2.6: openai-agents==0.2.6 openai_agents: pytest-asyncio huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.34.2: huggingface_hub==0.34.2 + huggingface_hub-v0.34.4: huggingface_hub==0.34.4 huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 @@ -542,7 +542,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.13.2: pymongo==4.13.2 + pymongo-v4.14.0: pymongo==4.14.0 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -551,7 +551,7 @@ deps = sqlalchemy-v1.3.24: sqlalchemy==1.3.24 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 - sqlalchemy-v2.0.41: sqlalchemy==2.0.41 + sqlalchemy-v2.0.43: sqlalchemy==2.0.43 # ~~~ Flags ~~~ @@ -561,7 +561,7 @@ deps = launchdarkly-v9.12.0: launchdarkly-server-sdk==9.12.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 - openfeature-v0.8.1: openfeature-sdk==0.8.1 + openfeature-v0.8.2: openfeature-sdk==0.8.2 statsig-v0.55.3: statsig==0.55.3 statsig-v0.57.3: statsig==0.57.3 @@ -599,7 +599,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.232.2: strawberry-graphql[fastapi,flask]==0.232.2 strawberry-v0.255.0: strawberry-graphql[fastapi,flask]==0.255.0 - strawberry-v0.278.0: strawberry-graphql[fastapi,flask]==0.278.0 + strawberry-v0.278.1: strawberry-graphql[fastapi,flask]==0.278.1 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.232.2: pydantic<2.11 @@ -646,7 +646,7 @@ deps = django-v3.2.25: django==3.2.25 django-v4.2.23: django==4.2.23 django-v5.0.14: django==5.0.14 - django-v5.2.4: django==5.2.4 + django-v5.2.5: django==5.2.5 django: psycopg2-binary django: djangorestframework django: pytest-django @@ -655,12 +655,12 @@ deps = django-v3.2.25: channels[daphne] django-v4.2.23: channels[daphne] django-v5.0.14: channels[daphne] - django-v5.2.4: channels[daphne] + django-v5.2.5: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio django-v4.2.23: pytest-asyncio django-v5.0.14: pytest-asyncio - django-v5.2.4: pytest-asyncio + django-v5.2.5: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 @@ -726,18 +726,18 @@ deps = falcon-v2.0.0: falcon==2.0.0 falcon-v3.1.3: falcon==3.1.3 falcon-v4.0.2: falcon==4.0.2 - falcon-v4.1.0a3: falcon==4.1.0a3 + falcon-v4.1.0: falcon==4.1.0 litestar-v2.0.1: litestar==2.0.1 - litestar-v2.5.5: litestar==2.5.5 - litestar-v2.10.0: litestar==2.10.0 - litestar-v2.16.0: litestar==2.16.0 + litestar-v2.6.4: litestar==2.6.4 + litestar-v2.12.1: litestar==2.12.1 + litestar-v2.17.0: litestar==2.17.0 litestar: pytest-asyncio litestar: python-multipart litestar: requests litestar: cryptography litestar-v2.0.1: httpx<0.28 - litestar-v2.5.5: httpx<0.28 + litestar-v2.6.4: httpx<0.28 pyramid-v1.8.6: pyramid==1.8.6 pyramid-v1.10.8: pyramid==1.10.8 @@ -758,7 +758,7 @@ deps = tornado-v6.0.4: tornado==6.0.4 tornado-v6.2: tornado==6.2 tornado-v6.4.2: tornado==6.4.2 - tornado-v6.5.1: tornado==6.5.1 + tornado-v6.5.2: tornado==6.5.2 tornado: pytest tornado-v6.0.4: pytest<8.2 tornado-v6.2: pytest<8.2 @@ -773,7 +773,7 @@ deps = trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 trytond-v7.0.34: trytond==7.0.34 - trytond-v7.6.4: trytond==7.6.4 + trytond-v7.6.5: trytond==7.6.5 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 775dae81d2cdcc639eff1fc75a1e577f9075a724 Mon Sep 17 00:00:00 2001 From: MeredithAnya Date: Tue, 12 Aug 2025 04:23:53 -0700 Subject: [PATCH 403/591] ref(gnu-integration): make path optional (#4688) I updated the GNU Integration in https://github.com/getsentry/sentry-python/pull/4598 but I didn't make the path optional so if we didn't have the path included, then the stacktrace didn't get parsed: ```python # got parsed "17. DB::TCPHandler::runImpl() @ 0x00000000121bb5d8 in /usr/bin/clickhouse" # didn't get parsed "17. DB::TCPHandler::runImpl() @ 0x00000000121bb5d8" ``` So updated it so that regardless of whether the path is present it will get parsed --- sentry_sdk/integrations/gnu_backtrace.py | 9 ++++--- tests/integrations/test_gnu_backtrace.py | 32 +++++++++++++++++++++++- 2 files changed, 37 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/gnu_backtrace.py b/sentry_sdk/integrations/gnu_backtrace.py index 21d8ea9b38..8241e27f13 100644 --- a/sentry_sdk/integrations/gnu_backtrace.py +++ b/sentry_sdk/integrations/gnu_backtrace.py @@ -11,13 +11,16 @@ from typing import Any from sentry_sdk._types import Event - -FUNCTION_RE = r"[^@]+?)\s+@\s+0x[0-9a-fA-F]+" +# function is everything between index at @ +# and then we match on the @ plus the hex val +FUNCTION_RE = r"[^@]+?" +HEX_ADDRESS = r"\s+@\s+0x[0-9a-fA-F]+" FRAME_RE = r""" -^(?P\d+)\.\s+(?P{FUNCTION_RE}\s+in\s+(?P.+)$ +^(?P\d+)\.\s+(?P{FUNCTION_RE}){HEX_ADDRESS}(?:\s+in\s+(?P.+))?$ """.format( FUNCTION_RE=FUNCTION_RE, + HEX_ADDRESS=HEX_ADDRESS, ) FRAME_RE = re.compile(FRAME_RE, re.MULTILINE | re.VERBOSE) diff --git a/tests/integrations/test_gnu_backtrace.py b/tests/integrations/test_gnu_backtrace.py index 63930f850d..be7346a2c3 100644 --- a/tests/integrations/test_gnu_backtrace.py +++ b/tests/integrations/test_gnu_backtrace.py @@ -31,8 +31,38 @@ 24. ? @ 0x00000000000d162c in /usr/lib/aarch64-linux-gnu/libc-2.31.so """ +LINES_NO_PATH = r""" +0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000bfc38a4 +1. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String&&, String&&) @ 0x00000000075d242c +2. DB::ActionsMatcher::visit(DB::ASTIdentifier const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1c648 +3. DB::ActionsMatcher::visit(DB::ASTFunction const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1f58c +4. DB::ActionsMatcher::visit(DB::ASTFunction const&, std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1f58c +5. DB::ActionsMatcher::visit(std::shared_ptr const&, DB::ActionsMatcher::Data&) @ 0x0000000010b1c394 +6. DB::InDepthNodeVisitor const>::doVisit(std::shared_ptr const&) @ 0x0000000010b154a0 +7. DB::ExpressionAnalyzer::getRootActions(std::shared_ptr const&, bool, std::shared_ptr&, bool) @ 0x0000000010af83b4 +8. DB::SelectQueryExpressionAnalyzer::appendSelect(DB::ExpressionActionsChain&, bool) @ 0x0000000010aff168 +9. DB::ExpressionAnalysisResult::ExpressionAnalysisResult(DB::SelectQueryExpressionAnalyzer&, std::shared_ptr const&, bool, bool, bool, std::shared_ptr const&, std::shared_ptr const&, DB::Block const&) @ 0x0000000010b05b74 +10. DB::InterpreterSelectQuery::getSampleBlockImpl() @ 0x00000000111559fc +11. DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, std::shared_ptr const&, std::optional, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&, std::shared_ptr const&, std::shared_ptr)::$_0::operator()(bool) const @ 0x0000000011148254 +12. DB::InterpreterSelectQuery::InterpreterSelectQuery(std::shared_ptr const&, std::shared_ptr const&, std::optional, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&, std::shared_ptr const&, std::shared_ptr) @ 0x00000000111413e8 +13. DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::shared_ptr const&, std::shared_ptr, DB::SelectQueryOptions const&, std::vector> const&) @ 0x00000000111d3708 +14. DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x0000000011100b64 +15. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, bool, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x00000000114c3f3c +16. DB::executeQuery(String const&, std::shared_ptr, bool, DB::QueryProcessingStage::Enum) @ 0x00000000114c0ec8 +17. DB::TCPHandler::runImpl() @ 0x00000000121bb5d8 +18. DB::TCPHandler::run() @ 0x00000000121cb728 +19. Poco::Net::TCPServerConnection::start() @ 0x00000000146d9404 +20. Poco::Net::TCPServerDispatcher::run() @ 0x00000000146da900 +21. Poco::PooledThread::run() @ 0x000000001484da7c +22. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001484bc24 +23. start_thread @ 0x0000000000007624 +24. ? @ 0x00000000000d162c +""" + -@pytest.mark.parametrize("input", LINES.strip().splitlines()) +@pytest.mark.parametrize( + "input", LINES.strip().splitlines() + LINES_NO_PATH.strip().splitlines() +) def test_basic(sentry_init, capture_events, input): sentry_init(integrations=[GnuBacktraceIntegration()]) events = capture_events() From 46eb82c80ef869a2e009b54f885b646a300844f3 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 12 Aug 2025 16:51:03 +0200 Subject: [PATCH 404/591] Remove performance papercuts (#4675) Removing some papercuts: * don't guess ASGI version everytime the middleware is initialized if we already know * remove debug logs (they're bad in async code) * use `.copy()` instead of copying stuff via constructor (e.g. `dict(old_old)`) * make UUID generation lazier Ref https://github.com/getsentry/sentry-python/issues/4660 Closes https://github.com/getsentry/sentry-python/issues/3908 --------- Co-authored-by: Neel Shah --- sentry_sdk/integrations/asgi.py | 36 ++++++++------------------ sentry_sdk/integrations/django/asgi.py | 2 +- sentry_sdk/integrations/fastapi.py | 8 +----- sentry_sdk/integrations/litestar.py | 2 +- sentry_sdk/integrations/quart.py | 2 +- sentry_sdk/integrations/starlette.py | 6 +---- sentry_sdk/integrations/starlite.py | 2 +- sentry_sdk/scope.py | 22 ++++++++-------- sentry_sdk/tracing.py | 35 +++++++++++++++++++++---- 9 files changed, 58 insertions(+), 57 deletions(-) diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index 1b020ebbc0..dde8128a33 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -12,7 +12,6 @@ import sentry_sdk from sentry_sdk.api import continue_trace from sentry_sdk.consts import OP - from sentry_sdk.integrations._asgi_common import ( _get_headers, _get_request_data, @@ -42,7 +41,6 @@ if TYPE_CHECKING: from typing import Any - from typing import Callable from typing import Dict from typing import Optional from typing import Tuple @@ -102,6 +100,7 @@ def __init__( mechanism_type="asgi", # type: str span_origin="manual", # type: str http_methods_to_capture=DEFAULT_HTTP_METHODS_TO_CAPTURE, # type: Tuple[str, ...] + asgi_version=None, # type: Optional[int] ): # type: (...) -> None """ @@ -140,10 +139,16 @@ def __init__( self.app = app self.http_methods_to_capture = http_methods_to_capture - if _looks_like_asgi3(app): - self.__call__ = self._run_asgi3 # type: Callable[..., Any] - else: - self.__call__ = self._run_asgi2 + if asgi_version is None: + if _looks_like_asgi3(app): + asgi_version = 3 + else: + asgi_version = 2 + + if asgi_version == 3: + self.__call__ = self._run_asgi3 + elif asgi_version == 2: + self.__call__ = self._run_asgi2 # type: ignore def _capture_lifespan_exception(self, exc): # type: (Exception) -> None @@ -217,10 +222,6 @@ async def _run_app(self, scope, receive, send, asgi_version): source=transaction_source, origin=self.span_origin, ) - logger.debug( - "[ASGI] Created transaction (continuing trace): %s", - transaction, - ) else: transaction = Transaction( op=OP.HTTP_SERVER, @@ -228,17 +229,9 @@ async def _run_app(self, scope, receive, send, asgi_version): source=transaction_source, origin=self.span_origin, ) - logger.debug( - "[ASGI] Created transaction (new): %s", transaction - ) if transaction: transaction.set_tag("asgi.type", ty) - logger.debug( - "[ASGI] Set transaction name and source on transaction: '%s' / '%s'", - transaction.name, - transaction.source, - ) with ( sentry_sdk.start_transaction( @@ -248,7 +241,6 @@ async def _run_app(self, scope, receive, send, asgi_version): if transaction is not None else nullcontext() ): - logger.debug("[ASGI] Started transaction: %s", transaction) try: async def _sentry_wrapped_send(event): @@ -303,12 +295,6 @@ def event_processor(self, event, hint, asgi_scope): event["transaction"] = name event["transaction_info"] = {"source": source} - logger.debug( - "[ASGI] Set transaction name and source in event_processor: '%s' / '%s'", - event["transaction"], - event["transaction_info"]["source"], - ) - return event # Helper functions. diff --git a/sentry_sdk/integrations/django/asgi.py b/sentry_sdk/integrations/django/asgi.py index 63a3f0b8f2..773c538045 100644 --- a/sentry_sdk/integrations/django/asgi.py +++ b/sentry_sdk/integrations/django/asgi.py @@ -155,7 +155,7 @@ async def sentry_patched_asgi_handler(self, receive, send): http_methods_to_capture=integration.http_methods_to_capture, ) - return await middleware(self.scope)(receive, send) + return await middleware(self.scope)(receive, send) # type: ignore cls.__call__ = sentry_patched_asgi_handler diff --git a/sentry_sdk/integrations/fastapi.py b/sentry_sdk/integrations/fastapi.py index 76c6adee0f..1473cbcab7 100644 --- a/sentry_sdk/integrations/fastapi.py +++ b/sentry_sdk/integrations/fastapi.py @@ -6,10 +6,7 @@ from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing import SOURCE_FOR_STYLE, TransactionSource -from sentry_sdk.utils import ( - transaction_from_function, - logger, -) +from sentry_sdk.utils import transaction_from_function from typing import TYPE_CHECKING @@ -66,9 +63,6 @@ def _set_transaction_name_and_source(scope, transaction_style, request): source = SOURCE_FOR_STYLE[transaction_style] scope.set_transaction_name(name, source=source) - logger.debug( - "[FastAPI] Set transaction name and source on scope: %s / %s", name, source - ) def patch_get_request_handler(): diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 4e15081cba..2be4d376e0 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -85,6 +85,7 @@ def __init__(self, app, span_origin=LitestarIntegration.origin): transaction_style="endpoint", mechanism_type="asgi", span_origin=span_origin, + asgi_version=3, ) def _capture_request_exception(self, exc): @@ -116,7 +117,6 @@ def injection_wrapper(self, *args, **kwargs): *(kwargs.get("after_exception") or []), ] - SentryLitestarASGIMiddleware.__call__ = SentryLitestarASGIMiddleware._run_asgi3 # type: ignore middleware = kwargs.get("middleware") or [] kwargs["middleware"] = [SentryLitestarASGIMiddleware, *middleware] old__init__(self, *args, **kwargs) diff --git a/sentry_sdk/integrations/quart.py b/sentry_sdk/integrations/quart.py index 51306bb4cd..64f7e0bcd2 100644 --- a/sentry_sdk/integrations/quart.py +++ b/sentry_sdk/integrations/quart.py @@ -95,8 +95,8 @@ async def sentry_patched_asgi_app(self, scope, receive, send): middleware = SentryAsgiMiddleware( lambda *a, **kw: old_app(self, *a, **kw), span_origin=QuartIntegration.origin, + asgi_version=3, ) - middleware.__call__ = middleware._run_asgi3 return await middleware(scope, receive, send) Quart.__call__ = sentry_patched_asgi_app diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index d0f0bf2045..c7ce40618b 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -29,7 +29,6 @@ capture_internal_exceptions, ensure_integration_enabled, event_from_exception, - logger, parse_version, transaction_from_function, ) @@ -403,9 +402,9 @@ async def _sentry_patched_asgi_app(self, scope, receive, send): if integration else DEFAULT_HTTP_METHODS_TO_CAPTURE ), + asgi_version=3, ) - middleware.__call__ = middleware._run_asgi3 return await middleware(scope, receive, send) Starlette.__call__ = _sentry_patched_asgi_app @@ -723,9 +722,6 @@ def _set_transaction_name_and_source(scope, transaction_style, request): source = TransactionSource.ROUTE scope.set_transaction_name(name, source=source) - logger.debug( - "[Starlette] Set transaction name and source on scope: %s / %s", name, source - ) def _get_transaction_from_middleware(app, asgi_scope, integration): diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index 24707a18b1..b402aa2184 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -65,6 +65,7 @@ def __init__(self, app, span_origin=StarliteIntegration.origin): transaction_style="endpoint", mechanism_type="asgi", span_origin=span_origin, + asgi_version=3, ) @@ -94,7 +95,6 @@ def injection_wrapper(self, *args, **kwargs): ] ) - SentryStarliteASGIMiddleware.__call__ = SentryStarliteASGIMiddleware._run_asgi3 # type: ignore middleware = kwargs.get("middleware") or [] kwargs["middleware"] = [SentryStarliteASGIMiddleware, *middleware] old__init__(self, *args, **kwargs) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 73bf43573e..3356de57a8 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -48,7 +48,7 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from collections.abc import Mapping, MutableMapping + from collections.abc import Mapping from typing import Any from typing import Callable @@ -238,24 +238,24 @@ def __copy__(self): rv._name = self._name rv._fingerprint = self._fingerprint rv._transaction = self._transaction - rv._transaction_info = dict(self._transaction_info) + rv._transaction_info = self._transaction_info.copy() rv._user = self._user - rv._tags = dict(self._tags) - rv._contexts = dict(self._contexts) - rv._extras = dict(self._extras) + rv._tags = self._tags.copy() + rv._contexts = self._contexts.copy() + rv._extras = self._extras.copy() rv._breadcrumbs = copy(self._breadcrumbs) - rv._n_breadcrumbs_truncated = copy(self._n_breadcrumbs_truncated) - rv._event_processors = list(self._event_processors) - rv._error_processors = list(self._error_processors) + rv._n_breadcrumbs_truncated = self._n_breadcrumbs_truncated + rv._event_processors = self._event_processors.copy() + rv._error_processors = self._error_processors.copy() rv._propagation_context = self._propagation_context rv._should_capture = self._should_capture rv._span = self._span rv._session = self._session rv._force_auto_session_tracking = self._force_auto_session_tracking - rv._attachments = list(self._attachments) + rv._attachments = self._attachments.copy() rv._profile = self._profile @@ -683,12 +683,12 @@ def clear(self): self._level = None # type: Optional[LogLevelStr] self._fingerprint = None # type: Optional[List[str]] self._transaction = None # type: Optional[str] - self._transaction_info = {} # type: MutableMapping[str, str] + self._transaction_info = {} # type: dict[str, str] self._user = None # type: Optional[Dict[str, Any]] self._tags = {} # type: Dict[str, Any] self._contexts = {} # type: Dict[str, Dict[str, Any]] - self._extras = {} # type: MutableMapping[str, Any] + self._extras = {} # type: dict[str, Any] self._attachments = [] # type: List[Attachment] self.clear_breadcrumbs() diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index e9d726cc66..92f7ae2073 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -257,8 +257,8 @@ class Span: """ __slots__ = ( - "trace_id", - "span_id", + "_trace_id", + "_span_id", "parent_span_id", "same_process_as_parent", "sampled", @@ -301,8 +301,8 @@ def __init__( name=None, # type: Optional[str] ): # type: (...) -> None - self.trace_id = trace_id or uuid.uuid4().hex - self.span_id = span_id or uuid.uuid4().hex[16:] + self._trace_id = trace_id + self._span_id = span_id self.parent_span_id = parent_span_id self.same_process_as_parent = same_process_as_parent self.sampled = sampled @@ -356,6 +356,32 @@ def init_span_recorder(self, maxlen): if self._span_recorder is None: self._span_recorder = _SpanRecorder(maxlen) + @property + def trace_id(self): + # type: () -> str + if not self._trace_id: + self._trace_id = uuid.uuid4().hex + + return self._trace_id + + @trace_id.setter + def trace_id(self, value): + # type: (str) -> None + self._trace_id = value + + @property + def span_id(self): + # type: () -> str + if not self._span_id: + self._span_id = uuid.uuid4().hex[16:] + + return self._span_id + + @span_id.setter + def span_id(self, value): + # type: (str) -> None + self._span_id = value + def _get_local_aggregator(self): # type: (...) -> LocalAggregator rv = self._local_aggregator @@ -822,7 +848,6 @@ def __init__( # type: ignore[misc] **kwargs, # type: Unpack[SpanKwargs] ): # type: (...) -> None - super().__init__(**kwargs) self.name = name From 1804955b12aa19bd3492bb15be062e9706c58c34 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 13 Aug 2025 13:04:04 +0200 Subject: [PATCH 405/591] feat(tracing): AI Agents templates for `@trace` decorator (#4676) Adding a `template` parameter to the `@trace` decorator to make it easier to manually create spans for insights modules. Currently there are three templates supported `ai_agent`, `ai_tool`, and `ai_chat`. --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- sentry_sdk/consts.py | 11 ++ sentry_sdk/tracing.py | 30 +++- sentry_sdk/tracing_utils.py | 280 +++++++++++++++++++++++++++++++- tests/tracing/test_decorator.py | 247 ++++++++++++++++++++++++++++ 4 files changed, 555 insertions(+), 13 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index d402467e5e..d880845011 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -100,6 +100,17 @@ class CompressionAlgo(Enum): ] +class SPANTEMPLATE(str, Enum): + DEFAULT = "default" + AI_AGENT = "ai_agent" + AI_TOOL = "ai_tool" + AI_CHAT = "ai_chat" + + def __str__(self): + # type: () -> str + return self.value + + class INSTRUMENTER: SENTRY = "sentry" OTEL = "otel" diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 92f7ae2073..c9b357305a 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -5,7 +5,7 @@ from enum import Enum import sentry_sdk -from sentry_sdk.consts import INSTRUMENTER, SPANSTATUS, SPANDATA +from sentry_sdk.consts import INSTRUMENTER, SPANSTATUS, SPANDATA, SPANTEMPLATE from sentry_sdk.profiler.continuous_profiler import get_profiler_id from sentry_sdk.utils import ( get_current_thread_meta, @@ -1365,8 +1365,10 @@ def _set_initial_sampling_decision(self, sampling_context): if TYPE_CHECKING: @overload - def trace(func=None, *, op=None, name=None, attributes=None): - # type: (None, Optional[str], Optional[str], Optional[dict[str, Any]]) -> Callable[[Callable[P, R]], Callable[P, R]] + def trace( + func=None, *, op=None, name=None, attributes=None, template=SPANTEMPLATE.DEFAULT + ): + # type: (None, Optional[str], Optional[str], Optional[dict[str, Any]], SPANTEMPLATE) -> Callable[[Callable[P, R]], Callable[P, R]] # Handles: @trace() and @trace(op="custom") pass @@ -1377,8 +1379,10 @@ def trace(func): pass -def trace(func=None, *, op=None, name=None, attributes=None): - # type: (Optional[Callable[P, R]], Optional[str], Optional[str], Optional[dict[str, Any]]) -> Union[Callable[P, R], Callable[[Callable[P, R]], Callable[P, R]]] +def trace( + func=None, *, op=None, name=None, attributes=None, template=SPANTEMPLATE.DEFAULT +): + # type: (Optional[Callable[P, R]], Optional[str], Optional[str], Optional[dict[str, Any]], SPANTEMPLATE) -> Union[Callable[P, R], Callable[[Callable[P, R]], Callable[P, R]]] """ Decorator to start a child span around a function call. @@ -1407,6 +1411,13 @@ def trace(func=None, *, op=None, name=None, attributes=None): attributes provide additional context about the span's execution. :type attributes: dict[str, Any] or None + :param template: The type of span to create. This determines what kind of + span instrumentation and data collection will be applied. Use predefined + constants from :py:class:`sentry_sdk.consts.SPANTEMPLATE`. + The default is `SPANTEMPLATE.DEFAULT` which is the right choice for most + use cases. + :type template: :py:class:`sentry_sdk.consts.SPANTEMPLATE` + :returns: When used as ``@trace``, returns the decorated function. When used as ``@trace(...)`` with parameters, returns a decorator function. :rtype: Callable or decorator function @@ -1414,7 +1425,7 @@ def trace(func=None, *, op=None, name=None, attributes=None): Example:: import sentry_sdk - from sentry_sdk.consts import OP + from sentry_sdk.consts import OP, SPANTEMPLATE # Simple usage with default values @sentry_sdk.trace @@ -1431,6 +1442,12 @@ def process_data(): def make_db_query(sql): # Function implementation pass + + # With a custom template + @sentry_sdk.trace(template=SPANTEMPLATE.AI_TOOL) + def calculate_interest_rate(amount, rate, years): + # Function implementation + pass """ from sentry_sdk.tracing_utils import create_span_decorator @@ -1438,6 +1455,7 @@ def make_db_query(sql): op=op, name=name, attributes=attributes, + template=template, ) if func: diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 447a708d4d..b31d3d85c5 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -12,7 +12,7 @@ import uuid import sentry_sdk -from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.consts import OP, SPANDATA, SPANTEMPLATE from sentry_sdk.utils import ( capture_internal_exceptions, filename_for_module, @@ -20,6 +20,7 @@ logger, match_regex_list, qualname_from_function, + safe_repr, to_string, try_convert, is_sentry_url, @@ -770,15 +771,27 @@ def normalize_incoming_data(incoming_data): return data -def create_span_decorator(op=None, name=None, attributes=None): - # type: (Optional[str], Optional[str], Optional[dict[str, Any]]) -> Any +def create_span_decorator( + op=None, name=None, attributes=None, template=SPANTEMPLATE.DEFAULT +): + # type: (Optional[Union[str, OP]], Optional[str], Optional[dict[str, Any]], SPANTEMPLATE) -> Any """ Create a span decorator that can wrap both sync and async functions. :param op: The operation type for the span. + :type op: str or :py:class:`sentry_sdk.consts.OP` or None :param name: The name of the span. + :type name: str or None :param attributes: Additional attributes to set on the span. + :type attributes: dict or None + :param template: The type of span to create. This determines what kind of + span instrumentation and data collection will be applied. Use predefined + constants from :py:class:`sentry_sdk.consts.SPANTEMPLATE`. + The default is `SPANTEMPLATE.DEFAULT` which is the right choice for most + use cases. + :type template: :py:class:`sentry_sdk.consts.SPANTEMPLATE` """ + from sentry_sdk.scope import should_send_default_pii def span_decorator(f): # type: (Any) -> Any @@ -799,15 +812,24 @@ async def async_wrapper(*args, **kwargs): ) return await f(*args, **kwargs) - span_op = op or OP.FUNCTION - span_name = name or qualname_from_function(f) or "" + span_op = op or _get_span_op(template) + function_name = name or qualname_from_function(f) or "" + span_name = _get_span_name(template, function_name, kwargs) + send_pii = should_send_default_pii() with current_span.start_child( op=span_op, name=span_name, ) as span: span.update_data(attributes or {}) + _set_input_attributes( + span, template, send_pii, function_name, f, args, kwargs + ) + result = await f(*args, **kwargs) + + _set_output_attributes(span, template, send_pii, result) + return result try: @@ -828,15 +850,24 @@ def sync_wrapper(*args, **kwargs): ) return f(*args, **kwargs) - span_op = op or OP.FUNCTION - span_name = name or qualname_from_function(f) or "" + span_op = op or _get_span_op(template) + function_name = name or qualname_from_function(f) or "" + span_name = _get_span_name(template, function_name, kwargs) + send_pii = should_send_default_pii() with current_span.start_child( op=span_op, name=span_name, ) as span: span.update_data(attributes or {}) + _set_input_attributes( + span, template, send_pii, function_name, f, args, kwargs + ) + result = f(*args, **kwargs) + + _set_output_attributes(span, template, send_pii, result) + return result try: @@ -912,6 +943,241 @@ def _sample_rand_range(parent_sampled, sample_rate): return sample_rate, 1.0 +def _get_value(source, key): + # type: (Any, str) -> Optional[Any] + """ + Gets a value from a source object. The source can be a dict or an object. + It is checked for dictionary keys and object attributes. + """ + value = None + if isinstance(source, dict): + value = source.get(key) + else: + if hasattr(source, key): + try: + value = getattr(source, key) + except Exception: + value = None + return value + + +def _get_span_name(template, name, kwargs=None): + # type: (Union[str, SPANTEMPLATE], str, Optional[dict[str, Any]]) -> str + """ + Get the name of the span based on the template and the name. + """ + span_name = name + + if template == SPANTEMPLATE.AI_CHAT: + model = None + if kwargs: + for key in ("model", "model_name"): + if kwargs.get(key) and isinstance(kwargs[key], str): + model = kwargs[key] + break + + span_name = f"chat {model}" if model else "chat" + + elif template == SPANTEMPLATE.AI_AGENT: + span_name = f"invoke_agent {name}" + + elif template == SPANTEMPLATE.AI_TOOL: + span_name = f"execute_tool {name}" + + return span_name + + +def _get_span_op(template): + # type: (Union[str, SPANTEMPLATE]) -> str + """ + Get the operation of the span based on the template. + """ + mapping = { + SPANTEMPLATE.AI_CHAT: OP.GEN_AI_CHAT, + SPANTEMPLATE.AI_AGENT: OP.GEN_AI_INVOKE_AGENT, + SPANTEMPLATE.AI_TOOL: OP.GEN_AI_EXECUTE_TOOL, + } # type: dict[Union[str, SPANTEMPLATE], Union[str, OP]] + op = mapping.get(template, OP.FUNCTION) + + return str(op) + + +def _get_input_attributes(template, send_pii, args, kwargs): + # type: (Union[str, SPANTEMPLATE], bool, tuple[Any, ...], dict[str, Any]) -> dict[str, Any] + """ + Get input attributes for the given span template. + """ + attributes = {} # type: dict[str, Any] + + if template in [SPANTEMPLATE.AI_AGENT, SPANTEMPLATE.AI_TOOL, SPANTEMPLATE.AI_CHAT]: + mapping = { + "model": (SPANDATA.GEN_AI_REQUEST_MODEL, str), + "model_name": (SPANDATA.GEN_AI_REQUEST_MODEL, str), + "agent": (SPANDATA.GEN_AI_AGENT_NAME, str), + "agent_name": (SPANDATA.GEN_AI_AGENT_NAME, str), + "max_tokens": (SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, int), + "frequency_penalty": (SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, float), + "presence_penalty": (SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, float), + "temperature": (SPANDATA.GEN_AI_REQUEST_TEMPERATURE, float), + "top_p": (SPANDATA.GEN_AI_REQUEST_TOP_P, float), + "top_k": (SPANDATA.GEN_AI_REQUEST_TOP_K, int), + } + + def _set_from_key(key, value): + # type: (str, Any) -> None + if key in mapping: + (attribute, data_type) = mapping[key] + if value is not None and isinstance(value, data_type): + attributes[attribute] = value + + for key, value in list(kwargs.items()): + if key == "prompt" and isinstance(value, str): + attributes.setdefault(SPANDATA.GEN_AI_REQUEST_MESSAGES, []).append( + {"role": "user", "content": value} + ) + continue + + if key == "system_prompt" and isinstance(value, str): + attributes.setdefault(SPANDATA.GEN_AI_REQUEST_MESSAGES, []).append( + {"role": "system", "content": value} + ) + continue + + _set_from_key(key, value) + + if template == SPANTEMPLATE.AI_TOOL and send_pii: + attributes[SPANDATA.GEN_AI_TOOL_INPUT] = safe_repr( + {"args": args, "kwargs": kwargs} + ) + + # Coerce to string + if SPANDATA.GEN_AI_REQUEST_MESSAGES in attributes: + attributes[SPANDATA.GEN_AI_REQUEST_MESSAGES] = safe_repr( + attributes[SPANDATA.GEN_AI_REQUEST_MESSAGES] + ) + + return attributes + + +def _get_usage_attributes(usage): + # type: (Any) -> dict[str, Any] + """ + Get usage attributes. + """ + attributes = {} + + def _set_from_keys(attribute, keys): + # type: (str, tuple[str, ...]) -> None + for key in keys: + value = _get_value(usage, key) + if value is not None and isinstance(value, int): + attributes[attribute] = value + + _set_from_keys( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, + ("prompt_tokens", "input_tokens"), + ) + _set_from_keys( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, + ("completion_tokens", "output_tokens"), + ) + _set_from_keys( + SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, + ("total_tokens",), + ) + + return attributes + + +def _get_output_attributes(template, send_pii, result): + # type: (Union[str, SPANTEMPLATE], bool, Any) -> dict[str, Any] + """ + Get output attributes for the given span template. + """ + attributes = {} # type: dict[str, Any] + + if template in [SPANTEMPLATE.AI_AGENT, SPANTEMPLATE.AI_TOOL, SPANTEMPLATE.AI_CHAT]: + with capture_internal_exceptions(): + # Usage from result, result.usage, and result.metadata.usage + usage_candidates = [result] + + usage = _get_value(result, "usage") + usage_candidates.append(usage) + + meta = _get_value(result, "metadata") + usage = _get_value(meta, "usage") + usage_candidates.append(usage) + + for usage_candidate in usage_candidates: + if usage_candidate is not None: + attributes.update(_get_usage_attributes(usage_candidate)) + + # Response model + model_name = _get_value(result, "model") + if model_name is not None and isinstance(model_name, str): + attributes[SPANDATA.GEN_AI_RESPONSE_MODEL] = model_name + + model_name = _get_value(result, "model_name") + if model_name is not None and isinstance(model_name, str): + attributes[SPANDATA.GEN_AI_RESPONSE_MODEL] = model_name + + # Tool output + if template == SPANTEMPLATE.AI_TOOL and send_pii: + attributes[SPANDATA.GEN_AI_TOOL_OUTPUT] = safe_repr(result) + + return attributes + + +def _set_input_attributes(span, template, send_pii, name, f, args, kwargs): + # type: (Span, Union[str, SPANTEMPLATE], bool, str, Any, tuple[Any, ...], dict[str, Any]) -> None + """ + Set span input attributes based on the given span template. + + :param span: The span to set attributes on. + :param template: The template to use to set attributes on the span. + :param send_pii: Whether to send PII data. + :param f: The wrapped function. + :param args: The arguments to the wrapped function. + :param kwargs: The keyword arguments to the wrapped function. + """ + attributes = {} # type: dict[str, Any] + + if template == SPANTEMPLATE.AI_AGENT: + attributes = { + SPANDATA.GEN_AI_OPERATION_NAME: "invoke_agent", + SPANDATA.GEN_AI_AGENT_NAME: name, + } + elif template == SPANTEMPLATE.AI_CHAT: + attributes = { + SPANDATA.GEN_AI_OPERATION_NAME: "chat", + } + elif template == SPANTEMPLATE.AI_TOOL: + attributes = { + SPANDATA.GEN_AI_OPERATION_NAME: "execute_tool", + SPANDATA.GEN_AI_TOOL_NAME: name, + } + + docstring = f.__doc__ + if docstring is not None: + attributes[SPANDATA.GEN_AI_TOOL_DESCRIPTION] = docstring + + attributes.update(_get_input_attributes(template, send_pii, args, kwargs)) + span.update_data(attributes or {}) + + +def _set_output_attributes(span, template, send_pii, result): + # type: (Span, Union[str, SPANTEMPLATE], bool, Any) -> None + """ + Set span output attributes based on the given span template. + + :param span: The span to set attributes on. + :param template: The template to use to set attributes on the span. + :param send_pii: Whether to send PII data. + :param result: The result of the wrapped function. + """ + span.update_data(_get_output_attributes(template, send_pii, result) or {}) + + # Circular imports from sentry_sdk.tracing import ( BAGGAGE_HEADER_NAME, diff --git a/tests/tracing/test_decorator.py b/tests/tracing/test_decorator.py index 9a7074c470..15432f5862 100644 --- a/tests/tracing/test_decorator.py +++ b/tests/tracing/test_decorator.py @@ -3,6 +3,8 @@ import pytest +import sentry_sdk +from sentry_sdk.consts import SPANTEMPLATE from sentry_sdk.tracing import trace from sentry_sdk.tracing_utils import create_span_decorator from sentry_sdk.utils import logger @@ -117,3 +119,248 @@ async def _some_function_traced(a, b, c): assert inspect.getcallargs(_some_function, 1, 2, 3) == inspect.getcallargs( _some_function_traced, 1, 2, 3 ) + + +def test_span_templates_ai_dicts(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_TOOL) + def my_tool(arg1, arg2): + return { + "output": "my_tool_result", + "usage": { + "prompt_tokens": 10, + "completion_tokens": 20, + "total_tokens": 30, + }, + } + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_CHAT) + def my_chat(model=None, **kwargs): + return { + "content": "my_chat_result", + "usage": { + "input_tokens": 11, + "output_tokens": 22, + "total_tokens": 33, + }, + "model": f"{model}-v123", + } + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_AGENT) + def my_agent(): + my_tool(1, 2) + my_chat( + model="my-gpt-4o-mini", + prompt="What is the weather in Tokyo?", + system_prompt="You are a helpful assistant that can answer questions about the weather.", + max_tokens=100, + temperature=0.5, + top_p=0.9, + top_k=40, + frequency_penalty=1.0, + presence_penalty=2.0, + ) + + with sentry_sdk.start_transaction(name="test-transaction"): + my_agent() + + (event,) = events + (agent_span, tool_span, chat_span) = event["spans"] + + assert agent_span["op"] == "gen_ai.invoke_agent" + assert ( + agent_span["description"] + == "invoke_agent test_decorator.test_span_templates_ai_dicts..my_agent" + ) + assert agent_span["data"] == { + "gen_ai.agent.name": "test_decorator.test_span_templates_ai_dicts..my_agent", + "gen_ai.operation.name": "invoke_agent", + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + assert tool_span["op"] == "gen_ai.execute_tool" + assert ( + tool_span["description"] + == "execute_tool test_decorator.test_span_templates_ai_dicts..my_tool" + ) + assert tool_span["data"] == { + "gen_ai.tool.name": "test_decorator.test_span_templates_ai_dicts..my_tool", + "gen_ai.operation.name": "execute_tool", + "gen_ai.usage.input_tokens": 10, + "gen_ai.usage.output_tokens": 20, + "gen_ai.usage.total_tokens": 30, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + assert "gen_ai.tool.description" not in tool_span["data"] + + assert chat_span["op"] == "gen_ai.chat" + assert chat_span["description"] == "chat my-gpt-4o-mini" + assert chat_span["data"] == { + "gen_ai.operation.name": "chat", + "gen_ai.request.frequency_penalty": 1.0, + "gen_ai.request.max_tokens": 100, + "gen_ai.request.messages": "[{'role': 'user', 'content': 'What is the weather in Tokyo?'}, {'role': 'system', 'content': 'You are a helpful assistant that can answer questions about the weather.'}]", + "gen_ai.request.model": "my-gpt-4o-mini", + "gen_ai.request.presence_penalty": 2.0, + "gen_ai.request.temperature": 0.5, + "gen_ai.request.top_k": 40, + "gen_ai.request.top_p": 0.9, + "gen_ai.response.model": "my-gpt-4o-mini-v123", + "gen_ai.usage.input_tokens": 11, + "gen_ai.usage.output_tokens": 22, + "gen_ai.usage.total_tokens": 33, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + +def test_span_templates_ai_objects(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_TOOL) + def my_tool(arg1, arg2): + """This is a tool function.""" + mock_usage = mock.Mock() + mock_usage.prompt_tokens = 10 + mock_usage.completion_tokens = 20 + mock_usage.total_tokens = 30 + + mock_result = mock.Mock() + mock_result.output = "my_tool_result" + mock_result.usage = mock_usage + + return mock_result + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_CHAT) + def my_chat(model=None, **kwargs): + mock_result = mock.Mock() + mock_result.content = "my_chat_result" + mock_result.usage = mock.Mock( + input_tokens=11, + output_tokens=22, + total_tokens=33, + ) + mock_result.model = f"{model}-v123" + + return mock_result + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_AGENT) + def my_agent(): + my_tool(1, 2) + my_chat( + model="my-gpt-4o-mini", + prompt="What is the weather in Tokyo?", + system_prompt="You are a helpful assistant that can answer questions about the weather.", + max_tokens=100, + temperature=0.5, + top_p=0.9, + top_k=40, + frequency_penalty=1.0, + presence_penalty=2.0, + ) + + with sentry_sdk.start_transaction(name="test-transaction"): + my_agent() + + (event,) = events + (agent_span, tool_span, chat_span) = event["spans"] + + assert agent_span["op"] == "gen_ai.invoke_agent" + assert ( + agent_span["description"] + == "invoke_agent test_decorator.test_span_templates_ai_objects..my_agent" + ) + assert agent_span["data"] == { + "gen_ai.agent.name": "test_decorator.test_span_templates_ai_objects..my_agent", + "gen_ai.operation.name": "invoke_agent", + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + assert tool_span["op"] == "gen_ai.execute_tool" + assert ( + tool_span["description"] + == "execute_tool test_decorator.test_span_templates_ai_objects..my_tool" + ) + assert tool_span["data"] == { + "gen_ai.tool.name": "test_decorator.test_span_templates_ai_objects..my_tool", + "gen_ai.tool.description": "This is a tool function.", + "gen_ai.operation.name": "execute_tool", + "gen_ai.usage.input_tokens": 10, + "gen_ai.usage.output_tokens": 20, + "gen_ai.usage.total_tokens": 30, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + assert chat_span["op"] == "gen_ai.chat" + assert chat_span["description"] == "chat my-gpt-4o-mini" + assert chat_span["data"] == { + "gen_ai.operation.name": "chat", + "gen_ai.request.frequency_penalty": 1.0, + "gen_ai.request.max_tokens": 100, + "gen_ai.request.messages": "[{'role': 'user', 'content': 'What is the weather in Tokyo?'}, {'role': 'system', 'content': 'You are a helpful assistant that can answer questions about the weather.'}]", + "gen_ai.request.model": "my-gpt-4o-mini", + "gen_ai.request.presence_penalty": 2.0, + "gen_ai.request.temperature": 0.5, + "gen_ai.request.top_k": 40, + "gen_ai.request.top_p": 0.9, + "gen_ai.response.model": "my-gpt-4o-mini-v123", + "gen_ai.usage.input_tokens": 11, + "gen_ai.usage.output_tokens": 22, + "gen_ai.usage.total_tokens": 33, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +def test_span_templates_ai_pii(sentry_init, capture_events, send_default_pii): + sentry_init(traces_sample_rate=1.0, send_default_pii=send_default_pii) + events = capture_events() + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_TOOL) + def my_tool(arg1, arg2, **kwargs): + """This is a tool function.""" + return "tool_output" + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_CHAT) + def my_chat(model=None, **kwargs): + return "chat_output" + + @sentry_sdk.trace(template=SPANTEMPLATE.AI_AGENT) + def my_agent(*args, **kwargs): + my_tool(1, 2, tool_arg1="3", tool_arg2="4") + my_chat( + model="my-gpt-4o-mini", + prompt="What is the weather in Tokyo?", + system_prompt="You are a helpful assistant that can answer questions about the weather.", + max_tokens=100, + temperature=0.5, + top_p=0.9, + top_k=40, + frequency_penalty=1.0, + presence_penalty=2.0, + ) + return "agent_output" + + with sentry_sdk.start_transaction(name="test-transaction"): + my_agent(22, 33, arg1=44, arg2=55) + + (event,) = events + (_, tool_span, _) = event["spans"] + + if send_default_pii: + assert ( + tool_span["data"]["gen_ai.tool.input"] + == "{'args': (1, 2), 'kwargs': {'tool_arg1': '3', 'tool_arg2': '4'}}" + ) + assert tool_span["data"]["gen_ai.tool.output"] == "'tool_output'" + else: + assert "gen_ai.tool.input" not in tool_span["data"] + assert "gen_ai.tool.output" not in tool_span["data"] From 4640531af6e3520ea91a660f54b523796b4f38ff Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 14 Aug 2025 11:37:30 +0200 Subject: [PATCH 406/591] build(deps): bump actions/create-github-app-token from 2.0.6 to 2.1.0 (#4684) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 2.0.6 to 2.1.0.
Release notes

Sourced from actions/create-github-app-token's releases.

v2.1.0

2.1.0 (2025-08-08)

Features

Commits
  • 0f859bf build(release): 2.1.0 [skip ci]
  • a1cbe0f feat: use node24 as runner (#267)
  • d7ee281 build(deps-dev): bump the development-dependencies group across 1 directory w...
  • 93c1f04 build(deps-dev): bump the development-dependencies group with 4 updates (#255)
  • dff4b11 ci(test): set permissions in test workflow (#247)
  • 6d44c9f docs(README): Client ID can be used as App ID (#251)
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=2.0.6&new-version=2.1.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 34815da549..6197f9023d 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@df432ceedc7162793a195dd1713ff69aefc7379e # v2.0.6 + uses: actions/create-github-app-token@0f859bf9e69e887678d5bbfbee594437cb440ffe # v2.1.0 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From e79c36e26a9df00f4ee3b359c7c531a392d938ab Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 14 Aug 2025 18:46:10 +0200 Subject: [PATCH 407/591] feat(langchain): update integration to from ai.* to gen_ai.* span attributes (#4678) - Update span attributes to new OTEL compatible schema Closes TET-992 --------- Co-authored-by: Anton Pirker --- pyproject.toml | 4 + sentry_sdk/consts.py | 6 +- sentry_sdk/integrations/langchain.py | 659 +++++++++++++----- .../integrations/langchain/test_langchain.py | 110 ++- 4 files changed, 543 insertions(+), 236 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index e5eae2c21f..deba247e39 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -126,6 +126,10 @@ ignore_missing_imports = true module = "langchain_core.*" ignore_missing_imports = true +[[tool.mypy.overrides]] +module = "langchain.*" +ignore_missing_imports = true + [[tool.mypy.overrides]] module = "executing.*" ignore_missing_imports = true diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index d880845011..a290697659 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -795,6 +795,7 @@ class OP: GEN_AI_EMBEDDINGS = "gen_ai.embeddings" GEN_AI_EXECUTE_TOOL = "gen_ai.execute_tool" GEN_AI_HANDOFF = "gen_ai.handoff" + GEN_AI_PIPELINE = "gen_ai.pipeline" GEN_AI_INVOKE_AGENT = "gen_ai.invoke_agent" GEN_AI_RESPONSES = "gen_ai.responses" GRAPHQL_EXECUTE = "graphql.execute" @@ -822,11 +823,6 @@ class OP: HUGGINGFACE_HUB_CHAT_COMPLETIONS_CREATE = ( "ai.chat_completions.create.huggingface_hub" ) - LANGCHAIN_PIPELINE = "ai.pipeline.langchain" - LANGCHAIN_RUN = "ai.run.langchain" - LANGCHAIN_TOOL = "ai.tool.langchain" - LANGCHAIN_AGENT = "ai.agent.langchain" - LANGCHAIN_CHAT_COMPLETIONS_CREATE = "ai.chat_completions.create.langchain" QUEUE_PROCESS = "queue.process" QUEUE_PUBLISH = "queue.publish" QUEUE_SUBMIT_ARQ = "queue.submit.arq" diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 8b67c4c994..7e04a740ed 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -3,55 +3,59 @@ from functools import wraps import sentry_sdk -from sentry_sdk.ai.monitoring import set_ai_pipeline_name, record_token_usage -from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.ai.monitoring import set_ai_pipeline_name from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing import Span -from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.tracing_utils import _get_value from sentry_sdk.utils import logger, capture_internal_exceptions from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Any, List, Callable, Dict, Union, Optional + from typing import ( + Any, + AsyncIterator, + Callable, + Dict, + Iterator, + List, + Optional, + Union, + ) from uuid import UUID + try: - from langchain_core.messages import BaseMessage - from langchain_core.outputs import LLMResult + from langchain.agents import AgentExecutor + from langchain_core.agents import AgentFinish from langchain_core.callbacks import ( - manager, BaseCallbackHandler, BaseCallbackManager, Callbacks, + manager, ) - from langchain_core.agents import AgentAction, AgentFinish + from langchain_core.messages import BaseMessage + from langchain_core.outputs import LLMResult + except ImportError: raise DidNotEnable("langchain not installed") DATA_FIELDS = { - "temperature": SPANDATA.AI_TEMPERATURE, - "top_p": SPANDATA.AI_TOP_P, - "top_k": SPANDATA.AI_TOP_K, - "function_call": SPANDATA.AI_FUNCTION_CALL, - "tool_calls": SPANDATA.AI_TOOL_CALLS, - "tools": SPANDATA.AI_TOOLS, - "response_format": SPANDATA.AI_RESPONSE_FORMAT, - "logit_bias": SPANDATA.AI_LOGIT_BIAS, - "tags": SPANDATA.AI_TAGS, + "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + "function_call": SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "tool_calls": SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + "tools": SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, + "top_k": SPANDATA.GEN_AI_REQUEST_TOP_K, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, } -# To avoid double collecting tokens, we do *not* measure -# token counts for models for which we have an explicit integration -NO_COLLECT_TOKEN_MODELS = [ - "openai-chat", - "anthropic-chat", - "cohere-chat", - "huggingface_endpoint", -] - class LangchainIntegration(Integration): identifier = "langchain" @@ -60,25 +64,23 @@ class LangchainIntegration(Integration): # The most number of spans (e.g., LLM calls) that can be processed at the same time. max_spans = 1024 - def __init__( - self, include_prompts=True, max_spans=1024, tiktoken_encoding_name=None - ): - # type: (LangchainIntegration, bool, int, Optional[str]) -> None + def __init__(self, include_prompts=True, max_spans=1024): + # type: (LangchainIntegration, bool, int) -> None self.include_prompts = include_prompts self.max_spans = max_spans - self.tiktoken_encoding_name = tiktoken_encoding_name @staticmethod def setup_once(): # type: () -> None manager._configure = _wrap_configure(manager._configure) + if AgentExecutor is not None: + AgentExecutor.invoke = _wrap_agent_executor_invoke(AgentExecutor.invoke) + AgentExecutor.stream = _wrap_agent_executor_stream(AgentExecutor.stream) + class WatchedSpan: span = None # type: Span - num_completion_tokens = 0 # type: int - num_prompt_tokens = 0 # type: int - no_collect_tokens = False # type: bool children = [] # type: List[WatchedSpan] is_pipeline = False # type: bool @@ -88,26 +90,14 @@ def __init__(self, span): class SentryLangchainCallback(BaseCallbackHandler): # type: ignore[misc] - """Base callback handler that can be used to handle callbacks from langchain.""" + """Callback handler that creates Sentry spans.""" - def __init__(self, max_span_map_size, include_prompts, tiktoken_encoding_name=None): - # type: (int, bool, Optional[str]) -> None + def __init__(self, max_span_map_size, include_prompts): + # type: (int, bool) -> None self.span_map = OrderedDict() # type: OrderedDict[UUID, WatchedSpan] self.max_span_map_size = max_span_map_size self.include_prompts = include_prompts - self.tiktoken_encoding = None - if tiktoken_encoding_name is not None: - import tiktoken # type: ignore - - self.tiktoken_encoding = tiktoken.get_encoding(tiktoken_encoding_name) - - def count_tokens(self, s): - # type: (str) -> int - if self.tiktoken_encoding is not None: - return len(self.tiktoken_encoding.encode_ordinary(s)) - return 0 - def gc_span_map(self): # type: () -> None @@ -117,39 +107,37 @@ def gc_span_map(self): def _handle_error(self, run_id, error): # type: (UUID, Any) -> None - if not run_id or run_id not in self.span_map: - return + with capture_internal_exceptions(): + if not run_id or run_id not in self.span_map: + return - span_data = self.span_map[run_id] - if not span_data: - return - sentry_sdk.capture_exception(error, span_data.span.scope) - span_data.span.__exit__(None, None, None) - del self.span_map[run_id] + span_data = self.span_map[run_id] + span = span_data.span + span.set_status("unknown") + + sentry_sdk.capture_exception(error, span.scope) + + span.__exit__(None, None, None) + del self.span_map[run_id] def _normalize_langchain_message(self, message): # type: (BaseMessage) -> Any - parsed = {"content": message.content, "role": message.type} + parsed = {"role": message.type, "content": message.content} parsed.update(message.additional_kwargs) return parsed def _create_span(self, run_id, parent_id, **kwargs): # type: (SentryLangchainCallback, UUID, Optional[Any], Any) -> WatchedSpan - watched_span = None # type: Optional[WatchedSpan] if parent_id: parent_span = self.span_map.get(parent_id) # type: Optional[WatchedSpan] if parent_span: watched_span = WatchedSpan(parent_span.span.start_child(**kwargs)) parent_span.children.append(watched_span) + if watched_span is None: watched_span = WatchedSpan(sentry_sdk.start_span(**kwargs)) - if kwargs.get("op", "").startswith("ai.pipeline."): - if kwargs.get("name"): - set_ai_pipeline_name(kwargs.get("name")) - watched_span.is_pipeline = True - watched_span.span.__enter__() self.span_map[run_id] = watched_span self.gc_span_map() @@ -157,7 +145,6 @@ def _create_span(self, run_id, parent_id, **kwargs): def _exit_span(self, span_data, run_id): # type: (SentryLangchainCallback, WatchedSpan, UUID) -> None - if span_data.is_pipeline: set_ai_pipeline_name(None) @@ -180,21 +167,44 @@ def on_llm_start( with capture_internal_exceptions(): if not run_id: return + all_params = kwargs.get("invocation_params", {}) all_params.update(serialized.get("kwargs", {})) + + model = ( + all_params.get("model") + or all_params.get("model_name") + or all_params.get("model_id") + or "" + ) + watched_span = self._create_span( run_id, - kwargs.get("parent_run_id"), - op=OP.LANGCHAIN_RUN, + parent_run_id, + op=OP.GEN_AI_PIPELINE, name=kwargs.get("name") or "Langchain LLM call", origin=LangchainIntegration.origin, ) span = watched_span.span + + if model: + span.set_data( + SPANDATA.GEN_AI_REQUEST_MODEL, + model, + ) + + ai_type = all_params.get("_type", "") + if "anthropic" in ai_type: + span.set_data(SPANDATA.GEN_AI_SYSTEM, "anthropic") + elif "openai" in ai_type: + span.set_data(SPANDATA.GEN_AI_SYSTEM, "openai") + + for key, attribute in DATA_FIELDS.items(): + if key in all_params and all_params[key] is not None: + set_data_normalized(span, attribute, all_params[key], unpack=False) + if should_send_default_pii() and self.include_prompts: - set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, prompts) - for k, v in DATA_FIELDS.items(): - if k in all_params: - set_data_normalized(span, v, all_params[k]) + set_data_normalized(span, SPANDATA.GEN_AI_REQUEST_MESSAGES, prompts) def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): # type: (SentryLangchainCallback, Dict[str, Any], List[List[BaseMessage]], UUID, Any) -> Any @@ -202,170 +212,150 @@ def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): with capture_internal_exceptions(): if not run_id: return + all_params = kwargs.get("invocation_params", {}) all_params.update(serialized.get("kwargs", {})) + + model = ( + all_params.get("model") + or all_params.get("model_name") + or all_params.get("model_id") + or "" + ) + watched_span = self._create_span( run_id, kwargs.get("parent_run_id"), - op=OP.LANGCHAIN_CHAT_COMPLETIONS_CREATE, - name=kwargs.get("name") or "Langchain Chat Model", + op=OP.GEN_AI_CHAT, + name=f"chat {model}".strip(), origin=LangchainIntegration.origin, ) span = watched_span.span - model = all_params.get( - "model", all_params.get("model_name", all_params.get("model_id")) - ) - watched_span.no_collect_tokens = any( - x in all_params.get("_type", "") for x in NO_COLLECT_TOKEN_MODELS - ) - if not model and "anthropic" in all_params.get("_type"): - model = "claude-2" + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") if model: - span.set_data(SPANDATA.AI_MODEL_ID, model) + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model) + + ai_type = all_params.get("_type", "") + if "anthropic" in ai_type: + span.set_data(SPANDATA.GEN_AI_SYSTEM, "anthropic") + elif "openai" in ai_type: + span.set_data(SPANDATA.GEN_AI_SYSTEM, "openai") + + for key, attribute in DATA_FIELDS.items(): + if key in all_params and all_params[key] is not None: + set_data_normalized(span, attribute, all_params[key], unpack=False) + if should_send_default_pii() and self.include_prompts: set_data_normalized( span, - SPANDATA.AI_INPUT_MESSAGES, + SPANDATA.GEN_AI_REQUEST_MESSAGES, [ [self._normalize_langchain_message(x) for x in list_] for list_ in messages ], ) - for k, v in DATA_FIELDS.items(): - if k in all_params: - set_data_normalized(span, v, all_params[k]) - if not watched_span.no_collect_tokens: - for list_ in messages: - for message in list_: - self.span_map[run_id].num_prompt_tokens += self.count_tokens( - message.content - ) + self.count_tokens(message.type) - - def on_llm_new_token(self, token, *, run_id, **kwargs): - # type: (SentryLangchainCallback, str, UUID, Any) -> Any - """Run on new LLM token. Only available when streaming is enabled.""" + + def on_chat_model_end(self, response, *, run_id, **kwargs): + # type: (SentryLangchainCallback, LLMResult, UUID, Any) -> Any + """Run when Chat Model ends running.""" with capture_internal_exceptions(): if not run_id or run_id not in self.span_map: return + span_data = self.span_map[run_id] - if not span_data or span_data.no_collect_tokens: - return - span_data.num_completion_tokens += self.count_tokens(token) + span = span_data.span + + if should_send_default_pii() and self.include_prompts: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + [[x.text for x in list_] for list_ in response.generations], + ) + + _record_token_usage(span, response) + self._exit_span(span_data, run_id) def on_llm_end(self, response, *, run_id, **kwargs): # type: (SentryLangchainCallback, LLMResult, UUID, Any) -> Any """Run when LLM ends running.""" with capture_internal_exceptions(): - if not run_id: + if not run_id or run_id not in self.span_map: return - token_usage = ( - response.llm_output.get("token_usage") if response.llm_output else None - ) - span_data = self.span_map[run_id] - if not span_data: - return + span = span_data.span + + try: + generation = response.generations[0][0] + except IndexError: + generation = None + + if generation is not None: + try: + response_model = generation.generation_info.get("model_name") + if response_model is not None: + span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, response_model) + except AttributeError: + pass + + try: + finish_reason = generation.generation_info.get("finish_reason") + if finish_reason is not None: + span.set_data( + SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, finish_reason + ) + except AttributeError: + pass + + try: + tool_calls = getattr(generation.message, "tool_calls", None) + if tool_calls is not None and tool_calls != []: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + tool_calls, + unpack=False, + ) + except AttributeError: + pass if should_send_default_pii() and self.include_prompts: set_data_normalized( - span_data.span, - SPANDATA.AI_RESPONSES, + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, [[x.text for x in list_] for list_ in response.generations], ) - if not span_data.no_collect_tokens: - if token_usage: - record_token_usage( - span_data.span, - input_tokens=token_usage.get("prompt_tokens"), - output_tokens=token_usage.get("completion_tokens"), - total_tokens=token_usage.get("total_tokens"), - ) - else: - record_token_usage( - span_data.span, - input_tokens=span_data.num_prompt_tokens, - output_tokens=span_data.num_completion_tokens, - ) - + _record_token_usage(span, response) self._exit_span(span_data, run_id) def on_llm_error(self, error, *, run_id, **kwargs): # type: (SentryLangchainCallback, Union[Exception, KeyboardInterrupt], UUID, Any) -> Any """Run when LLM errors.""" - with capture_internal_exceptions(): - self._handle_error(run_id, error) - - def on_chain_start(self, serialized, inputs, *, run_id, **kwargs): - # type: (SentryLangchainCallback, Dict[str, Any], Dict[str, Any], UUID, Any) -> Any - """Run when chain starts running.""" - with capture_internal_exceptions(): - if not run_id: - return - watched_span = self._create_span( - run_id, - kwargs.get("parent_run_id"), - op=( - OP.LANGCHAIN_RUN - if kwargs.get("parent_run_id") is not None - else OP.LANGCHAIN_PIPELINE - ), - name=kwargs.get("name") or "Chain execution", - origin=LangchainIntegration.origin, - ) - metadata = kwargs.get("metadata") - if metadata: - set_data_normalized(watched_span.span, SPANDATA.AI_METADATA, metadata) - - def on_chain_end(self, outputs, *, run_id, **kwargs): - # type: (SentryLangchainCallback, Dict[str, Any], UUID, Any) -> Any - """Run when chain ends running.""" - with capture_internal_exceptions(): - if not run_id or run_id not in self.span_map: - return - - span_data = self.span_map[run_id] - if not span_data: - return - self._exit_span(span_data, run_id) + self._handle_error(run_id, error) - def on_chain_error(self, error, *, run_id, **kwargs): + def on_chat_model_error(self, error, *, run_id, **kwargs): # type: (SentryLangchainCallback, Union[Exception, KeyboardInterrupt], UUID, Any) -> Any - """Run when chain errors.""" + """Run when Chat Model errors.""" self._handle_error(run_id, error) - def on_agent_action(self, action, *, run_id, **kwargs): - # type: (SentryLangchainCallback, AgentAction, UUID, Any) -> Any - with capture_internal_exceptions(): - if not run_id: - return - watched_span = self._create_span( - run_id, - kwargs.get("parent_run_id"), - op=OP.LANGCHAIN_AGENT, - name=action.tool or "AI tool usage", - origin=LangchainIntegration.origin, - ) - if action.tool_input and should_send_default_pii() and self.include_prompts: - set_data_normalized( - watched_span.span, SPANDATA.AI_INPUT_MESSAGES, action.tool_input - ) - def on_agent_finish(self, finish, *, run_id, **kwargs): # type: (SentryLangchainCallback, AgentFinish, UUID, Any) -> Any with capture_internal_exceptions(): - if not run_id: + if not run_id or run_id not in self.span_map: return span_data = self.span_map[run_id] - if not span_data: - return + span = span_data.span + if should_send_default_pii() and self.include_prompts: set_data_normalized( - span_data.span, SPANDATA.AI_RESPONSES, finish.return_values.items() + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + finish.return_values.items(), ) + self._exit_span(span_data, run_id) def on_tool_start(self, serialized, input_str, *, run_id, **kwargs): @@ -374,23 +364,31 @@ def on_tool_start(self, serialized, input_str, *, run_id, **kwargs): with capture_internal_exceptions(): if not run_id: return + + tool_name = serialized.get("name") or kwargs.get("name") or "" + watched_span = self._create_span( run_id, kwargs.get("parent_run_id"), - op=OP.LANGCHAIN_TOOL, - name=serialized.get("name") or kwargs.get("name") or "AI tool usage", + op=OP.GEN_AI_EXECUTE_TOOL, + name=f"execute_tool {tool_name}".strip(), origin=LangchainIntegration.origin, ) + span = watched_span.span + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "execute_tool") + span.set_data(SPANDATA.GEN_AI_TOOL_NAME, tool_name) + + tool_description = serialized.get("description") + if tool_description is not None: + span.set_data(SPANDATA.GEN_AI_TOOL_DESCRIPTION, tool_description) + if should_send_default_pii() and self.include_prompts: set_data_normalized( - watched_span.span, - SPANDATA.AI_INPUT_MESSAGES, + span, + SPANDATA.GEN_AI_TOOL_INPUT, kwargs.get("inputs", [input_str]), ) - if kwargs.get("metadata"): - set_data_normalized( - watched_span.span, SPANDATA.AI_METADATA, kwargs.get("metadata") - ) def on_tool_end(self, output, *, run_id, **kwargs): # type: (SentryLangchainCallback, str, UUID, Any) -> Any @@ -400,10 +398,11 @@ def on_tool_end(self, output, *, run_id, **kwargs): return span_data = self.span_map[run_id] - if not span_data: - return + span = span_data.span + if should_send_default_pii() and self.include_prompts: - set_data_normalized(span_data.span, SPANDATA.AI_RESPONSES, output) + set_data_normalized(span, SPANDATA.GEN_AI_TOOL_OUTPUT, output) + self._exit_span(span_data, run_id) def on_tool_error(self, error, *args, run_id, **kwargs): @@ -412,6 +411,126 @@ def on_tool_error(self, error, *args, run_id, **kwargs): self._handle_error(run_id, error) +def _extract_tokens(token_usage): + # type: (Any) -> tuple[Optional[int], Optional[int], Optional[int]] + if not token_usage: + return None, None, None + + input_tokens = _get_value(token_usage, "prompt_tokens") or _get_value( + token_usage, "input_tokens" + ) + output_tokens = _get_value(token_usage, "completion_tokens") or _get_value( + token_usage, "output_tokens" + ) + total_tokens = _get_value(token_usage, "total_tokens") + + return input_tokens, output_tokens, total_tokens + + +def _extract_tokens_from_generations(generations): + # type: (Any) -> tuple[Optional[int], Optional[int], Optional[int]] + """Extract token usage from response.generations structure.""" + if not generations: + return None, None, None + + total_input = 0 + total_output = 0 + total_total = 0 + + for gen_list in generations: + for gen in gen_list: + token_usage = _get_token_usage(gen) + input_tokens, output_tokens, total_tokens = _extract_tokens(token_usage) + total_input += input_tokens if input_tokens is not None else 0 + total_output += output_tokens if output_tokens is not None else 0 + total_total += total_tokens if total_tokens is not None else 0 + + return ( + total_input if total_input > 0 else None, + total_output if total_output > 0 else None, + total_total if total_total > 0 else None, + ) + + +def _get_token_usage(obj): + # type: (Any) -> Optional[Dict[str, Any]] + """ + Check multiple paths to extract token usage from different objects. + """ + possible_names = ("usage", "token_usage", "usage_metadata") + + message = _get_value(obj, "message") + if message is not None: + for name in possible_names: + usage = _get_value(message, name) + if usage is not None: + return usage + + llm_output = _get_value(obj, "llm_output") + if llm_output is not None: + for name in possible_names: + usage = _get_value(llm_output, name) + if usage is not None: + return usage + + # check for usage in the object itself + for name in possible_names: + usage = _get_value(obj, name) + if usage is not None: + return usage + + # no usage found anywhere + return None + + +def _record_token_usage(span, response): + # type: (Span, Any) -> None + token_usage = _get_token_usage(response) + if token_usage: + input_tokens, output_tokens, total_tokens = _extract_tokens(token_usage) + else: + input_tokens, output_tokens, total_tokens = _extract_tokens_from_generations( + response.generations + ) + + if input_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, input_tokens) + + if output_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, output_tokens) + + if total_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, total_tokens) + + +def _get_request_data(obj, args, kwargs): + # type: (Any, Any, Any) -> tuple[Optional[str], Optional[List[Any]]] + """ + Get the agent name and available tools for the agent. + """ + agent = getattr(obj, "agent", None) + runnable = getattr(agent, "runnable", None) + runnable_config = getattr(runnable, "config", {}) + tools = ( + getattr(obj, "tools", None) + or getattr(agent, "tools", None) + or runnable_config.get("tools") + or runnable_config.get("available_tools") + ) + tools = tools if tools and len(tools) > 0 else None + + try: + agent_name = None + if len(args) > 1: + agent_name = args[1].get("run_name") + if agent_name is None: + agent_name = runnable_config.get("run_name") + except Exception: + pass + + return (agent_name, tools) + + def _wrap_configure(f): # type: (Callable[..., Any]) -> Callable[..., Any] @@ -473,7 +592,6 @@ def new_configure( sentry_handler = SentryLangchainCallback( integration.max_spans, integration.include_prompts, - integration.tiktoken_encoding_name, ) if isinstance(local_callbacks, BaseCallbackManager): local_callbacks = local_callbacks.copy() @@ -495,3 +613,158 @@ def new_configure( ) return new_configure + + +def _wrap_agent_executor_invoke(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + + @wraps(f) + def new_invoke(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(LangchainIntegration) + if integration is None: + return f(self, *args, **kwargs) + + agent_name, tools = _get_request_data(self, args, kwargs) + + with sentry_sdk.start_span( + op=OP.GEN_AI_INVOKE_AGENT, + name=f"invoke_agent {agent_name}" if agent_name else "invoke_agent", + origin=LangchainIntegration.origin, + ) as span: + if agent_name: + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, agent_name) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, False) + + if tools: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, tools, unpack=False + ) + + # Run the agent + result = f(self, *args, **kwargs) + + input = result.get("input") + if ( + input is not None + and should_send_default_pii() + and integration.include_prompts + ): + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + [ + input, + ], + ) + + output = result.get("output") + if ( + output is not None + and should_send_default_pii() + and integration.include_prompts + ): + span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + + return result + + return new_invoke + + +def _wrap_agent_executor_stream(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + + @wraps(f) + def new_stream(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(LangchainIntegration) + if integration is None: + return f(self, *args, **kwargs) + + agent_name, tools = _get_request_data(self, args, kwargs) + + span = sentry_sdk.start_span( + op=OP.GEN_AI_INVOKE_AGENT, + name=f"invoke_agent {agent_name}".strip(), + origin=LangchainIntegration.origin, + ) + span.__enter__() + + if agent_name: + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, agent_name) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + if tools: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, tools, unpack=False + ) + + input = args[0].get("input") if len(args) >= 1 else None + if ( + input is not None + and should_send_default_pii() + and integration.include_prompts + ): + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + [ + input, + ], + ) + + # Run the agent + result = f(self, *args, **kwargs) + + old_iterator = result + + def new_iterator(): + # type: () -> Iterator[Any] + for event in old_iterator: + yield event + + try: + output = event.get("output") + except Exception: + output = None + + if ( + output is not None + and should_send_default_pii() + and integration.include_prompts + ): + span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + + span.__exit__(None, None, None) + + async def new_iterator_async(): + # type: () -> AsyncIterator[Any] + async for event in old_iterator: + yield event + + try: + output = event.get("output") + except Exception: + output = None + + if ( + output is not None + and should_send_default_pii() + and integration.include_prompts + ): + span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + + span.__exit__(None, None, None) + + if str(type(result)) == "": + result = new_iterator_async() + else: + result = new_iterator() + + return result + + return new_stream diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 9d55a49f82..9a06ac05d4 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -54,15 +54,7 @@ def _llm_type(self) -> str: return llm_type -def tiktoken_encoding_if_installed(): - try: - import tiktoken # type: ignore # noqa # pylint: disable=unused-import - - return "cl100k_base" - except ImportError: - return None - - +@pytest.mark.xfail @pytest.mark.parametrize( "send_default_pii, include_prompts, use_unknown_llm_type", [ @@ -82,7 +74,6 @@ def test_langchain_agent( integrations=[ LangchainIntegration( include_prompts=include_prompts, - tiktoken_encoding_name=tiktoken_encoding_if_installed(), ) ], traces_sample_rate=1.0, @@ -144,7 +135,16 @@ def test_langchain_agent( ), ChatGenerationChunk( type="ChatGenerationChunk", - message=AIMessageChunk(content="5"), + message=AIMessageChunk( + content="5", + usage_metadata={ + "input_tokens": 142, + "output_tokens": 50, + "total_tokens": 192, + "input_token_details": {"audio": 0, "cache_read": 0}, + "output_token_details": {"audio": 0, "reasoning": 0}, + }, + ), generation_info={"finish_reason": "function_call"}, ), ], @@ -152,7 +152,16 @@ def test_langchain_agent( ChatGenerationChunk( text="The word eudca has 5 letters.", type="ChatGenerationChunk", - message=AIMessageChunk(content="The word eudca has 5 letters."), + message=AIMessageChunk( + content="The word eudca has 5 letters.", + usage_metadata={ + "input_tokens": 89, + "output_tokens": 28, + "total_tokens": 117, + "input_token_details": {"audio": 0, "cache_read": 0}, + "output_token_details": {"audio": 0, "reasoning": 0}, + }, + ), ), ChatGenerationChunk( type="ChatGenerationChunk", @@ -176,42 +185,49 @@ def test_langchain_agent( tx = events[0] assert tx["type"] == "transaction" - chat_spans = list( - x for x in tx["spans"] if x["op"] == "ai.chat_completions.create.langchain" - ) - tool_exec_span = next(x for x in tx["spans"] if x["op"] == "ai.tool.langchain") + chat_spans = list(x for x in tx["spans"] if x["op"] == "gen_ai.chat") + tool_exec_span = next(x for x in tx["spans"] if x["op"] == "gen_ai.execute_tool") assert len(chat_spans) == 2 # We can't guarantee anything about the "shape" of the langchain execution graph - assert len(list(x for x in tx["spans"] if x["op"] == "ai.run.langchain")) > 0 + assert len(list(x for x in tx["spans"] if x["op"] == "gen_ai.chat")) > 0 - if use_unknown_llm_type: - assert "gen_ai.usage.input_tokens" in chat_spans[0]["data"] - assert "gen_ai.usage.total_tokens" in chat_spans[0]["data"] - else: - # important: to avoid double counting, we do *not* measure - # tokens used if we have an explicit integration (e.g. OpenAI) - assert "measurements" not in chat_spans[0] + assert "gen_ai.usage.input_tokens" in chat_spans[0]["data"] + assert "gen_ai.usage.output_tokens" in chat_spans[0]["data"] + assert "gen_ai.usage.total_tokens" in chat_spans[0]["data"] + + assert chat_spans[0]["data"]["gen_ai.usage.input_tokens"] == 142 + assert chat_spans[0]["data"]["gen_ai.usage.output_tokens"] == 50 + assert chat_spans[0]["data"]["gen_ai.usage.total_tokens"] == 192 + + assert "gen_ai.usage.input_tokens" in chat_spans[1]["data"] + assert "gen_ai.usage.output_tokens" in chat_spans[1]["data"] + assert "gen_ai.usage.total_tokens" in chat_spans[1]["data"] + assert chat_spans[1]["data"]["gen_ai.usage.input_tokens"] == 89 + assert chat_spans[1]["data"]["gen_ai.usage.output_tokens"] == 28 + assert chat_spans[1]["data"]["gen_ai.usage.total_tokens"] == 117 if send_default_pii and include_prompts: assert ( - "You are very powerful" in chat_spans[0]["data"][SPANDATA.AI_INPUT_MESSAGES] + "You are very powerful" + in chat_spans[0]["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] ) - assert "5" in chat_spans[0]["data"][SPANDATA.AI_RESPONSES] - assert "word" in tool_exec_span["data"][SPANDATA.AI_INPUT_MESSAGES] - assert 5 == int(tool_exec_span["data"][SPANDATA.AI_RESPONSES]) + assert "5" in chat_spans[0]["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + assert "word" in tool_exec_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert 5 == int(tool_exec_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT]) assert ( - "You are very powerful" in chat_spans[1]["data"][SPANDATA.AI_INPUT_MESSAGES] + "You are very powerful" + in chat_spans[1]["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] ) - assert "5" in chat_spans[1]["data"][SPANDATA.AI_RESPONSES] + assert "5" in chat_spans[1]["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] else: - assert SPANDATA.AI_INPUT_MESSAGES not in chat_spans[0].get("data", {}) - assert SPANDATA.AI_RESPONSES not in chat_spans[0].get("data", {}) - assert SPANDATA.AI_INPUT_MESSAGES not in chat_spans[1].get("data", {}) - assert SPANDATA.AI_RESPONSES not in chat_spans[1].get("data", {}) - assert SPANDATA.AI_INPUT_MESSAGES not in tool_exec_span.get("data", {}) - assert SPANDATA.AI_RESPONSES not in tool_exec_span.get("data", {}) + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in chat_spans[0].get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in chat_spans[0].get("data", {}) + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in chat_spans[1].get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in chat_spans[1].get("data", {}) + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in tool_exec_span.get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in tool_exec_span.get("data", {}) def test_langchain_error(sentry_init, capture_events): @@ -311,7 +327,16 @@ def test_span_origin(sentry_init, capture_events): ), ChatGenerationChunk( type="ChatGenerationChunk", - message=AIMessageChunk(content="5"), + message=AIMessageChunk( + content="5", + usage_metadata={ + "input_tokens": 142, + "output_tokens": 50, + "total_tokens": 192, + "input_token_details": {"audio": 0, "cache_read": 0}, + "output_token_details": {"audio": 0, "reasoning": 0}, + }, + ), generation_info={"finish_reason": "function_call"}, ), ], @@ -319,7 +344,16 @@ def test_span_origin(sentry_init, capture_events): ChatGenerationChunk( text="The word eudca has 5 letters.", type="ChatGenerationChunk", - message=AIMessageChunk(content="The word eudca has 5 letters."), + message=AIMessageChunk( + content="The word eudca has 5 letters.", + usage_metadata={ + "input_tokens": 89, + "output_tokens": 28, + "total_tokens": 117, + "input_token_details": {"audio": 0, "cache_read": 0}, + "output_token_details": {"audio": 0, "reasoning": 0}, + }, + ), ), ChatGenerationChunk( type="ChatGenerationChunk", From 71f61af19282d68ed2e8f7b6009adb36db9d720e Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 14 Aug 2025 16:47:40 +0000 Subject: [PATCH 408/591] release: 2.35.0 --- CHANGELOG.md | 24 ++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 27 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 21b1d5fec9..d3fe6de4e4 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,29 @@ # Changelog +## 2.35.0 + +### Various fixes & improvements + +- feat(langchain): update integration to from ai.* to gen_ai.* span attributes (#4678) by @shellmayr +- build(deps): bump actions/create-github-app-token from 2.0.6 to 2.1.0 (#4684) by @dependabot +- feat(tracing): AI Agents templates for `@trace` decorator (#4676) by @antonpirker +- Remove performance papercuts (#4675) by @sentrivana +- ref(gnu-integration): make path optional (#4688) by @MeredithAnya +- Update tox.ini (#4689) by @sentrivana +- Fix Redis CI (#4691) by @sentrivana +- Help for debugging Cron problems (#4686) by @antonpirker +- feat(anthropic) Update span attributes to use `gen_ai.*` namespace instead of `ai.*` (#4674) by @constantinius +- fix(clickhouse): Don't eat the generator data (#4669) by @szokeasaurusrex +- ref(clickhouse): List `send_data` parameters (#4667) by @szokeasaurusrex +- feat(tracing): Improve `@trace` decorator. (#4648) by @antonpirker +- feat(tracing): Add convenience function `update_current_span`. (#4673) by @antonpirker +- Update `gen_ai.*` and `ai.*` attributes (#4665) by @antonpirker +- Add `update_data` to `Span`. (#4666) by @antonpirker +- Fix plugins key codecov (#4655) by @sl0thentr0py +- Add `enable_logs`, `before_send_log` as top-level options (#4644) by @sentrivana +- Fix mypy (#4649) by @sentrivana +- Better checking for empty tools list (#4647) by @antonpirker + ## 2.34.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index f5d0b9e121..465e29a4e8 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.34.1" +release = "2.35.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index a290697659..f307e526af 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1329,4 +1329,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.34.1" +VERSION = "2.35.0" diff --git a/setup.py b/setup.py index 11b02cbca8..dd91f8bb37 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.34.1", + version="2.35.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 895c3cd600aee70c7aed68e017f4e082bdaeccfe Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 14 Aug 2025 19:01:46 +0200 Subject: [PATCH 409/591] updated changelog --- CHANGELOG.md | 34 +++++++++++++++++----------------- 1 file changed, 17 insertions(+), 17 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index d3fe6de4e4..d90d46abe0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,25 +4,25 @@ ### Various fixes & improvements -- feat(langchain): update integration to from ai.* to gen_ai.* span attributes (#4678) by @shellmayr -- build(deps): bump actions/create-github-app-token from 2.0.6 to 2.1.0 (#4684) by @dependabot -- feat(tracing): AI Agents templates for `@trace` decorator (#4676) by @antonpirker -- Remove performance papercuts (#4675) by @sentrivana -- ref(gnu-integration): make path optional (#4688) by @MeredithAnya -- Update tox.ini (#4689) by @sentrivana -- Fix Redis CI (#4691) by @sentrivana -- Help for debugging Cron problems (#4686) by @antonpirker -- feat(anthropic) Update span attributes to use `gen_ai.*` namespace instead of `ai.*` (#4674) by @constantinius -- fix(clickhouse): Don't eat the generator data (#4669) by @szokeasaurusrex -- ref(clickhouse): List `send_data` parameters (#4667) by @szokeasaurusrex -- feat(tracing): Improve `@trace` decorator. (#4648) by @antonpirker -- feat(tracing): Add convenience function `update_current_span`. (#4673) by @antonpirker +- [Langchain Integration](https://docs.sentry.io/platforms/python/integrations/langchain/) now supports Sentry [AI Insights dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4678) by @shellmayr +- [Anthropic Integration](https://docs.sentry.io/platforms/python/integrations/anthropic/) now supports Sentry [AI Insights dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4674) by @constantinius +- AI Agents templates for `@trace` decorator (#4676) by @antonpirker +- Sentry Logs: Add `enable_logs`, `before_send_log` as top-level `sentry_sdk.init()` options (#4644) by @sentrivana +- Tracing: Improve `@trace` decorator. Allows to set `span.op`, `span.name`, and `span.attributes` (#4648) by @antonpirker +- Tracing: Add convenience function `sentry_sdk.update_current_span`. (#4673) by @antonpirker +- Tracing: Add `Span.update_data()` to update multiple `span.data` items at once. (#4666) by @antonpirker +- GNU-integration: make path optional (#4688) by @MeredithAnya +- Clickhouse: Don't eat the generator data (#4669) by @szokeasaurusrex +- Clickhouse: List `send_data` parameters (#4667) by @szokeasaurusrex - Update `gen_ai.*` and `ai.*` attributes (#4665) by @antonpirker -- Add `update_data` to `Span`. (#4666) by @antonpirker -- Fix plugins key codecov (#4655) by @sl0thentr0py -- Add `enable_logs`, `before_send_log` as top-level options (#4644) by @sentrivana -- Fix mypy (#4649) by @sentrivana - Better checking for empty tools list (#4647) by @antonpirker +- Remove performance paper cuts (#4675) by @sentrivana +- Help for debugging Cron problems (#4686) by @antonpirker +- Fix Redis CI (#4691) by @sentrivana +- Fix plugins key codecov (#4655) by @sl0thentr0py +- Fix Mypy (#4649) by @sentrivana +- Update tox.ini (#4689) by @sentrivana +- build(deps): bump actions/create-github-app-token from 2.0.6 to 2.1.0 (#4684) by @dependabot ## 2.34.1 From 5eafb784e8e71ba9b4d7c16b42cea18448944d5d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 14 Aug 2025 19:04:30 +0200 Subject: [PATCH 410/591] update changelog --- CHANGELOG.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index d90d46abe0..6e06e61e32 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,8 +4,8 @@ ### Various fixes & improvements -- [Langchain Integration](https://docs.sentry.io/platforms/python/integrations/langchain/) now supports Sentry [AI Insights dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4678) by @shellmayr -- [Anthropic Integration](https://docs.sentry.io/platforms/python/integrations/anthropic/) now supports Sentry [AI Insights dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4674) by @constantinius +- [Langchain Integration](https://docs.sentry.io/platforms/python/integrations/langchain/) now supports the Sentry [AI dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4678) by @shellmayr +- [Anthropic Integration](https://docs.sentry.io/platforms/python/integrations/anthropic/) now supports the Sentry [AI dashboard](https://docs.sentry.io/product/insights/ai/agents/dashboard/). (#4674) by @constantinius - AI Agents templates for `@trace` decorator (#4676) by @antonpirker - Sentry Logs: Add `enable_logs`, `before_send_log` as top-level `sentry_sdk.init()` options (#4644) by @sentrivana - Tracing: Improve `@trace` decorator. Allows to set `span.op`, `span.name`, and `span.attributes` (#4648) by @antonpirker From 8787ee4d0e1250c3ff1f85aca215c0151873c089 Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Wed, 20 Aug 2025 00:53:14 -0400 Subject: [PATCH 411/591] fix(tracing): Do not attach stacktrace to transaction (#4713) The `attach_stacktrace` option was attaching stack traces to transactions. This is an expensive operation but the results aren't used anywhere. --- sentry_sdk/client.py | 6 ++++-- tests/test_client.py | 10 ++++++++++ 2 files changed, 14 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 5d584a5537..c45d5e2f4f 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -516,8 +516,9 @@ def _prepare_event( if event.get("timestamp") is None: event["timestamp"] = datetime.now(timezone.utc) + is_transaction = event.get("type") == "transaction" + if scope is not None: - is_transaction = event.get("type") == "transaction" spans_before = len(cast(List[Dict[str, object]], event.get("spans", []))) event_ = scope.apply_to_event(event, hint, self.options) @@ -560,7 +561,8 @@ def _prepare_event( ) if ( - self.options["attach_stacktrace"] + not is_transaction + and self.options["attach_stacktrace"] and "exception" not in event and "stacktrace" not in event and "threads" not in event diff --git a/tests/test_client.py b/tests/test_client.py index 0468fcbb7b..a02ea6e56a 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -21,6 +21,7 @@ capture_exception, capture_event, set_tag, + start_transaction, ) from sentry_sdk.spotlight import DEFAULT_SPOTLIGHT_URL from sentry_sdk.utils import capture_internal_exception @@ -562,6 +563,15 @@ def test_attach_stacktrace_disabled(sentry_init, capture_events): assert "threads" not in event +def test_attach_stacktrace_transaction(sentry_init, capture_events): + sentry_init(traces_sample_rate=1.0, attach_stacktrace=True) + events = capture_events() + with start_transaction(name="transaction"): + pass + (event,) = events + assert "threads" not in event + + def test_capture_event_works(sentry_init): sentry_init(transport=_TestTransport()) pytest.raises(EnvelopeCapturedError, lambda: capture_event({})) From 9e154f7c15934fc14c091d5abc4e729dcaa374a9 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 25 Aug 2025 12:39:59 +0200 Subject: [PATCH 412/591] build(deps): bump actions/create-github-app-token from 2.1.0 to 2.1.1 (#4710) Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 2.1.0 to 2.1.1.
Release notes

Sourced from actions/create-github-app-token's releases.

v2.1.1

2.1.1 (2025-08-11)

Bug Fixes

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=2.1.0&new-version=2.1.1)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 6197f9023d..066c58595d 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@0f859bf9e69e887678d5bbfbee594437cb440ffe # v2.1.0 + uses: actions/create-github-app-token@a8d616148505b5069dccd32f177bb87d7f39123b # v2.1.1 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From b8248a39a27374179c7b7f03c0aca1678dead01c Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 25 Aug 2025 10:42:55 +0000 Subject: [PATCH 413/591] build(deps): bump codecov/codecov-action from 5.4.3 to 5.5.0 (#4717) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.4.3 to 5.5.0.
Release notes

Sourced from codecov/codecov-action's releases.

v5.5.0

What's Changed

New Contributors

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.3...v5.5.0

Changelog

Sourced from codecov/codecov-action's changelog.

v5.5.0

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.4.3..v5.5.0

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=codecov/codecov-action&package-manager=github_actions&previous-version=5.4.3&new-version=5.5.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 13 files changed, 19 insertions(+), 19 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index dd57f5909b..702496acd9 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index e79c9513ef..c64c955855 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -166,7 +166,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index c7e356420c..dc46d8d475 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 6c203379fe..aa938a3ccb 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -107,7 +107,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -206,7 +206,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 926465990d..64529064e9 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index a08e91c909..f8babbecee 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 9bbeee6c6a..454bc1d5ea 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 3595640ce1..b049ad5642 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 3ac5508dab..a79dc0dd2b 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -75,7 +75,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 13c34224be..868d43a6f0 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -102,7 +102,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -196,7 +196,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index e52a903208..87c0054362 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index c703cfafce..a991d4f84f 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 96faefc54e..b81e964a18 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -100,7 +100,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.4.3 + uses: codecov/codecov-action@v5.5.0 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From a2a9413372130a90247a59fe1b275d508258f926 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 25 Aug 2025 10:43:56 +0000 Subject: [PATCH 414/591] build(deps): bump actions/setup-java from 4 to 5 (#4716) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/setup-java](https://github.com/actions/setup-java) from 4 to 5.
Release notes

Sourced from actions/setup-java's releases.

v5.0.0

What's Changed

Breaking Changes

Make sure your runner is updated to this version or newer to use this release. v2.327.1 Release Notes

Dependency Upgrades

Bug Fixes

New Contributors

Full Changelog: https://github.com/actions/setup-java/compare/v4...v5.0.0

v4.7.1

What's Changed

Documentation changes

Dependency updates:

Full Changelog: https://github.com/actions/setup-java/compare/v4...v4.7.1

v4.7.0

What's Changed

... (truncated)

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/setup-java&package-manager=github_actions&previous-version=4&new-version=5)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-tasks.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 868d43a6f0..fa8e405d7f 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -47,7 +47,7 @@ jobs: - name: Start Redis uses: supercharge/redis-github-action@1.8.0 - name: Install Java - uses: actions/setup-java@v4 + uses: actions/setup-java@v5 with: distribution: 'temurin' java-version: '21' @@ -141,7 +141,7 @@ jobs: - name: Start Redis uses: supercharge/redis-github-action@1.8.0 - name: Install Java - uses: actions/setup-java@v4 + uses: actions/setup-java@v5 with: distribution: 'temurin' java-version: '21' diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index b81e964a18..9c30cd1a75 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -61,7 +61,7 @@ {% if needs_java %} - name: Install Java - uses: actions/setup-java@v4 + uses: actions/setup-java@v5 with: distribution: 'temurin' java-version: '21' From 28d0dddf41b7c10f9ba056aee659d2da4d490fbe Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 25 Aug 2025 10:46:36 +0000 Subject: [PATCH 415/591] build(deps): bump actions/checkout from 4.2.2 to 5.0.0 (#4709) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/checkout](https://github.com/actions/checkout) from 4.2.2 to 5.0.0.
Release notes

Sourced from actions/checkout's releases.

v5.0.0

What's Changed

⚠️ Minimum Compatible Runner Version

v2.327.1
Release Notes

Make sure your runner is updated to this version or newer to use this release.

Full Changelog: https://github.com/actions/checkout/compare/v4...v5.0.0

v4.3.0

What's Changed

New Contributors

Full Changelog: https://github.com/actions/checkout/compare/v4...v4.3.0

Changelog

Sourced from actions/checkout's changelog.

V5.0.0

V4.3.0

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/checkout&package-manager=github_actions&previous-version=4.2.2&new-version=5.0.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/ci.yml | 8 ++++---- .github/workflows/codeql-analysis.yml | 2 +- .github/workflows/release.yml | 2 +- .github/workflows/test-integrations-ai.yml | 4 ++-- .github/workflows/test-integrations-cloud.yml | 4 ++-- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 16 files changed, 25 insertions(+), 25 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 03ed8de742..ffc0a741fc 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -24,7 +24,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -39,7 +39,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -58,7 +58,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 with: python-version: 3.12 @@ -89,7 +89,7 @@ jobs: timeout-minutes: 10 steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 with: python-version: 3.12 diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index d824757ee9..74664add46 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -48,7 +48,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v4.2.2 + uses: actions/checkout@v5.0.0 # Initializes the CodeQL tools for scanning. - name: Initialize CodeQL diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 066c58595d..f5e952d0de 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -24,7 +24,7 @@ jobs: with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 with: token: ${{ steps.token.outputs.token }} fetch-depth: 0 diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 702496acd9..a784f9fc47 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -117,7 +117,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index c64c955855..a04d57497a 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -42,7 +42,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -121,7 +121,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index dc46d8d475..1c0c9b80d2 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index aa938a3ccb..5fc0be029b 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -56,7 +56,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -155,7 +155,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 64529064e9..f744f514ee 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index f8babbecee..382e6a5f15 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 454bc1d5ea..93675fb4fe 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index b049ad5642..e8937708bc 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index a79dc0dd2b..867681d3a3 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -105,7 +105,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index fa8e405d7f..a489f64410 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -132,7 +132,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 87c0054362..ba802faa01 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -56,7 +56,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index a991d4f84f..e79a54ef67 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -38,7 +38,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: @@ -133,7 +133,7 @@ jobs: # Use Docker container only for Python 3.6 container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 if: ${{ matrix.python-version != '3.6' }} with: diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 9c30cd1a75..4ac0d03eb2 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -42,7 +42,7 @@ # Use Docker container only for Python 3.6 {% raw %}container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }}{% endraw %} steps: - - uses: actions/checkout@v4.2.2 + - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v5 {% raw %}if: ${{ matrix.python-version != '3.6' }}{% endraw %} with: From eee4c4b0186c8aed964151a8e2af56420b7ad288 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 26 Aug 2025 09:53:37 +0200 Subject: [PATCH 416/591] fix(openai-agents): Isolate agent run (#4720) It looks like we're running into https://github.com/getsentry/sentry-python/issues/4718 (and probably https://github.com/getsentry/sentry-python/issues/4690) because the different agent runs are not properly isolated. This only seems to be a problem when multiple agent runs are awaited at once (e.g. via `asyncio.gather`) -- it seems that leads to some scope bleed. ```python import asyncio import sentry_sdk from agents import Agent, Runner from sentry_sdk.integrations.asyncio import AsyncioIntegration from sentry_sdk.integrations.openai_agents import OpenAIAgentsIntegration sentry_sdk.init(...) main_agent = Agent( name="main_agent", model="gpt-5", ) async def run_agent() -> None: runner = await Runner.run( starting_agent=main_agent, input="How are you?", ) print(runner.final_output) async def main() -> None: await asyncio.gather(*[run_agent() for _ in range(2)]) # throws an error # await asyncio.gather(run_agent()) # works ``` --- .../openai_agents/patches/agent_run.py | 3 -- .../openai_agents/patches/runner.py | 33 +++++++------- .../openai_agents/test_openai_agents.py | 43 +++++++++++++++++++ 3 files changed, 61 insertions(+), 18 deletions(-) diff --git a/sentry_sdk/integrations/openai_agents/patches/agent_run.py b/sentry_sdk/integrations/openai_agents/patches/agent_run.py index 084100878c..29002f6619 100644 --- a/sentry_sdk/integrations/openai_agents/patches/agent_run.py +++ b/sentry_sdk/integrations/openai_agents/patches/agent_run.py @@ -1,7 +1,6 @@ from functools import wraps from sentry_sdk.integrations import DidNotEnable - from ..spans import invoke_agent_span, update_invoke_agent_span, handoff_span from typing import TYPE_CHECKING @@ -9,7 +8,6 @@ if TYPE_CHECKING: from typing import Any, Optional - try: import agents except ImportError: @@ -62,7 +60,6 @@ def _get_current_agent(context_wrapper): async def patched_run_single_turn(cls, *args, **kwargs): # type: (agents.Runner, *Any, **Any) -> Any """Patched _run_single_turn that creates agent invocation spans""" - agent = kwargs.get("agent") context_wrapper = kwargs.get("context_wrapper") should_run_agent_start_hooks = kwargs.get("should_run_agent_start_hooks") diff --git a/sentry_sdk/integrations/openai_agents/patches/runner.py b/sentry_sdk/integrations/openai_agents/patches/runner.py index e1e9a3b50c..745f30a38e 100644 --- a/sentry_sdk/integrations/openai_agents/patches/runner.py +++ b/sentry_sdk/integrations/openai_agents/patches/runner.py @@ -23,20 +23,23 @@ def _create_run_wrapper(original_func): @wraps(original_func) async def wrapper(*args, **kwargs): # type: (*Any, **Any) -> Any - agent = args[0] - with agent_workflow_span(agent): - result = None - try: - result = await original_func(*args, **kwargs) - return result - except Exception as exc: - _capture_exception(exc) - - # It could be that there is a "invoke agent" span still open - current_span = sentry_sdk.get_current_span() - if current_span is not None and current_span.timestamp is None: - current_span.__exit__(None, None, None) - - raise exc from None + # Isolate each workflow so that when agents are run in asyncio tasks they + # don't touch each other's scopes + with sentry_sdk.isolation_scope(): + agent = args[0] + with agent_workflow_span(agent): + result = None + try: + result = await original_func(*args, **kwargs) + return result + except Exception as exc: + _capture_exception(exc) + + # It could be that there is a "invoke agent" span still open + current_span = sentry_sdk.get_current_span() + if current_span is not None and current_span.timestamp is None: + current_span.__exit__(None, None, None) + + raise exc from None return wrapper diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index 3f64e5c45c..09fca2fbf3 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -1,3 +1,4 @@ +import asyncio import re import pytest from unittest.mock import MagicMock, patch @@ -637,3 +638,45 @@ async def test_error_handling(sentry_init, capture_events, test_agent): assert ai_client_span["description"] == "chat gpt-4" assert ai_client_span["origin"] == "auto.ai.openai_agents" assert ai_client_span["tags"]["status"] == "internal_error" + + +@pytest.mark.asyncio +async def test_multiple_agents_asyncio( + sentry_init, capture_events, test_agent, mock_model_response +): + """ + Test that multiple agents can be run at the same time in asyncio tasks + without interfering with each other. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.return_value = mock_model_response + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + async def run(): + await agents.Runner.run( + starting_agent=test_agent, + input="Test input", + run_config=test_run_config, + ) + + await asyncio.gather(*[run() for _ in range(3)]) + + assert len(events) == 3 + txn1, txn2, txn3 = events + + assert txn1["type"] == "transaction" + assert txn1["transaction"] == "test_agent workflow" + assert txn2["type"] == "transaction" + assert txn2["transaction"] == "test_agent workflow" + assert txn3["type"] == "transaction" + assert txn3["transaction"] == "test_agent workflow" From c2a21aada390aa28ca4ccf7880fe1e5fde31ac52 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 26 Aug 2025 07:59:20 +0000 Subject: [PATCH 417/591] release: 2.35.1 --- CHANGELOG.md | 11 +++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 14 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 6e06e61e32..c88533ebe3 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,16 @@ # Changelog +## 2.35.1 + +### Various fixes & improvements + +- fix(openai-agents): Isolate agent run (#4720) by @sentrivana +- build(deps): bump actions/checkout from 4.2.2 to 5.0.0 (#4709) by @dependabot +- build(deps): bump actions/setup-java from 4 to 5 (#4716) by @dependabot +- build(deps): bump codecov/codecov-action from 5.4.3 to 5.5.0 (#4717) by @dependabot +- build(deps): bump actions/create-github-app-token from 2.1.0 to 2.1.1 (#4710) by @dependabot +- fix(tracing): Do not attach stacktrace to transaction (#4713) by @Zylphrex + ## 2.35.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 465e29a4e8..7ad137b9ed 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.35.0" +release = "2.35.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index f307e526af..2d3ab230b6 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1329,4 +1329,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.35.0" +VERSION = "2.35.1" diff --git a/setup.py b/setup.py index dd91f8bb37..f16f4e3fd0 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.35.0", + version="2.35.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From fb4faf6090bae29000a1b8c4cd07dee0d25a59f4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 26 Aug 2025 10:08:00 +0200 Subject: [PATCH 418/591] Update CHANGELOG.md --- CHANGELOG.md | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index c88533ebe3..a3af3f63a0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,12 +4,8 @@ ### Various fixes & improvements -- fix(openai-agents): Isolate agent run (#4720) by @sentrivana -- build(deps): bump actions/checkout from 4.2.2 to 5.0.0 (#4709) by @dependabot -- build(deps): bump actions/setup-java from 4 to 5 (#4716) by @dependabot -- build(deps): bump codecov/codecov-action from 5.4.3 to 5.5.0 (#4717) by @dependabot -- build(deps): bump actions/create-github-app-token from 2.1.0 to 2.1.1 (#4710) by @dependabot -- fix(tracing): Do not attach stacktrace to transaction (#4713) by @Zylphrex +- OpenAI Agents: Isolate agent run (#4720) by @sentrivana +- Tracing: Do not attach stacktrace to transaction (#4713) by @Zylphrex ## 2.35.0 From 57a340568fddd0c566ba59d01f80d747e6c19a7f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 26 Aug 2025 11:10:07 +0200 Subject: [PATCH 419/591] Update tox.ini (#4721) Regular update --- .../openai_agents/test_openai_agents.py | 2 +- tox.ini | 72 ++++++++++--------- 2 files changed, 38 insertions(+), 36 deletions(-) diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index 09fca2fbf3..a3075e6415 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -87,7 +87,7 @@ def test_agent_custom_model(): name="test_agent_custom_model", instructions="You are a helpful test assistant.", # the model could be agents.OpenAIChatCompletionsModel() - model=MagicMock(model="my-custom-model"), + model="my-custom-model", model_settings=ModelSettings( max_tokens=100, temperature=0.7, diff --git a/tox.ini b/tox.ini index a1b1327af5..bbc1d57c12 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-08-12T07:16:34.585160+00:00 +# Last generated: 2025-08-26T08:59:42.512502+00:00 [tox] requires = @@ -136,18 +136,18 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.31.2 - {py3.8,py3.11,py3.12}-anthropic-v0.46.0 - {py3.8,py3.12,py3.13}-anthropic-v0.62.0 + {py3.8,py3.11,py3.12}-anthropic-v0.32.0 + {py3.8,py3.11,py3.12}-anthropic-v0.48.0 + {py3.8,py3.12,py3.13}-anthropic-v0.64.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 {py3.9,py3.11,py3.12}-cohere-v5.13.12 - {py3.9,py3.11,py3.12}-cohere-v5.16.3 + {py3.9,py3.11,py3.12}-cohere-v5.17.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 - {py3.10,py3.12,py3.13}-openai_agents-v0.2.6 + {py3.10,py3.12,py3.13}-openai_agents-v0.2.9 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 @@ -162,7 +162,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.14.0 + {py3.9,py3.12,py3.13}-pymongo-v4.14.1 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -183,9 +183,9 @@ envlist = {py3.9,py3.12,py3.13}-openfeature-v0.8.2 {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.57.3 - {py3.7,py3.12,py3.13}-statsig-v0.59.1 + {py3.7,py3.12,py3.13}-statsig-v0.58.4 {py3.7,py3.12,py3.13}-statsig-v0.61.0 + {py3.7,py3.12,py3.13}-statsig-v0.63.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -201,15 +201,16 @@ envlist = {py3.6,py3.9,py3.10}-gql-v3.4.1 {py3.7,py3.11,py3.12}-gql-v3.5.3 - {py3.9,py3.12,py3.13}-gql-v4.0.0b0 + {py3.9,py3.12,py3.13}-gql-v4.0.0 + {py3.9,py3.12,py3.13}-gql-v4.1.0b0 {py3.6,py3.9,py3.10}-graphene-v3.3 {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.232.2 - {py3.8,py3.12,py3.13}-strawberry-v0.255.0 - {py3.9,py3.12,py3.13}-strawberry-v0.278.1 + {py3.8,py3.11,py3.12}-strawberry-v0.233.3 + {py3.9,py3.12,py3.13}-strawberry-v0.257.0 + {py3.9,py3.12,py3.13}-strawberry-v0.280.0 # ~~~ Network ~~~ @@ -250,12 +251,12 @@ envlist = {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 {py3.8,py3.12,py3.13}-flask-v3.0.3 - {py3.9,py3.12,py3.13}-flask-v3.1.1 + {py3.9,py3.12,py3.13}-flask-v3.1.2 {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.26.1 {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.47.2 + {py3.9,py3.12,py3.13}-starlette-v0.47.3 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 @@ -309,7 +310,7 @@ envlist = {py3.9,py3.12,py3.13}-trytond-v7.6.5 {py3.7,py3.12,py3.13}-typer-v0.15.4 - {py3.7,py3.12,py3.13}-typer-v0.16.0 + {py3.7,py3.12,py3.13}-typer-v0.16.1 @@ -511,22 +512,22 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.31.2: anthropic==0.31.2 - anthropic-v0.46.0: anthropic==0.46.0 - anthropic-v0.62.0: anthropic==0.62.0 + anthropic-v0.32.0: anthropic==0.32.0 + anthropic-v0.48.0: anthropic==0.48.0 + anthropic-v0.64.0: anthropic==0.64.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.31.2: httpx<0.28.0 - anthropic-v0.46.0: httpx<0.28.0 + anthropic-v0.32.0: httpx<0.28.0 + anthropic-v0.48.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 cohere-v5.13.12: cohere==5.13.12 - cohere-v5.16.3: cohere==5.16.3 + cohere-v5.17.0: cohere==5.17.0 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 - openai_agents-v0.2.6: openai-agents==0.2.6 + openai_agents-v0.2.9: openai-agents==0.2.9 openai_agents: pytest-asyncio huggingface_hub-v0.22.2: huggingface_hub==0.22.2 @@ -542,7 +543,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.14.0: pymongo==4.14.0 + pymongo-v4.14.1: pymongo==4.14.1 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -564,9 +565,9 @@ deps = openfeature-v0.8.2: openfeature-sdk==0.8.2 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.57.3: statsig==0.57.3 - statsig-v0.59.1: statsig==0.59.1 + statsig-v0.58.4: statsig==0.58.4 statsig-v0.61.0: statsig==0.61.0 + statsig-v0.63.0: statsig==0.63.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -586,7 +587,8 @@ deps = gql-v3.4.1: gql[all]==3.4.1 gql-v3.5.3: gql[all]==3.5.3 - gql-v4.0.0b0: gql[all]==4.0.0b0 + gql-v4.0.0: gql[all]==4.0.0 + gql-v4.1.0b0: gql[all]==4.1.0b0 graphene-v3.3: graphene==3.3 graphene-v3.4.3: graphene==3.4.3 @@ -597,13 +599,13 @@ deps = py3.6-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.232.2: strawberry-graphql[fastapi,flask]==0.232.2 - strawberry-v0.255.0: strawberry-graphql[fastapi,flask]==0.255.0 - strawberry-v0.278.1: strawberry-graphql[fastapi,flask]==0.278.1 + strawberry-v0.233.3: strawberry-graphql[fastapi,flask]==0.233.3 + strawberry-v0.257.0: strawberry-graphql[fastapi,flask]==0.257.0 + strawberry-v0.280.0: strawberry-graphql[fastapi,flask]==0.280.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.232.2: pydantic<2.11 - strawberry-v0.255.0: pydantic<2.11 + strawberry-v0.233.3: pydantic<2.11 + strawberry-v0.257.0: pydantic<2.11 # ~~~ Network ~~~ @@ -673,7 +675,7 @@ deps = flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 flask-v3.0.3: flask==3.0.3 - flask-v3.1.1: flask==3.1.1 + flask-v3.1.2: flask==3.1.2 flask: flask-login flask: werkzeug flask-v1.1.4: werkzeug<2.1.0 @@ -682,7 +684,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.26.1: starlette==0.26.1 starlette-v0.36.3: starlette==0.36.3 - starlette-v0.47.2: starlette==0.47.2 + starlette-v0.47.3: starlette==0.47.3 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -779,7 +781,7 @@ deps = trytond-v4.8.18: werkzeug<1.0 typer-v0.15.4: typer==0.15.4 - typer-v0.16.0: typer==0.16.0 + typer-v0.16.1: typer==0.16.1 From bf4d921a5d779076e693545652359b4c0668f384 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 29 Aug 2025 14:20:30 +0200 Subject: [PATCH 420/591] fix(logs): Do not attach template if there are no parameters (#4728) Closes https://github.com/getsentry/sentry-python/issues/4725 --- sentry_sdk/integrations/logging.py | 15 ++++++++++++-- sentry_sdk/logger.py | 7 ++++--- tests/integrations/logging/test_logging.py | 19 +++++++++++++++++ tests/integrations/loguru/test_loguru.py | 18 ++++++++++++++++ tests/test_logs.py | 24 ++++++++++++++++++++++ 5 files changed, 78 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index 15ff2ed233..bfb30fc67b 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -356,12 +356,14 @@ def _capture_log_from_record(self, client, record): record.levelno, SEVERITY_TO_OTEL_SEVERITY ) project_root = client.options["project_root"] + attrs = self._extra_from_record(record) # type: Any attrs["sentry.origin"] = "auto.logger.log" - if isinstance(record.msg, str): - attrs["sentry.message.template"] = record.msg + + parameters_set = False if record.args is not None: if isinstance(record.args, tuple): + parameters_set = bool(record.args) for i, arg in enumerate(record.args): attrs[f"sentry.message.parameter.{i}"] = ( arg @@ -369,19 +371,28 @@ def _capture_log_from_record(self, client, record): else safe_repr(arg) ) elif isinstance(record.args, dict): + parameters_set = bool(record.args) for key, value in record.args.items(): attrs[f"sentry.message.parameter.{key}"] = ( value if isinstance(value, (str, float, int, bool)) else safe_repr(value) ) + + if parameters_set and isinstance(record.msg, str): + # only include template if there is at least one + # sentry.message.parameter.X set + attrs["sentry.message.template"] = record.msg + if record.lineno: attrs["code.line.number"] = record.lineno + if record.pathname: if project_root is not None and record.pathname.startswith(project_root): attrs["code.file.path"] = record.pathname[len(project_root) + 1 :] else: attrs["code.file.path"] = record.pathname + if record.funcName: attrs["code.function.name"] = record.funcName diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index c18cf91ff2..bc98f35155 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -22,13 +22,14 @@ def _capture_log(severity_text, severity_number, template, **kwargs): # type: (str, int, str, **Any) -> None client = get_client() - attrs = { - "sentry.message.template": template, - } # type: dict[str, str | bool | float | int] + attrs = {} # type: dict[str, str | bool | float | int] if "attributes" in kwargs: attrs.update(kwargs.pop("attributes")) for k, v in kwargs.items(): attrs[f"sentry.message.parameter.{k}"] = v + if kwargs: + # only attach template if there are parameters + attrs["sentry.message.template"] = template attrs = { k: ( diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 7ecdf42500..7a00ceadd2 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -571,3 +571,22 @@ def test_sentry_logs_named_parameters_complex_values(sentry_init, capture_envelo assert isinstance(complex_param, str) assert "nested" in complex_param assert "data" in complex_param + + +def test_sentry_logs_no_parameters_no_template(sentry_init, capture_envelopes): + """ + There shouldn't be a template if there are no parameters. + """ + sentry_init(enable_logs=True) + envelopes = capture_envelopes() + + python_logger = logging.Logger("test-logger") + python_logger.warning("Warning about something without any parameters.") + + get_client().flush() + logs = envelopes_to_logs(envelopes) + + assert len(logs) == 1 + + attrs = logs[0]["attributes"] + assert "sentry.message.template" not in attrs diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 38093d24cb..3d04d7d1ea 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -467,3 +467,21 @@ def test_logger_with_all_attributes( "sentry.severity_number": 13, "sentry.severity_text": "warn", } + + +def test_no_parameters_no_template( + sentry_init, capture_envelopes, uninstall_integration, request +): + uninstall_integration("loguru") + request.addfinalizer(logger.remove) + + sentry_init(enable_logs=True) + envelopes = capture_envelopes() + + logger.warning("Logging a hardcoded warning") + sentry_sdk.get_client().flush() + + logs = envelopes_to_logs(envelopes) + + attributes = logs[0]["attributes"] + assert "sentry.message.template" not in attributes diff --git a/tests/test_logs.py b/tests/test_logs.py index b2578d83d5..596a31922e 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -254,30 +254,54 @@ def test_logs_message_params(sentry_init, capture_envelopes): sentry_sdk.logger.error( "The recorded error was '{error}'", error=Exception("some error") ) + sentry_sdk.logger.warning("The recorded value was hardcoded.") get_client().flush() logs = envelopes_to_logs(envelopes) assert logs[0]["body"] == "The recorded value was '1'" assert logs[0]["attributes"]["sentry.message.parameter.int_var"] == 1 + assert ( + logs[0]["attributes"]["sentry.message.template"] + == "The recorded value was '{int_var}'" + ) assert logs[1]["body"] == "The recorded value was '2.0'" assert logs[1]["attributes"]["sentry.message.parameter.float_var"] == 2.0 + assert ( + logs[1]["attributes"]["sentry.message.template"] + == "The recorded value was '{float_var}'" + ) assert logs[2]["body"] == "The recorded value was 'False'" assert logs[2]["attributes"]["sentry.message.parameter.bool_var"] is False + assert ( + logs[2]["attributes"]["sentry.message.template"] + == "The recorded value was '{bool_var}'" + ) assert logs[3]["body"] == "The recorded value was 'some string value'" assert ( logs[3]["attributes"]["sentry.message.parameter.string_var"] == "some string value" ) + assert ( + logs[3]["attributes"]["sentry.message.template"] + == "The recorded value was '{string_var}'" + ) assert logs[4]["body"] == "The recorded error was 'some error'" assert ( logs[4]["attributes"]["sentry.message.parameter.error"] == "Exception('some error')" ) + assert ( + logs[4]["attributes"]["sentry.message.template"] + == "The recorded error was '{error}'" + ) + + assert logs[5]["body"] == "The recorded value was hardcoded." + assert "sentry.message.template" not in logs[5]["attributes"] @minimum_python_37 From d2cb532459b3e954edbee3605af5e631749c547d Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 1 Sep 2025 09:56:39 +0000 Subject: [PATCH 421/591] release: 2.35.2 --- CHANGELOG.md | 7 +++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 10 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a3af3f63a0..92cd4573b0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,12 @@ # Changelog +## 2.35.2 + +### Various fixes & improvements + +- fix(logs): Do not attach template if there are no parameters (#4728) by @sentrivana +- Update tox.ini (#4721) by @sentrivana + ## 2.35.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 7ad137b9ed..0863980aac 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.35.1" +release = "2.35.2" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 2d3ab230b6..d7a0603a10 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1329,4 +1329,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.35.1" +VERSION = "2.35.2" diff --git a/setup.py b/setup.py index f16f4e3fd0..ecb24290c8 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.35.1", + version="2.35.2", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 814cd5959b58350cb81fa8b21502fcdfe3adf960 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 1 Sep 2025 12:01:32 +0200 Subject: [PATCH 422/591] Update CHANGELOG.md --- CHANGELOG.md | 1 - 1 file changed, 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 92cd4573b0..19f734976f 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,7 +5,6 @@ ### Various fixes & improvements - fix(logs): Do not attach template if there are no parameters (#4728) by @sentrivana -- Update tox.ini (#4721) by @sentrivana ## 2.35.1 From 173df643f8847e8e0bb0a59da7cda0a019af283e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 1 Sep 2025 15:19:24 +0200 Subject: [PATCH 423/591] Update tox.ini (#4731) Regular update --- tox.ini | 14 +++++++++----- 1 file changed, 9 insertions(+), 5 deletions(-) diff --git a/tox.ini b/tox.ini index bbc1d57c12..0dbcef2c64 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-08-26T08:59:42.512502+00:00 +# Last generated: 2025-09-01T12:08:33.833560+00:00 [tox] requires = @@ -147,7 +147,7 @@ envlist = {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 - {py3.10,py3.12,py3.13}-openai_agents-v0.2.9 + {py3.10,py3.12,py3.13}-openai_agents-v0.2.10 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 @@ -210,7 +210,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.233.3 {py3.9,py3.12,py3.13}-strawberry-v0.257.0 - {py3.9,py3.12,py3.13}-strawberry-v0.280.0 + {py3.9,py3.12,py3.13}-strawberry-v0.281.0 # ~~~ Network ~~~ @@ -218,6 +218,7 @@ envlist = {py3.7,py3.9,py3.10}-grpc-v1.46.5 {py3.7,py3.11,py3.12}-grpc-v1.60.2 {py3.9,py3.12,py3.13}-grpc-v1.74.0 + {py3.9,py3.12,py3.13}-grpc-v1.75.0rc1 # ~~~ Tasks ~~~ @@ -311,6 +312,7 @@ envlist = {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 + {py3.7,py3.12,py3.13}-typer-v0.17.3 @@ -527,7 +529,7 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 - openai_agents-v0.2.9: openai-agents==0.2.9 + openai_agents-v0.2.10: openai-agents==0.2.10 openai_agents: pytest-asyncio huggingface_hub-v0.22.2: huggingface_hub==0.22.2 @@ -601,7 +603,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.233.3: strawberry-graphql[fastapi,flask]==0.233.3 strawberry-v0.257.0: strawberry-graphql[fastapi,flask]==0.257.0 - strawberry-v0.280.0: strawberry-graphql[fastapi,flask]==0.280.0 + strawberry-v0.281.0: strawberry-graphql[fastapi,flask]==0.281.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.233.3: pydantic<2.11 @@ -613,6 +615,7 @@ deps = grpc-v1.46.5: grpcio==1.46.5 grpc-v1.60.2: grpcio==1.60.2 grpc-v1.74.0: grpcio==1.74.0 + grpc-v1.75.0rc1: grpcio==1.75.0rc1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -782,6 +785,7 @@ deps = typer-v0.15.4: typer==0.15.4 typer-v0.16.1: typer==0.16.1 + typer-v0.17.3: typer==0.17.3 From 1d473b62490508cc3f8070bdaccc2e5bd20b182a Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 2 Sep 2025 10:44:22 +0200 Subject: [PATCH 424/591] toxgen: Add variants & move OpenAI under toxgen (#4730) Adds supports for variants, i.e., the same test suite running with a slightly different setup (for instance, a different set of dependencies, like `openai` and `openai_notiktoken`). To add a variant, simply add a new test suite to the config. The tricky part is naming. I had to rename `openai` to `openai_base` since otherwise the `openai_notiktoken` and `openai_agents` test suite would be run with `tox -e py-openai` / `./scripts/runtox.sh py-openai` due to how tox works. They should be treated as three different suites. Closes https://github.com/getsentry/sentry-python/issues/4507 --- .github/workflows/test-integrations-ai.yml | 16 ++++-- scripts/populate_tox/README.md | 8 +++ scripts/populate_tox/config.py | 18 +++++++ scripts/populate_tox/populate_tox.py | 8 +-- scripts/populate_tox/tox.jinja | 24 +-------- .../split_tox_gh_actions.py | 3 +- tox.ini | 53 +++++++++++-------- 7 files changed, 77 insertions(+), 53 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index a784f9fc47..a6995fa268 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -62,10 +62,14 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-latest" - - name: Test openai latest + - name: Test openai_base latest run: | set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openai-latest" + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_base-latest" + - name: Test openai_notiktoken latest + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_notiktoken-latest" - name: Test openai_agents latest run: | set -x # print commands that are executed @@ -141,10 +145,14 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain" - - name: Test openai pinned + - name: Test openai_base pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_base" + - name: Test openai_notiktoken pinned run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai" + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_notiktoken" - name: Test openai_agents pinned run: | set -x # print commands that are executed diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index c9a3b67ba0..c48d57734d 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -153,6 +153,14 @@ be expressed like so: } ``` +### `integration_name` + +Sometimes, the name of the test suite doesn't match the name of the integration. +For example, we have the `openai_base` and `openai_notiktoken` test suites, both +of which are actually testing the `openai` integration. If this is the case, you can use the `integration_name` key to define the name of the integration. If not provided, it will default to the name of the test suite. + +Linking an integration to a test suite allows the script to access integration configuration like for example the minimum version defined in `sentry_sdk/integrations/__init__.py`. + ## How-Tos diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f395289b4a..65e463a947 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -139,6 +139,24 @@ "loguru": { "package": "loguru", }, + "openai_base": { + "package": "openai", + "integration_name": "openai", + "deps": { + "*": ["pytest-asyncio", "tiktoken"], + "<1.55": ["httpx<0.28"], + }, + "python": ">=3.8", + }, + "openai_notiktoken": { + "package": "openai", + "integration_name": "openai", + "deps": { + "*": ["pytest-asyncio"], + "<1.55": ["httpx<0.28"], + }, + "python": ">=3.8", + }, "openai_agents": { "package": "openai-agents", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 3ca5ab18c8..53d5609d50 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -76,8 +76,6 @@ "httpx", "langchain", "langchain_notiktoken", - "openai", - "openai_notiktoken", "pure_eval", "quart", "ray", @@ -141,7 +139,11 @@ def _prefilter_releases( - the list of prefiltered releases - an optional prerelease if there is one that should be tested """ - min_supported = _MIN_VERSIONS.get(integration) + integration_name = ( + TEST_SUITE_CONFIG[integration].get("integration_name") or integration + ) + + min_supported = _MIN_VERSIONS.get(integration_name) if min_supported is not None: min_supported = Version(".".join(map(str, min_supported))) else: diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 4c3b86af81..632ce7c71b 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -83,13 +83,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # OpenAI - {py3.9,py3.11,py3.12}-openai-v1.0 - {py3.9,py3.11,py3.12}-openai-v1.22 - {py3.9,py3.11,py3.12}-openai-v1.55 - {py3.9,py3.11,py3.12}-openai-latest - {py3.9,py3.11,py3.12}-openai-notiktoken - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -252,20 +245,6 @@ deps = langchain-{latest,notiktoken}: openai>=1.6.1 langchain-latest: tiktoken~=0.6.0 - # OpenAI - openai: pytest-asyncio - openai-v1.0: openai~=1.0.0 - openai-v1.0: tiktoken - openai-v1.0: httpx<0.28.0 - openai-v1.22: openai~=1.22.0 - openai-v1.22: tiktoken - openai-v1.22: httpx<0.28.0 - openai-v1.55: openai~=1.55.0 - openai-v1.55: tiktoken - openai-latest: openai - openai-latest: tiktoken~=0.6.0 - openai-notiktoken: openai - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -401,7 +380,8 @@ setenv = launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru - openai: TESTPATH=tests/integrations/openai + openai_base: TESTPATH=tests/integrations/openai + openai_notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index af1ff84cd6..305ceeae76 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -62,7 +62,8 @@ "anthropic", "cohere", "langchain", - "openai", + "openai_base", + "openai_notiktoken", "openai_agents", "huggingface_hub", ], diff --git a/tox.ini b/tox.ini index 0dbcef2c64..eea115876b 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-01T12:08:33.833560+00:00 +# Last generated: 2025-09-01T14:09:50.564158+00:00 [tox] requires = @@ -83,13 +83,6 @@ envlist = {py3.9,py3.11,py3.12}-langchain-latest {py3.9,py3.11,py3.12}-langchain-notiktoken - # OpenAI - {py3.9,py3.11,py3.12}-openai-v1.0 - {py3.9,py3.11,py3.12}-openai-v1.22 - {py3.9,py3.11,py3.12}-openai-v1.55 - {py3.9,py3.11,py3.12}-openai-latest - {py3.9,py3.11,py3.12}-openai-notiktoken - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -145,6 +138,16 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.13.12 {py3.9,py3.11,py3.12}-cohere-v5.17.0 + {py3.8,py3.11,py3.12}-openai_base-v1.0.1 + {py3.8,py3.11,py3.12}-openai_base-v1.35.15 + {py3.8,py3.11,py3.12}-openai_base-v1.69.0 + {py3.8,py3.12,py3.13}-openai_base-v1.102.0 + + {py3.8,py3.11,py3.12}-openai_notiktoken-v1.0.1 + {py3.8,py3.11,py3.12}-openai_notiktoken-v1.35.15 + {py3.8,py3.11,py3.12}-openai_notiktoken-v1.69.0 + {py3.8,py3.12,py3.13}-openai_notiktoken-v1.102.0 + {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.10 @@ -423,20 +426,6 @@ deps = langchain-{latest,notiktoken}: openai>=1.6.1 langchain-latest: tiktoken~=0.6.0 - # OpenAI - openai: pytest-asyncio - openai-v1.0: openai~=1.0.0 - openai-v1.0: tiktoken - openai-v1.0: httpx<0.28.0 - openai-v1.22: openai~=1.22.0 - openai-v1.22: tiktoken - openai-v1.22: httpx<0.28.0 - openai-v1.55: openai~=1.55.0 - openai-v1.55: tiktoken - openai-latest: openai - openai-latest: tiktoken~=0.6.0 - openai-notiktoken: openai - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -527,6 +516,23 @@ deps = cohere-v5.13.12: cohere==5.13.12 cohere-v5.17.0: cohere==5.17.0 + openai_base-v1.0.1: openai==1.0.1 + openai_base-v1.35.15: openai==1.35.15 + openai_base-v1.69.0: openai==1.69.0 + openai_base-v1.102.0: openai==1.102.0 + openai_base: pytest-asyncio + openai_base: tiktoken + openai_base-v1.0.1: httpx<0.28 + openai_base-v1.35.15: httpx<0.28 + + openai_notiktoken-v1.0.1: openai==1.0.1 + openai_notiktoken-v1.35.15: openai==1.35.15 + openai_notiktoken-v1.69.0: openai==1.69.0 + openai_notiktoken-v1.102.0: openai==1.102.0 + openai_notiktoken: pytest-asyncio + openai_notiktoken-v1.0.1: httpx<0.28 + openai_notiktoken-v1.35.15: httpx<0.28 + openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.10: openai-agents==0.2.10 @@ -831,7 +837,8 @@ setenv = launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru - openai: TESTPATH=tests/integrations/openai + openai_base: TESTPATH=tests/integrations/openai + openai_notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry From 65755f95351581bd89101ce8eba9ff4768c9474e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 2 Sep 2025 13:42:55 +0200 Subject: [PATCH 425/591] tests: Move langchain under toxgen (#4734) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - move the langchain test suites to be governed by toxgen - this indirectly results in removing the `-latest` tests in the AI group 🎉 - `-latest` tests predate toxgen and langchain was the last non-toxgen test suite (the rest was just skipped) -- all AI tests are now pinned - updated the naming scheme to use dashes instead of underscores for variants so that it's clearer if something is part of the name of the integration or if it denotes a variant - for instance, `openai-base` means this is the `base` variant of the `openai` test suite, but `openai_agents` means this is the `openai_agents` test suite (no variant) I'm explicitly ignoring the two alpha versions of 1.0 since adapting the integration to work with those is out of scope: [dedicated issue](https://github.com/getsentry/sentry-python/issues/4735) Part of https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-ai.yml | 99 ++--------------- scripts/populate_tox/config.py | 24 ++++- scripts/populate_tox/tox.jinja | 27 +---- .../split_tox_gh_actions.py | 7 +- sentry_sdk/integrations/__init__.py | 2 +- tox.ini | 100 +++++++++--------- 6 files changed, 93 insertions(+), 166 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index a6995fa268..72a4253744 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -22,89 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-ai-latest: - name: AI (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.9","3.11","3.12"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test anthropic latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-anthropic-latest" - - name: Test cohere latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-cohere-latest" - - name: Test langchain latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-latest" - - name: Test openai_base latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_base-latest" - - name: Test openai_notiktoken latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_notiktoken-latest" - - name: Test openai_agents latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_agents-latest" - - name: Test huggingface_hub latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-huggingface_hub-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-ai-pinned: name: AI (pinned) timeout-minutes: 30 @@ -141,18 +58,22 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-cohere" - - name: Test langchain pinned + - name: Test langchain-base pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain-base" + - name: Test langchain-notiktoken pinned run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain" - - name: Test openai_base pinned + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain-notiktoken" + - name: Test openai-base pinned run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_base" - - name: Test openai_notiktoken pinned + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai-base" + - name: Test openai-notiktoken pinned run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_notiktoken" + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai-notiktoken" - name: Test openai_agents pinned run: | set -x # print commands that are executed diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 65e463a947..0d4d0fe6ee 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -126,6 +126,26 @@ "huggingface_hub": { "package": "huggingface_hub", }, + "langchain-base": { + "package": "langchain", + "integration_name": "langchain", + "deps": { + "*": ["openai", "tiktoken", "langchain-openai"], + "<=0.1": ["httpx<0.28.0"], + ">=0.3": ["langchain-community"], + }, + "include": "<1.0", + }, + "langchain-notiktoken": { + "package": "langchain", + "integration_name": "langchain", + "deps": { + "*": ["openai", "langchain-openai"], + "<=0.1": ["httpx<0.28.0"], + ">=0.3": ["langchain-community"], + }, + "include": "<1.0", + }, "launchdarkly": { "package": "launchdarkly-server-sdk", }, @@ -139,7 +159,7 @@ "loguru": { "package": "loguru", }, - "openai_base": { + "openai-base": { "package": "openai", "integration_name": "openai", "deps": { @@ -148,7 +168,7 @@ }, "python": ">=3.8", }, - "openai_notiktoken": { + "openai-notiktoken": { "package": "openai", "integration_name": "openai", "deps": { diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 632ce7c71b..2b968b7aa1 100644 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -77,12 +77,6 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} {py3.9,py3.12,py3.13}-httpx-latest - # Langchain - {py3.9,py3.11,py3.12}-langchain-v0.1 - {py3.9,py3.11,py3.12}-langchain-v0.3 - {py3.9,py3.11,py3.12}-langchain-latest - {py3.9,py3.11,py3.12}-langchain-notiktoken - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -231,20 +225,6 @@ deps = httpx-v0.27: httpx~=0.27.0 httpx-latest: httpx - # Langchain - langchain-v0.1: openai~=1.0.0 - langchain-v0.1: langchain~=0.1.11 - langchain-v0.1: tiktoken~=0.6.0 - langchain-v0.1: httpx<0.28.0 - langchain-v0.3: langchain~=0.3.0 - langchain-v0.3: langchain-community - langchain-v0.3: tiktoken - langchain-v0.3: openai - langchain-{latest,notiktoken}: langchain - langchain-{latest,notiktoken}: langchain-openai - langchain-{latest,notiktoken}: openai>=1.6.1 - langchain-latest: tiktoken~=0.6.0 - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -376,12 +356,13 @@ setenv = httpx: TESTPATH=tests/integrations/httpx huey: TESTPATH=tests/integrations/huey huggingface_hub: TESTPATH=tests/integrations/huggingface_hub - langchain: TESTPATH=tests/integrations/langchain + langchain-base: TESTPATH=tests/integrations/langchain + langchain-notiktoken: TESTPATH=tests/integrations/langchain launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru - openai_base: TESTPATH=tests/integrations/openai - openai_notiktoken: TESTPATH=tests/integrations/openai + openai-base: TESTPATH=tests/integrations/openai + openai-notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 305ceeae76..1c3435f43b 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -61,9 +61,10 @@ "AI": [ "anthropic", "cohere", - "langchain", - "openai_base", - "openai_notiktoken", + "langchain-base", + "langchain-notiktoken", + "openai-base", + "openai-notiktoken", "openai_agents", "huggingface_hub", ], diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index e2eadd523d..6f0109aced 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -141,7 +141,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "graphene": (3, 3), "grpc": (1, 32, 0), # grpcio "huggingface_hub": (0, 22), - "langchain": (0, 0, 210), + "langchain": (0, 1, 0), "launchdarkly": (9, 8, 0), "loguru": (0, 7, 0), "openai": (1, 0, 0), diff --git a/tox.ini b/tox.ini index eea115876b..0898bc888f 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-01T14:09:50.564158+00:00 +# Last generated: 2025-09-02T10:59:55.249513+00:00 [tox] requires = @@ -77,12 +77,6 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} {py3.9,py3.12,py3.13}-httpx-latest - # Langchain - {py3.9,py3.11,py3.12}-langchain-v0.1 - {py3.9,py3.11,py3.12}-langchain-v0.3 - {py3.9,py3.11,py3.12}-langchain-latest - {py3.9,py3.11,py3.12}-langchain-notiktoken - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -138,15 +132,23 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.13.12 {py3.9,py3.11,py3.12}-cohere-v5.17.0 - {py3.8,py3.11,py3.12}-openai_base-v1.0.1 - {py3.8,py3.11,py3.12}-openai_base-v1.35.15 - {py3.8,py3.11,py3.12}-openai_base-v1.69.0 - {py3.8,py3.12,py3.13}-openai_base-v1.102.0 + {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 + {py3.9,py3.11,py3.12}-langchain-base-v0.2.17 + {py3.9,py3.12,py3.13}-langchain-base-v0.3.27 + + {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.1.20 + {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.2.17 + {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 - {py3.8,py3.11,py3.12}-openai_notiktoken-v1.0.1 - {py3.8,py3.11,py3.12}-openai_notiktoken-v1.35.15 - {py3.8,py3.11,py3.12}-openai_notiktoken-v1.69.0 - {py3.8,py3.12,py3.13}-openai_notiktoken-v1.102.0 + {py3.8,py3.11,py3.12}-openai-base-v1.0.1 + {py3.8,py3.11,py3.12}-openai-base-v1.35.15 + {py3.8,py3.11,py3.12}-openai-base-v1.69.0 + {py3.8,py3.12,py3.13}-openai-base-v1.102.0 + + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.35.15 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.69.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.102.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 @@ -412,20 +414,6 @@ deps = httpx-v0.27: httpx~=0.27.0 httpx-latest: httpx - # Langchain - langchain-v0.1: openai~=1.0.0 - langchain-v0.1: langchain~=0.1.11 - langchain-v0.1: tiktoken~=0.6.0 - langchain-v0.1: httpx<0.28.0 - langchain-v0.3: langchain~=0.3.0 - langchain-v0.3: langchain-community - langchain-v0.3: tiktoken - langchain-v0.3: openai - langchain-{latest,notiktoken}: langchain - langchain-{latest,notiktoken}: langchain-openai - langchain-{latest,notiktoken}: openai>=1.6.1 - langchain-latest: tiktoken~=0.6.0 - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -516,22 +504,37 @@ deps = cohere-v5.13.12: cohere==5.13.12 cohere-v5.17.0: cohere==5.17.0 - openai_base-v1.0.1: openai==1.0.1 - openai_base-v1.35.15: openai==1.35.15 - openai_base-v1.69.0: openai==1.69.0 - openai_base-v1.102.0: openai==1.102.0 - openai_base: pytest-asyncio - openai_base: tiktoken - openai_base-v1.0.1: httpx<0.28 - openai_base-v1.35.15: httpx<0.28 - - openai_notiktoken-v1.0.1: openai==1.0.1 - openai_notiktoken-v1.35.15: openai==1.35.15 - openai_notiktoken-v1.69.0: openai==1.69.0 - openai_notiktoken-v1.102.0: openai==1.102.0 - openai_notiktoken: pytest-asyncio - openai_notiktoken-v1.0.1: httpx<0.28 - openai_notiktoken-v1.35.15: httpx<0.28 + langchain-base-v0.1.20: langchain==0.1.20 + langchain-base-v0.2.17: langchain==0.2.17 + langchain-base-v0.3.27: langchain==0.3.27 + langchain-base: openai + langchain-base: tiktoken + langchain-base: langchain-openai + langchain-base-v0.3.27: langchain-community + + langchain-notiktoken-v0.1.20: langchain==0.1.20 + langchain-notiktoken-v0.2.17: langchain==0.2.17 + langchain-notiktoken-v0.3.27: langchain==0.3.27 + langchain-notiktoken: openai + langchain-notiktoken: langchain-openai + langchain-notiktoken-v0.3.27: langchain-community + + openai-base-v1.0.1: openai==1.0.1 + openai-base-v1.35.15: openai==1.35.15 + openai-base-v1.69.0: openai==1.69.0 + openai-base-v1.102.0: openai==1.102.0 + openai-base: pytest-asyncio + openai-base: tiktoken + openai-base-v1.0.1: httpx<0.28 + openai-base-v1.35.15: httpx<0.28 + + openai-notiktoken-v1.0.1: openai==1.0.1 + openai-notiktoken-v1.35.15: openai==1.35.15 + openai-notiktoken-v1.69.0: openai==1.69.0 + openai-notiktoken-v1.102.0: openai==1.102.0 + openai-notiktoken: pytest-asyncio + openai-notiktoken-v1.0.1: httpx<0.28 + openai-notiktoken-v1.35.15: httpx<0.28 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 @@ -833,12 +836,13 @@ setenv = httpx: TESTPATH=tests/integrations/httpx huey: TESTPATH=tests/integrations/huey huggingface_hub: TESTPATH=tests/integrations/huggingface_hub - langchain: TESTPATH=tests/integrations/langchain + langchain-base: TESTPATH=tests/integrations/langchain + langchain-notiktoken: TESTPATH=tests/integrations/langchain launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru - openai_base: TESTPATH=tests/integrations/openai - openai_notiktoken: TESTPATH=tests/integrations/openai + openai-base: TESTPATH=tests/integrations/openai + openai-notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry From b1a8b6333ceafd116fbaf1f50e1e14967f5d9f94 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Tue, 2 Sep 2025 14:14:36 +0200 Subject: [PATCH 426/591] fix(openai): Avoid double exit causing an unraisable exception (#4736) Add parameter to the method capturing exceptions in the OpenAI integration, to determine if the span context is closed with __exit__() or not. The option is used to prevent double exit scenarios when a span context is managed automatically. Related to: https://github.com/getsentry/sentry-python/issues/4723 --- sentry_sdk/integrations/openai.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 187f795807..6ea545322c 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -78,12 +78,12 @@ def count_tokens(self, s): return 0 -def _capture_exception(exc): - # type: (Any) -> None +def _capture_exception(exc, manual_span_cleanup=True): + # type: (Any, bool) -> None # Close an eventually open span # We need to do this by hand because we are not using the start_span context manager current_span = sentry_sdk.get_current_span() - if current_span is not None: + if manual_span_cleanup and current_span is not None: current_span.__exit__(None, None, None) event, hint = event_from_exception( @@ -516,7 +516,7 @@ def _execute_sync(f, *args, **kwargs): try: result = f(*args, **kwargs) except Exception as e: - _capture_exception(e) + _capture_exception(e, manual_span_cleanup=False) raise e from None return gen.send(result) @@ -550,7 +550,7 @@ async def _execute_async(f, *args, **kwargs): try: result = await f(*args, **kwargs) except Exception as e: - _capture_exception(e) + _capture_exception(e, manual_span_cleanup=False) raise e from None return gen.send(result) From 9c4eb5e272910aafadd65e8ae92696034647e47f Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Tue, 2 Sep 2025 15:27:41 +0200 Subject: [PATCH 427/591] tests: Trigger Pytest failure when an unraisable exception occurs (#4738) Set Pytest command-line argument to return non-zero exit code when an unraisable exception is encountered. Closes https://github.com/getsentry/sentry-python/issues/4723. --- scripts/populate_tox/tox.jinja | 2 +- tox.ini | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) mode change 100644 => 100755 scripts/populate_tox/tox.jinja diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja old mode 100644 new mode 100755 index 2b968b7aa1..42c570b111 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -429,7 +429,7 @@ commands = ; Running `pytest` as an executable suffers from an import error ; when loading tests in scenarios. In particular, django fails to ; load the settings from the test module. - python -m pytest {env:TESTPATH} -o junit_suite_name={envname} {posargs} + python -m pytest -W error::pytest.PytestUnraisableExceptionWarning {env:TESTPATH} -o junit_suite_name={envname} {posargs} [testenv:linters] commands = diff --git a/tox.ini b/tox.ini index 0898bc888f..a8e66cb80f 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-02T10:59:55.249513+00:00 +# Last generated: 2025-09-02T12:34:09.591543+00:00 [tox] requires = @@ -909,7 +909,7 @@ commands = ; Running `pytest` as an executable suffers from an import error ; when loading tests in scenarios. In particular, django fails to ; load the settings from the test module. - python -m pytest {env:TESTPATH} -o junit_suite_name={envname} {posargs} + python -m pytest -W error::pytest.PytestUnraisableExceptionWarning {env:TESTPATH} -o junit_suite_name={envname} {posargs} [testenv:linters] commands = From c213abf4a4ea9a09f8387fd192e8ee1992851657 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 2 Sep 2025 16:01:12 +0200 Subject: [PATCH 428/591] Remove old langchain test suites from ignore list (#4737) Forgot to remove these two from the toxgen ignore list. Shouldn't have any actual effect on tests since the test suites are now called differently. --- scripts/populate_tox/populate_tox.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 53d5609d50..179a466944 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -74,8 +74,6 @@ "chalice", "gcp", "httpx", - "langchain", - "langchain_notiktoken", "pure_eval", "quart", "ray", From 4456351b9156fd44b5797583d37889ed2af70517 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 3 Sep 2025 08:46:15 +0200 Subject: [PATCH 429/591] Fix `openai_agents` in CI (#4742) A new version of `openai`, which is a dependency of `openai_agents`, [came out an hour ago](https://pypi.org/project/openai/#history), which [broke](https://github.com/getsentry/sentry-python/actions/runs/17405958869/job/49410259073) our CI. Pinning for now. --- scripts/populate_tox/config.py | 1 + tox.ini | 13 ++++++++----- 2 files changed, 9 insertions(+), 5 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 0d4d0fe6ee..69f7b02e21 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -181,6 +181,7 @@ "package": "openai-agents", "deps": { "*": ["pytest-asyncio"], + "<=0.2.10": ["openai<1.103.0"], }, "python": ">=3.10", }, diff --git a/tox.ini b/tox.ini index a8e66cb80f..c45c72bf85 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-02T12:34:09.591543+00:00 +# Last generated: 2025-09-02T14:49:13.002983+00:00 [tox] requires = @@ -143,12 +143,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.35.15 {py3.8,py3.11,py3.12}-openai-base-v1.69.0 - {py3.8,py3.12,py3.13}-openai-base-v1.102.0 + {py3.8,py3.12,py3.13}-openai-base-v1.103.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.35.15 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.69.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.102.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.103.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 @@ -522,7 +522,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.35.15: openai==1.35.15 openai-base-v1.69.0: openai==1.69.0 - openai-base-v1.102.0: openai==1.102.0 + openai-base-v1.103.0: openai==1.103.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -531,7 +531,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.35.15: openai==1.35.15 openai-notiktoken-v1.69.0: openai==1.69.0 - openai-notiktoken-v1.102.0: openai==1.102.0 + openai-notiktoken-v1.103.0: openai==1.103.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.35.15: httpx<0.28 @@ -540,6 +540,9 @@ deps = openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.10: openai-agents==0.2.10 openai_agents: pytest-asyncio + openai_agents-v0.0.19: openai<1.103.0 + openai_agents-v0.1.0: openai<1.103.0 + openai_agents-v0.2.10: openai<1.103.0 huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 From f702ec94badc17bd09d7d3ccf7414fde01a173c8 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 3 Sep 2025 09:44:55 +0200 Subject: [PATCH 430/591] fix: Constrain types of ai_track decorator (#4745) I followed how other functions in the SDK are typed. For example, other wrappers have the signature `(F) -> F` for a type variable `F`, although here the function can be async as well. Closes https://github.com/getsentry/sentry-python/issues/4663 --- sentry_sdk/ai/monitoring.py | 12 +++++++----- 1 file changed, 7 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/ai/monitoring.py b/sentry_sdk/ai/monitoring.py index e3f372c3ba..9dd1aa132c 100644 --- a/sentry_sdk/ai/monitoring.py +++ b/sentry_sdk/ai/monitoring.py @@ -10,7 +10,9 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Optional, Callable, Any + from typing import Optional, Callable, Awaitable, Any, Union, TypeVar + + F = TypeVar("F", bound=Union[Callable[..., Any], Callable[..., Awaitable[Any]]]) _ai_pipeline_name = ContextVar("ai_pipeline_name", default=None) @@ -26,9 +28,9 @@ def get_ai_pipeline_name(): def ai_track(description, **span_kwargs): - # type: (str, Any) -> Callable[..., Any] + # type: (str, Any) -> Callable[[F], F] def decorator(f): - # type: (Callable[..., Any]) -> Callable[..., Any] + # type: (F) -> F def sync_wrapped(*args, **kwargs): # type: (Any, Any) -> Any curr_pipeline = _ai_pipeline_name.get() @@ -88,9 +90,9 @@ async def async_wrapped(*args, **kwargs): return res if inspect.iscoroutinefunction(f): - return wraps(f)(async_wrapped) + return wraps(f)(async_wrapped) # type: ignore else: - return wraps(f)(sync_wrapped) + return wraps(f)(sync_wrapped) # type: ignore return decorator From 5f2adcffecff85b1f736f93701cf154d58f85653 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 3 Sep 2025 16:11:11 +0200 Subject: [PATCH 431/591] Wrap span restoration in `__exit__` in `capture_internal_exceptions` (#4719) Ref https://github.com/getsentry/sentry-python/issues/4718 Does not solve the underlying issue and might leave things in an inconsistent state, but it's still preferable to letting an error bubble up to the user. --- sentry_sdk/tracing.py | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index c9b357305a..0d1fcc45da 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -8,6 +8,7 @@ from sentry_sdk.consts import INSTRUMENTER, SPANSTATUS, SPANDATA, SPANTEMPLATE from sentry_sdk.profiler.continuous_profiler import get_profiler_id from sentry_sdk.utils import ( + capture_internal_exceptions, get_current_thread_meta, is_valid_sample_rate, logger, @@ -418,10 +419,11 @@ def __exit__(self, ty, value, tb): if value is not None and should_be_treated_as_error(ty, value): self.set_status(SPANSTATUS.INTERNAL_ERROR) - scope, old_span = self._context_manager_state - del self._context_manager_state - self.finish(scope) - scope.span = old_span + with capture_internal_exceptions(): + scope, old_span = self._context_manager_state + del self._context_manager_state + self.finish(scope) + scope.span = old_span @property def containing_transaction(self): From 6d6e8a2e70fee7553675288b82e28ba311ffca3c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 3 Sep 2025 16:13:51 +0200 Subject: [PATCH 432/591] Don't fail if there is no `_context_manager_state` (#4698) This is not a fix -- it just makes the SDK not propagate an internal SDK exception upwards. --- sentry_sdk/profiler/transaction_profiler.py | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/profiler/transaction_profiler.py b/sentry_sdk/profiler/transaction_profiler.py index 3743b7c905..d228f77de9 100644 --- a/sentry_sdk/profiler/transaction_profiler.py +++ b/sentry_sdk/profiler/transaction_profiler.py @@ -45,6 +45,7 @@ ) from sentry_sdk.utils import ( capture_internal_exception, + capture_internal_exceptions, get_current_thread_meta, is_gevent, is_valid_sample_rate, @@ -369,12 +370,13 @@ def __enter__(self): def __exit__(self, ty, value, tb): # type: (Optional[Any], Optional[Any], Optional[Any]) -> None - self.stop() + with capture_internal_exceptions(): + self.stop() - scope, old_profile = self._context_manager_state - del self._context_manager_state + scope, old_profile = self._context_manager_state + del self._context_manager_state - scope.profile = old_profile + scope.profile = old_profile def write(self, ts, sample): # type: (int, ExtractedSample) -> None From a6e3b50004c13f32de8c30e7632c164a39d7babe Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 3 Sep 2025 16:49:29 +0200 Subject: [PATCH 433/591] feat(integrations): Add unraisable exception integration (#4733) Adds an uncaught exception integration, enabled by default. The integration forwards the exception to Sentry only if the exception value and stacktrace are set. Closes https://github.com/getsentry/sentry-python/issues/374 --- sentry_sdk/integrations/unraisablehook.py | 53 ++++++++++++++++++ .../unraisablehook/test_unraisablehook.py | 56 +++++++++++++++++++ tests/test_basics.py | 1 + 3 files changed, 110 insertions(+) create mode 100644 sentry_sdk/integrations/unraisablehook.py create mode 100644 tests/integrations/unraisablehook/test_unraisablehook.py diff --git a/sentry_sdk/integrations/unraisablehook.py b/sentry_sdk/integrations/unraisablehook.py new file mode 100644 index 0000000000..cfb8212c71 --- /dev/null +++ b/sentry_sdk/integrations/unraisablehook.py @@ -0,0 +1,53 @@ +import sys + +import sentry_sdk +from sentry_sdk.utils import ( + capture_internal_exceptions, + event_from_exception, +) +from sentry_sdk.integrations import Integration + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Callable + from typing import Any + + +class UnraisablehookIntegration(Integration): + identifier = "unraisablehook" + + @staticmethod + def setup_once(): + # type: () -> None + sys.unraisablehook = _make_unraisable(sys.unraisablehook) + + +def _make_unraisable(old_unraisablehook): + # type: (Callable[[sys.UnraisableHookArgs], Any]) -> Callable[[sys.UnraisableHookArgs], Any] + def sentry_sdk_unraisablehook(unraisable): + # type: (sys.UnraisableHookArgs) -> None + integration = sentry_sdk.get_client().get_integration(UnraisablehookIntegration) + + # Note: If we replace this with ensure_integration_enabled then + # we break the exceptiongroup backport; + # See: https://github.com/getsentry/sentry-python/issues/3097 + if integration is None: + return old_unraisablehook(unraisable) + + if unraisable.exc_value and unraisable.exc_traceback: + with capture_internal_exceptions(): + event, hint = event_from_exception( + ( + unraisable.exc_type, + unraisable.exc_value, + unraisable.exc_traceback, + ), + client_options=sentry_sdk.get_client().options, + mechanism={"type": "unraisablehook", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + return old_unraisablehook(unraisable) + + return sentry_sdk_unraisablehook diff --git a/tests/integrations/unraisablehook/test_unraisablehook.py b/tests/integrations/unraisablehook/test_unraisablehook.py new file mode 100644 index 0000000000..2f97886ce8 --- /dev/null +++ b/tests/integrations/unraisablehook/test_unraisablehook.py @@ -0,0 +1,56 @@ +import pytest +import sys +import subprocess + +from textwrap import dedent + + +TEST_PARAMETERS = [ + ("", "HttpTransport"), + ('_experiments={"transport_http2": True}', "Http2Transport"), +] + +minimum_python_38 = pytest.mark.skipif( + sys.version_info < (3, 8), + reason="The unraisable exception hook is only available in Python 3.8 and above.", +) + + +@minimum_python_38 +@pytest.mark.parametrize("options, transport", TEST_PARAMETERS) +def test_unraisablehook(tmpdir, options, transport): + app = tmpdir.join("app.py") + app.write( + dedent( + """ + from sentry_sdk import init, transport + from sentry_sdk.integrations.unraisablehook import UnraisablehookIntegration + + class Undeletable: + def __del__(self): + 1 / 0 + + def capture_envelope(self, envelope): + print("capture_envelope was called") + event = envelope.get_event() + if event is not None: + print(event) + + transport.{transport}.capture_envelope = capture_envelope + + init("http://foobar@localhost/123", integrations=[UnraisablehookIntegration()], {options}) + + undeletable = Undeletable() + del undeletable + """.format( + transport=transport, options=options + ) + ) + ) + + output = subprocess.check_output( + [sys.executable, str(app)], stderr=subprocess.STDOUT + ) + + assert b"ZeroDivisionError" in output + assert b"capture_envelope was called" in output diff --git a/tests/test_basics.py b/tests/test_basics.py index 2eeba78216..45303c9a59 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -870,6 +870,7 @@ def foo(event, hint): (["celery"], "sentry.python"), (["dedupe"], "sentry.python"), (["excepthook"], "sentry.python"), + (["unraisablehook"], "sentry.python"), (["executing"], "sentry.python"), (["modules"], "sentry.python"), (["pure_eval"], "sentry.python"), From 4e845d5767f8fd43f7eee310afc269bc070c9b3f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 3 Sep 2025 16:51:28 +0200 Subject: [PATCH 434/591] tests: Support dashes in test suite names (#4740) - actually support dashes in integration names in `tox.ini`/toxgen - make `split_tox_gh_actions.py` actually fail if it fails to parse `tox.ini` Context: `split_tox_gh_actions.py` was actually failing because it assumed there can't be dashes in test suite names, but since it was just printing the error instead of actually exiting with an error code, we didn't notice this in CI. --- .../split_tox_gh_actions.py | 41 ++++++++++++++----- 1 file changed, 31 insertions(+), 10 deletions(-) diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 1c3435f43b..cf83e0a3fe 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -17,6 +17,7 @@ import configparser import hashlib +import re import sys from collections import defaultdict from functools import reduce @@ -25,6 +26,18 @@ from jinja2 import Environment, FileSystemLoader +TOXENV_REGEX = re.compile( + r""" + {?(?P(py\d+\.\d+,?)+)}? + -(?P[a-z](?:[a-z_]|-(?!v{?\d|latest))*[a-z0-9]) + (?:-( + (v{?(?P[0-9.]+[0-9a-z,.]*}?)) + | + (?Platest) + ))? +""", + re.VERBOSE, +) OUT_DIR = Path(__file__).resolve().parent.parent.parent / ".github" / "workflows" TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" @@ -202,29 +215,37 @@ def parse_tox(): py_versions_pinned = defaultdict(set) py_versions_latest = defaultdict(set) + parsed_correctly = True + for line in lines: # normalize lines line = line.strip().lower() try: # parse tox environment definition - try: - (raw_python_versions, framework, framework_versions) = line.split("-") - except ValueError: - (raw_python_versions, framework) = line.split("-") - framework_versions = [] + parsed = TOXENV_REGEX.match(line) + if not parsed: + print(f"ERROR reading line {line}") + raise ValueError("Failed to parse tox environment definition") + + groups = parsed.groupdict() + raw_python_versions = groups["py_versions"] + framework = groups["framework"] + framework_versions_latest = groups.get("framework_versions_latest") or False # collect python versions to test the framework in - raw_python_versions = set( - raw_python_versions.replace("{", "").replace("}", "").split(",") - ) - if "latest" in framework_versions: + raw_python_versions = set(raw_python_versions.split(",")) + if framework_versions_latest: py_versions_latest[framework] |= raw_python_versions else: py_versions_pinned[framework] |= raw_python_versions - except ValueError: + except Exception: print(f"ERROR reading line {line}") + parsed_correctly = False + + if not parsed_correctly: + raise RuntimeError("Failed to parse tox.ini") py_versions_pinned = _normalize_py_versions(py_versions_pinned) py_versions_latest = _normalize_py_versions(py_versions_latest) From 7bc91eda417db023b76bd4193c80288943a27a65 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 3 Sep 2025 17:10:33 +0200 Subject: [PATCH 435/591] tests: Move arq under toxgen (#4739) Remove hardcoded arq config, generate with toxgen instead. --- scripts/populate_tox/config.py | 7 +++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 12 ---- tox.ini | 87 ++++++++++++++-------------- 4 files changed, 51 insertions(+), 56 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 69f7b02e21..f6093b0250 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -29,6 +29,13 @@ }, "python": ">=3.8", }, + "arq": { + "package": "arq", + "deps": { + "*": ["async-timeout", "pytest-asyncio", "fakeredis>=2.2.0,<2.8"], + "<=0.23": ["pydantic<2"], + }, + }, "bottle": { "package": "bottle", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 179a466944..a8c58938ae 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "arq", "asyncpg", "beam", "boto3", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 42c570b111..115b99fd5c 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -36,10 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # Arq - {py3.7,py3.11}-arq-v{0.23} - {py3.7,py3.12,py3.13}-arq-latest - # Asgi {py3.7,py3.12,py3.13}-asgi @@ -164,14 +160,6 @@ deps = # === Integrations === - # Arq - arq-v0.23: arq~=0.23.0 - arq-v0.23: pydantic<2 - arq-latest: arq - arq: fakeredis>=2.2.0,<2.8 - arq: pytest-asyncio - arq: async-timeout - # Asgi asgi: pytest-asyncio asgi: async-asgi-testclient diff --git a/tox.ini b/tox.ini index c45c72bf85..f2ad720a25 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-02T14:49:13.002983+00:00 +# Last generated: 2025-09-03T15:01:21.035943+00:00 [tox] requires = @@ -36,10 +36,6 @@ envlist = # At a minimum, we should test against at least the lowest # and the latest supported version of a framework. - # Arq - {py3.7,py3.11}-arq-v{0.23} - {py3.7,py3.12,py3.13}-arq-latest - # Asgi {py3.7,py3.12,py3.13}-asgi @@ -123,9 +119,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.32.0 - {py3.8,py3.11,py3.12}-anthropic-v0.48.0 - {py3.8,py3.12,py3.13}-anthropic-v0.64.0 + {py3.8,py3.11,py3.12}-anthropic-v0.33.1 + {py3.8,py3.11,py3.12}-anthropic-v0.50.0 + {py3.8,py3.12,py3.13}-anthropic-v0.66.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 @@ -141,14 +137,14 @@ envlist = {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 - {py3.8,py3.11,py3.12}-openai-base-v1.35.15 - {py3.8,py3.11,py3.12}-openai-base-v1.69.0 - {py3.8,py3.12,py3.13}-openai-base-v1.103.0 + {py3.8,py3.11,py3.12}-openai-base-v1.36.1 + {py3.8,py3.11,py3.12}-openai-base-v1.71.0 + {py3.8,py3.12,py3.13}-openai-base-v1.105.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.35.15 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.69.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.103.0 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.105.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 @@ -227,6 +223,11 @@ envlist = # ~~~ Tasks ~~~ + {py3.7,py3.9,py3.10}-arq-v0.23 + {py3.7,py3.10,py3.11}-arq-v0.24.0 + {py3.7,py3.10,py3.11}-arq-v0.25.0 + {py3.8,py3.11,py3.12}-arq-v0.26.3 + {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.12,py3.13}-celery-v5.5.3 @@ -250,9 +251,9 @@ envlist = {py3.6,py3.7}-django-v1.11.29 {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 - {py3.8,py3.11,py3.12}-django-v4.2.23 + {py3.8,py3.11,py3.12}-django-v4.2.24 {py3.10,py3.11,py3.12}-django-v5.0.14 - {py3.10,py3.12,py3.13}-django-v5.2.5 + {py3.10,py3.12,py3.13}-django-v5.2.6 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -353,14 +354,6 @@ deps = # === Integrations === - # Arq - arq-v0.23: arq~=0.23.0 - arq-v0.23: pydantic<2 - arq-latest: arq - arq: fakeredis>=2.2.0,<2.8 - arq: pytest-asyncio - arq: async-timeout - # Asgi asgi: pytest-asyncio asgi: async-asgi-testclient @@ -491,13 +484,12 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.32.0: anthropic==0.32.0 - anthropic-v0.48.0: anthropic==0.48.0 - anthropic-v0.64.0: anthropic==0.64.0 + anthropic-v0.33.1: anthropic==0.33.1 + anthropic-v0.50.0: anthropic==0.50.0 + anthropic-v0.66.0: anthropic==0.66.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.32.0: httpx<0.28.0 - anthropic-v0.48.0: httpx<0.28.0 + anthropic-v0.33.1: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 @@ -520,21 +512,21 @@ deps = langchain-notiktoken-v0.3.27: langchain-community openai-base-v1.0.1: openai==1.0.1 - openai-base-v1.35.15: openai==1.35.15 - openai-base-v1.69.0: openai==1.69.0 - openai-base-v1.103.0: openai==1.103.0 + openai-base-v1.36.1: openai==1.36.1 + openai-base-v1.71.0: openai==1.71.0 + openai-base-v1.105.0: openai==1.105.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 - openai-base-v1.35.15: httpx<0.28 + openai-base-v1.36.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 - openai-notiktoken-v1.35.15: openai==1.35.15 - openai-notiktoken-v1.69.0: openai==1.69.0 - openai-notiktoken-v1.103.0: openai==1.103.0 + openai-notiktoken-v1.36.1: openai==1.36.1 + openai-notiktoken-v1.71.0: openai==1.71.0 + openai-notiktoken-v1.105.0: openai==1.105.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 - openai-notiktoken-v1.35.15: httpx<0.28 + openai-notiktoken-v1.36.1: httpx<0.28 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 @@ -635,6 +627,15 @@ deps = # ~~~ Tasks ~~~ + arq-v0.23: arq==0.23 + arq-v0.24.0: arq==0.24.0 + arq-v0.25.0: arq==0.25.0 + arq-v0.26.3: arq==0.26.3 + arq: async-timeout + arq: pytest-asyncio + arq: fakeredis>=2.2.0,<2.8 + arq-v0.23: pydantic<2 + celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.5.3: celery==5.5.3 @@ -661,23 +662,23 @@ deps = django-v1.11.29: django==1.11.29 django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 - django-v4.2.23: django==4.2.23 + django-v4.2.24: django==4.2.24 django-v5.0.14: django==5.0.14 - django-v5.2.5: django==5.2.5 + django-v5.2.6: django==5.2.6 django: psycopg2-binary django: djangorestframework django: pytest-django django: Werkzeug django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] - django-v4.2.23: channels[daphne] + django-v4.2.24: channels[daphne] django-v5.0.14: channels[daphne] - django-v5.2.5: channels[daphne] + django-v5.2.6: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio - django-v4.2.23: pytest-asyncio + django-v4.2.24: pytest-asyncio django-v5.0.14: pytest-asyncio - django-v5.2.5: pytest-asyncio + django-v5.2.6: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 From 6f396f490a0a2c9bee4f4f8e66db49cd1f738d81 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 4 Sep 2025 09:24:28 +0200 Subject: [PATCH 436/591] meta: Update instructions on release process (#4755) --- CONTRIBUTING.md | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 024a374f85..313910fe56 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -138,18 +138,18 @@ _(only relevant for Python SDK core team)_ - On GitHub in the `sentry-python` repository, go to "Actions" and select the "Release" workflow. - Click on "Run workflow" on the right side, and make sure the `master` branch is selected. -- Set the "Version to release" input field. Here you decide if it is a major, minor or patch release. (See "Versioning Policy" below) +- Set the "Version to release" input field. Here you decide if it is a major, minor or patch release (see "Versioning Policy" below). - Click "Run Workflow". -This will trigger [Craft](https://github.com/getsentry/craft) to prepare everything needed for a release. (For more information, see [craft prepare](https://github.com/getsentry/craft#craft-prepare-preparing-a-new-release).) At the end of this process a release issue is created in the [Publish](https://github.com/getsentry/publish) repository. (Example release issue: https://github.com/getsentry/publish/issues/815) +This will trigger [Craft](https://github.com/getsentry/craft) to prepare everything needed for a release. (For more information, see [craft prepare](https://github.com/getsentry/craft#craft-prepare-preparing-a-new-release).) At the end of this process a release issue is created in the [Publish](https://github.com/getsentry/publish) repository (example issue: https://github.com/getsentry/publish/issues/815). -Now one of the persons with release privileges (most probably your engineering manager) will review this issue and then add the `accepted` label to the issue. +At the same time, the action will create a release branch in the `sentry-python` repository called `release/`. You may want to check out this branch and polish the auto-generated `CHANGELOG.md` before proceeding by including code snippets, descriptions, reordering and reformatting entries, in order to make the changelog as useful and actionable to users as possible. -There are always two persons involved in a release. +CI must be passing on the release branch; if there's any failure, Craft will not create a release. -If you are in a hurry and the release should be out immediately, there is a Slack channel called `#proj-release-approval` where you can see your release issue and where you can ping people to please have a look immediately. +Once the release branch is ready and green, notify your team (or your manager). They will need to add the `accepted` label to the issue in the `publish` repo. There are always two people involved in a release. Do not accept your own releases. -When the release issue is labeled `accepted`, [Craft](https://github.com/getsentry/craft) is triggered again to publish the release to all the right platforms. (See [craft publish](https://github.com/getsentry/craft#craft-publish-publishing-the-release) for more information.) At the end of this process the release issue on GitHub will be closed and the release is completed! Congratulations! +When the release issue is labeled `accepted`, [Craft](https://github.com/getsentry/craft) is triggered again to publish the release to all the right platforms. See [craft publish](https://github.com/getsentry/craft#craft-publish-publishing-the-release) for more information. At the end of this process, the release issue on GitHub will be closed and the release is completed! Congratulations! There is a sequence diagram visualizing all this in the [README.md](https://github.com/getsentry/publish) of the `Publish` repository. From c7f3b396920f652c9776baf4be7ebfedfbda7df7 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 4 Sep 2025 07:31:06 +0000 Subject: [PATCH 437/591] release: 2.36.0 --- CHANGELOG.md | 19 +++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 22 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 19f734976f..a0b3c1647e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,24 @@ # Changelog +## 2.36.0 + +### Various fixes & improvements + +- meta: Update instructions on release process (#4755) by @sentrivana +- tests: Move arq under toxgen (#4739) by @sentrivana +- tests: Support dashes in test suite names (#4740) by @sentrivana +- feat(integrations): Add unraisable exception integration (#4733) by @alexander-alderman-webb +- Don't fail if there is no `_context_manager_state` (#4698) by @sentrivana +- Wrap span restoration in `__exit__` in `capture_internal_exceptions` (#4719) by @sentrivana +- fix: Constrain types of ai_track decorator (#4745) by @alexander-alderman-webb +- Fix `openai_agents` in CI (#4742) by @sentrivana +- Remove old langchain test suites from ignore list (#4737) by @sentrivana +- tests: Trigger Pytest failure when an unraisable exception occurs (#4738) by @alexander-alderman-webb +- fix(openai): Avoid double exit causing an unraisable exception (#4736) by @alexander-alderman-webb +- tests: Move langchain under toxgen (#4734) by @sentrivana +- toxgen: Add variants & move OpenAI under toxgen (#4730) by @sentrivana +- Update tox.ini (#4731) by @sentrivana + ## 2.35.2 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 0863980aac..835c20b112 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.35.2" +release = "2.36.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index d7a0603a10..3ed8efd506 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1329,4 +1329,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.35.2" +VERSION = "2.36.0" diff --git a/setup.py b/setup.py index ecb24290c8..828dd43461 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.35.2", + version="2.36.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From ea097ede7d8c90b65b6412250070033b66aa8283 Mon Sep 17 00:00:00 2001 From: Alexander Alderman Webb Date: Thu, 4 Sep 2025 09:44:07 +0200 Subject: [PATCH 438/591] docs: Add snippet to configure sending unraisable exceptions to Sentry --- CHANGELOG.md | 16 +++++++++++++++- 1 file changed, 15 insertions(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a0b3c1647e..2c10ef8b7f 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,10 +4,24 @@ ### Various fixes & improvements +- **New integration:** Unraisable exceptions (#4733) by @alexander-alderman-webb + + Add the unraisable exception integration to your sentry_sdk.init call: +```python +import sentry_sdk +from sentry_sdk.integrations.unraisablehook import UnraisablehookIntegration + +sentry_sdk.init( + dsn="...", + integrations=[ + UnraisablehookIntegration(), + ] +) +``` + - meta: Update instructions on release process (#4755) by @sentrivana - tests: Move arq under toxgen (#4739) by @sentrivana - tests: Support dashes in test suite names (#4740) by @sentrivana -- feat(integrations): Add unraisable exception integration (#4733) by @alexander-alderman-webb - Don't fail if there is no `_context_manager_state` (#4698) by @sentrivana - Wrap span restoration in `__exit__` in `capture_internal_exceptions` (#4719) by @sentrivana - fix: Constrain types of ai_track decorator (#4745) by @alexander-alderman-webb From ff9b1c37f2ccbdad39e1950dab09546b29e2d247 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 4 Sep 2025 10:23:49 +0200 Subject: [PATCH 439/591] tests: Remove openai pin and update tox (#4748) [v1.104.2](https://github.com/openai/openai-python/releases/tag/v1.104.2) of openai added back the missing export that was causing problems for `openai_agents`, so we can remove the version pin again. --- scripts/populate_tox/config.py | 1 - tox.ini | 17 +++++++---------- 2 files changed, 7 insertions(+), 11 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f6093b0250..a2c4c8770c 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -188,7 +188,6 @@ "package": "openai-agents", "deps": { "*": ["pytest-asyncio"], - "<=0.2.10": ["openai<1.103.0"], }, "python": ">=3.10", }, diff --git a/tox.ini b/tox.ini index f2ad720a25..67ba6eadc6 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-03T15:01:21.035943+00:00 +# Last generated: 2025-09-04T07:00:53.509946+00:00 [tox] requires = @@ -148,7 +148,7 @@ envlist = {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 - {py3.10,py3.12,py3.13}-openai_agents-v0.2.10 + {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 @@ -313,8 +313,8 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.34 - {py3.9,py3.12,py3.13}-trytond-v7.6.5 + {py3.8,py3.11,py3.12}-trytond-v7.0.35 + {py3.9,py3.12,py3.13}-trytond-v7.6.6 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 @@ -530,11 +530,8 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 - openai_agents-v0.2.10: openai-agents==0.2.10 + openai_agents-v0.2.11: openai-agents==0.2.11 openai_agents: pytest-asyncio - openai_agents-v0.0.19: openai<1.103.0 - openai_agents-v0.1.0: openai<1.103.0 - openai_agents-v0.2.10: openai<1.103.0 huggingface_hub-v0.22.2: huggingface_hub==0.22.2 huggingface_hub-v0.26.5: huggingface_hub==0.26.5 @@ -790,8 +787,8 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.34: trytond==7.0.34 - trytond-v7.6.5: trytond==7.6.5 + trytond-v7.0.35: trytond==7.0.35 + trytond-v7.6.6: trytond==7.6.6 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From c378c2d8d9032a50f4371df20a1929756342b245 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 4 Sep 2025 13:45:07 +0200 Subject: [PATCH 440/591] tests: Move beam under toxgen (#4759) - move beam under toxgen - lower the waiting time between pypi requests Ref https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-tasks.yml | 2 +- scripts/populate_tox/config.py | 4 ++++ scripts/populate_tox/populate_tox.py | 3 +-- scripts/populate_tox/tox.jinja | 8 -------- tox.ini | 20 ++++++++++--------- 5 files changed, 17 insertions(+), 20 deletions(-) diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index a489f64410..f842683285 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.8","3.10","3.11","3.12","3.13"] + python-version: ["3.7","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index a2c4c8770c..689253e889 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -36,6 +36,10 @@ "<=0.23": ["pydantic<2"], }, }, + "beam": { + "package": "apache-beam", + "python": ">=3.7", + }, "bottle": { "package": "bottle", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index a8c58938ae..3d9ef23b66 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -40,7 +40,7 @@ lstrip_blocks=True, ) -PYPI_COOLDOWN = 0.15 # seconds to wait between requests to PyPI +PYPI_COOLDOWN = 0.1 # seconds to wait between requests to PyPI PYPI_PROJECT_URL = "https://pypi.python.org/pypi/%7Bproject%7D/json" PYPI_VERSION_URL = "https://pypi.python.org/pypi/%7Bproject%7D/%7Bversion%7D/json" @@ -68,7 +68,6 @@ # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list "asyncpg", - "beam", "boto3", "chalice", "gcp", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 115b99fd5c..65a5ba3f36 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -46,10 +46,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Beam - {py3.7}-beam-v{2.12} - {py3.8,py3.11}-beam-latest - # Boto3 {py3.6,py3.7}-boto3-v{1.12} {py3.7,py3.11,py3.12}-boto3-v{1.23} @@ -177,10 +173,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Beam - beam-v2.12: apache-beam~=2.12.0 - beam-latest: apache-beam - # Boto3 boto3-v1.12: boto3~=1.12.0 boto3-v1.23: boto3~=1.23.0 diff --git a/tox.ini b/tox.ini index 67ba6eadc6..fd633654be 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-04T07:00:53.509946+00:00 +# Last generated: 2025-09-04T10:35:13.756355+00:00 [tox] requires = @@ -46,10 +46,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Beam - {py3.7}-beam-v{2.12} - {py3.8,py3.11}-beam-latest - # Boto3 {py3.6,py3.7}-boto3-v{1.12} {py3.7,py3.11,py3.12}-boto3-v{1.23} @@ -228,6 +224,11 @@ envlist = {py3.7,py3.10,py3.11}-arq-v0.25.0 {py3.8,py3.11,py3.12}-arq-v0.26.3 + {py3.7}-beam-v2.14.0 + {py3.7,py3.8}-beam-v2.32.0 + {py3.8,py3.10,py3.11}-beam-v2.50.0 + {py3.9,py3.12,py3.13}-beam-v2.67.0 + {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.12,py3.13}-celery-v5.5.3 @@ -371,10 +372,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Beam - beam-v2.12: apache-beam~=2.12.0 - beam-latest: apache-beam - # Boto3 boto3-v1.12: boto3~=1.12.0 boto3-v1.23: boto3~=1.23.0 @@ -633,6 +630,11 @@ deps = arq: fakeredis>=2.2.0,<2.8 arq-v0.23: pydantic<2 + beam-v2.14.0: apache-beam==2.14.0 + beam-v2.32.0: apache-beam==2.32.0 + beam-v2.50.0: apache-beam==2.50.0 + beam-v2.67.0: apache-beam==2.67.0 + celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.5.3: celery==5.5.3 From 58a9827e1a5bb207a34651409a303bc21890fb66 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 4 Sep 2025 15:38:33 +0200 Subject: [PATCH 441/591] feat: Add LangGraph integration (#4727) - Add LangGraph integration - Compilation of StateGraphs results in an agent creation span (according to [OTEL semantic conventions](https://opentelemetry.io/docs/specs/semconv/gen-ai/gen-ai-agent-spans/)) - Runtime executions are done on Pregel instances - we are wrapping their invoke & ainvoke which produces the invoke_agent spans - There's some internals that automatically switch between invoke & stream on CompiledStateGraph (which is a subclass of Pregel), which results in duplicate spans if both are instrumented. For now, only invoke is wrapped to prevent this duplication. - Agent handoffs in LangGraph are done via tools - so there is no real possibility to create handoff spans within the SDK. Looks like this will be handled in product logic instead. Closes TET-991 Closes PY-1799 --------- Co-authored-by: Anton Pirker --- .github/workflows/test-integrations-ai.yml | 4 + pyproject.toml | 4 + scripts/populate_tox/config.py | 3 + scripts/populate_tox/tox.jinja | 1 + .../split_tox_gh_actions.py | 1 + sentry_sdk/consts.py | 1 + sentry_sdk/integrations/__init__.py | 2 + sentry_sdk/integrations/langgraph.py | 321 +++++++++ setup.py | 1 + tests/integrations/langgraph/__init__.py | 3 + .../integrations/langgraph/test_langgraph.py | 632 ++++++++++++++++++ tox.ini | 9 +- 12 files changed, 981 insertions(+), 1 deletion(-) create mode 100644 sentry_sdk/integrations/langgraph.py create mode 100644 tests/integrations/langgraph/__init__.py create mode 100644 tests/integrations/langgraph/test_langgraph.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 72a4253744..26a8bdb8bb 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -74,6 +74,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai-notiktoken" + - name: Test langgraph pinned + run: | + set -x # print commands that are executed + ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langgraph" - name: Test openai_agents pinned run: | set -x # print commands that are executed diff --git a/pyproject.toml b/pyproject.toml index deba247e39..44eded7641 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -130,6 +130,10 @@ ignore_missing_imports = true module = "langchain.*" ignore_missing_imports = true +[[tool.mypy.overrides]] +module = "langgraph.*" +ignore_missing_imports = true + [[tool.mypy.overrides]] module = "executing.*" ignore_missing_imports = true diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 689253e889..6795e36303 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -157,6 +157,9 @@ }, "include": "<1.0", }, + "langgraph": { + "package": "langgraph", + }, "launchdarkly": { "package": "launchdarkly-server-sdk", }, diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 65a5ba3f36..241e0ca288 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -338,6 +338,7 @@ setenv = huggingface_hub: TESTPATH=tests/integrations/huggingface_hub langchain-base: TESTPATH=tests/integrations/langchain langchain-notiktoken: TESTPATH=tests/integrations/langchain + langgraph: TESTPATH=tests/integrations/langgraph launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index cf83e0a3fe..51ee614d04 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -78,6 +78,7 @@ "langchain-notiktoken", "openai-base", "openai-notiktoken", + "langgraph", "openai_agents", "huggingface_hub", ], diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 3ed8efd506..5480ef5dce 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -792,6 +792,7 @@ class OP: FUNCTION_AWS = "function.aws" FUNCTION_GCP = "function.gcp" GEN_AI_CHAT = "gen_ai.chat" + GEN_AI_CREATE_AGENT = "gen_ai.create_agent" GEN_AI_EMBEDDINGS = "gen_ai.embeddings" GEN_AI_EXECUTE_TOOL = "gen_ai.execute_tool" GEN_AI_HANDOFF = "gen_ai.handoff" diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 6f0109aced..7f202221a7 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -95,6 +95,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "sentry_sdk.integrations.huey.HueyIntegration", "sentry_sdk.integrations.huggingface_hub.HuggingfaceHubIntegration", "sentry_sdk.integrations.langchain.LangchainIntegration", + "sentry_sdk.integrations.langgraph.LanggraphIntegration", "sentry_sdk.integrations.litestar.LitestarIntegration", "sentry_sdk.integrations.loguru.LoguruIntegration", "sentry_sdk.integrations.openai.OpenAIIntegration", @@ -142,6 +143,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "grpc": (1, 32, 0), # grpcio "huggingface_hub": (0, 22), "langchain": (0, 1, 0), + "langgraph": (0, 6, 6), "launchdarkly": (9, 8, 0), "loguru": (0, 7, 0), "openai": (1, 0, 0), diff --git a/sentry_sdk/integrations/langgraph.py b/sentry_sdk/integrations/langgraph.py new file mode 100644 index 0000000000..4b241fe895 --- /dev/null +++ b/sentry_sdk/integrations/langgraph.py @@ -0,0 +1,321 @@ +from functools import wraps +from typing import Any, Callable, List, Optional + +import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import safe_serialize + + +try: + from langgraph.graph import StateGraph + from langgraph.pregel import Pregel +except ImportError: + raise DidNotEnable("langgraph not installed") + + +class LanggraphIntegration(Integration): + identifier = "langgraph" + origin = f"auto.ai.{identifier}" + + def __init__(self, include_prompts=True): + # type: (LanggraphIntegration, bool) -> None + self.include_prompts = include_prompts + + @staticmethod + def setup_once(): + # type: () -> None + # LangGraph lets users create agents using a StateGraph or the Functional API. + # StateGraphs are then compiled to a CompiledStateGraph. Both CompiledStateGraph and + # the functional API execute on a Pregel instance. Pregel is the runtime for the graph + # and the invocation happens on Pregel, so patching the invoke methods takes care of both. + # The streaming methods are not patched, because due to some internal reasons, LangGraph + # will automatically patch the streaming methods to run through invoke, and by doing this + # we prevent duplicate spans for invocations. + StateGraph.compile = _wrap_state_graph_compile(StateGraph.compile) + if hasattr(Pregel, "invoke"): + Pregel.invoke = _wrap_pregel_invoke(Pregel.invoke) + if hasattr(Pregel, "ainvoke"): + Pregel.ainvoke = _wrap_pregel_ainvoke(Pregel.ainvoke) + + +def _get_graph_name(graph_obj): + # type: (Any) -> Optional[str] + for attr in ["name", "graph_name", "__name__", "_name"]: + if hasattr(graph_obj, attr): + name = getattr(graph_obj, attr) + if name and isinstance(name, str): + return name + return None + + +def _normalize_langgraph_message(message): + # type: (Any) -> Any + if not hasattr(message, "content"): + return None + + parsed = {"role": getattr(message, "type", None), "content": message.content} + + for attr in ["name", "tool_calls", "function_call", "tool_call_id"]: + if hasattr(message, attr): + value = getattr(message, attr) + if value is not None: + parsed[attr] = value + + return parsed + + +def _parse_langgraph_messages(state): + # type: (Any) -> Optional[List[Any]] + if not state: + return None + + messages = None + + if isinstance(state, dict): + messages = state.get("messages") + elif hasattr(state, "messages"): + messages = state.messages + elif hasattr(state, "get") and callable(state.get): + try: + messages = state.get("messages") + except Exception: + pass + + if not messages or not isinstance(messages, (list, tuple)): + return None + + normalized_messages = [] + for message in messages: + try: + normalized = _normalize_langgraph_message(message) + if normalized: + normalized_messages.append(normalized) + except Exception: + continue + + return normalized_messages if normalized_messages else None + + +def _wrap_state_graph_compile(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + @wraps(f) + def new_compile(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(LanggraphIntegration) + if integration is None: + return f(self, *args, **kwargs) + with sentry_sdk.start_span( + op=OP.GEN_AI_CREATE_AGENT, + origin=LanggraphIntegration.origin, + ) as span: + compiled_graph = f(self, *args, **kwargs) + + compiled_graph_name = getattr(compiled_graph, "name", None) + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "create_agent") + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, compiled_graph_name) + + if compiled_graph_name: + span.description = f"create_agent {compiled_graph_name}" + else: + span.description = "create_agent" + + if kwargs.get("model", None) is not None: + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, kwargs.get("model")) + + tools = None + get_graph = getattr(compiled_graph, "get_graph", None) + if get_graph and callable(get_graph): + graph_obj = compiled_graph.get_graph() + nodes = getattr(graph_obj, "nodes", None) + if nodes and isinstance(nodes, dict): + tools_node = nodes.get("tools") + if tools_node: + data = getattr(tools_node, "data", None) + if data and hasattr(data, "tools_by_name"): + tools = list(data.tools_by_name.keys()) + + if tools is not None: + span.set_data(SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, tools) + + return compiled_graph + + return new_compile + + +def _wrap_pregel_invoke(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + + @wraps(f) + def new_invoke(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(LanggraphIntegration) + if integration is None: + return f(self, *args, **kwargs) + + graph_name = _get_graph_name(self) + span_name = ( + f"invoke_agent {graph_name}".strip() if graph_name else "invoke_agent" + ) + + with sentry_sdk.start_span( + op=OP.GEN_AI_INVOKE_AGENT, + name=span_name, + origin=LanggraphIntegration.origin, + ) as span: + if graph_name: + span.set_data(SPANDATA.GEN_AI_PIPELINE_NAME, graph_name) + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, graph_name) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + + # Store input messages to later compare with output + input_messages = None + if ( + len(args) > 0 + and should_send_default_pii() + and integration.include_prompts + ): + input_messages = _parse_langgraph_messages(args[0]) + if input_messages: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + safe_serialize(input_messages), + ) + + result = f(self, *args, **kwargs) + + _set_response_attributes(span, input_messages, result, integration) + + return result + + return new_invoke + + +def _wrap_pregel_ainvoke(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + + @wraps(f) + async def new_ainvoke(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(LanggraphIntegration) + if integration is None: + return await f(self, *args, **kwargs) + + graph_name = _get_graph_name(self) + span_name = ( + f"invoke_agent {graph_name}".strip() if graph_name else "invoke_agent" + ) + + with sentry_sdk.start_span( + op=OP.GEN_AI_INVOKE_AGENT, + name=span_name, + origin=LanggraphIntegration.origin, + ) as span: + if graph_name: + span.set_data(SPANDATA.GEN_AI_PIPELINE_NAME, graph_name) + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, graph_name) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + + input_messages = None + if ( + len(args) > 0 + and should_send_default_pii() + and integration.include_prompts + ): + input_messages = _parse_langgraph_messages(args[0]) + if input_messages: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + safe_serialize(input_messages), + ) + + result = await f(self, *args, **kwargs) + + _set_response_attributes(span, input_messages, result, integration) + + return result + + return new_ainvoke + + +def _get_new_messages(input_messages, output_messages): + # type: (Optional[List[Any]], Optional[List[Any]]) -> Optional[List[Any]] + """Extract only the new messages added during this invocation.""" + if not output_messages: + return None + + if not input_messages: + return output_messages + + # only return the new messages, aka the output messages that are not in the input messages + input_count = len(input_messages) + new_messages = ( + output_messages[input_count:] if len(output_messages) > input_count else [] + ) + + return new_messages if new_messages else None + + +def _extract_llm_response_text(messages): + # type: (Optional[List[Any]]) -> Optional[str] + if not messages: + return None + + for message in reversed(messages): + if isinstance(message, dict): + role = message.get("role") + if role in ["assistant", "ai"]: + content = message.get("content") + if content and isinstance(content, str): + return content + + return None + + +def _extract_tool_calls(messages): + # type: (Optional[List[Any]]) -> Optional[List[Any]] + if not messages: + return None + + tool_calls = [] + for message in messages: + if isinstance(message, dict): + msg_tool_calls = message.get("tool_calls") + if msg_tool_calls and isinstance(msg_tool_calls, list): + tool_calls.extend(msg_tool_calls) + + return tool_calls if tool_calls else None + + +def _set_response_attributes(span, input_messages, result, integration): + # type: (Any, Optional[List[Any]], Any, LanggraphIntegration) -> None + if not (should_send_default_pii() and integration.include_prompts): + return + + parsed_response_messages = _parse_langgraph_messages(result) + new_messages = _get_new_messages(input_messages, parsed_response_messages) + + llm_response_text = _extract_llm_response_text(new_messages) + if llm_response_text: + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, llm_response_text) + elif new_messages: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(new_messages) + ) + else: + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(result)) + + tool_calls = _extract_tool_calls(new_messages) + if tool_calls: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + safe_serialize(tool_calls), + unpack=False, + ) diff --git a/setup.py b/setup.py index 828dd43461..ca6e7ec534 100644 --- a/setup.py +++ b/setup.py @@ -63,6 +63,7 @@ def get_file_text(file_name): "huey": ["huey>=2"], "huggingface_hub": ["huggingface_hub>=0.22"], "langchain": ["langchain>=0.0.210"], + "langgraph": ["langgraph>=0.6.6"], "launchdarkly": ["launchdarkly-server-sdk>=9.8.0"], "litestar": ["litestar>=2.0.0"], "loguru": ["loguru>=0.5"], diff --git a/tests/integrations/langgraph/__init__.py b/tests/integrations/langgraph/__init__.py new file mode 100644 index 0000000000..b7dd1cb562 --- /dev/null +++ b/tests/integrations/langgraph/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("langgraph") diff --git a/tests/integrations/langgraph/test_langgraph.py b/tests/integrations/langgraph/test_langgraph.py new file mode 100644 index 0000000000..5e35f772f5 --- /dev/null +++ b/tests/integrations/langgraph/test_langgraph.py @@ -0,0 +1,632 @@ +import asyncio +import sys +from unittest.mock import MagicMock, patch + +import pytest + +from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANDATA, OP + + +def mock_langgraph_imports(): + """Mock langgraph modules to prevent import errors.""" + mock_state_graph = MagicMock() + mock_pregel = MagicMock() + + langgraph_graph_mock = MagicMock() + langgraph_graph_mock.StateGraph = mock_state_graph + + langgraph_pregel_mock = MagicMock() + langgraph_pregel_mock.Pregel = mock_pregel + + sys.modules["langgraph"] = MagicMock() + sys.modules["langgraph.graph"] = langgraph_graph_mock + sys.modules["langgraph.pregel"] = langgraph_pregel_mock + + return mock_state_graph, mock_pregel + + +mock_state_graph, mock_pregel = mock_langgraph_imports() + +from sentry_sdk.integrations.langgraph import ( # noqa: E402 + LanggraphIntegration, + _parse_langgraph_messages, + _wrap_state_graph_compile, + _wrap_pregel_invoke, + _wrap_pregel_ainvoke, +) + + +class MockStateGraph: + def __init__(self, schema=None): + self.name = "test_graph" + self.schema = schema + self._compiled_graph = None + + def compile(self, *args, **kwargs): + compiled = MockCompiledGraph(self.name) + compiled.graph = self + return compiled + + +class MockCompiledGraph: + def __init__(self, name="test_graph"): + self.name = name + self._graph = None + + def get_graph(self): + return MockGraphRepresentation() + + def invoke(self, state, config=None): + return {"messages": [MockMessage("Response from graph")]} + + async def ainvoke(self, state, config=None): + return {"messages": [MockMessage("Async response from graph")]} + + +class MockGraphRepresentation: + def __init__(self): + self.nodes = {"tools": MockToolsNode()} + + +class MockToolsNode: + def __init__(self): + self.data = MockToolsData() + + +class MockToolsData: + def __init__(self): + self.tools_by_name = { + "search_tool": MockTool("search_tool"), + "calculator": MockTool("calculator"), + } + + +class MockTool: + def __init__(self, name): + self.name = name + + +class MockMessage: + def __init__( + self, + content, + name=None, + tool_calls=None, + function_call=None, + role=None, + type=None, + ): + self.content = content + self.name = name + self.tool_calls = tool_calls + self.function_call = function_call + self.role = role + # The integration uses getattr(message, "type", None) for the role in _normalize_langgraph_message + # Set default type based on name if type not explicitly provided + if type is None and name in ["assistant", "ai", "user", "system", "function"]: + self.type = name + else: + self.type = type + + +class MockPregelInstance: + def __init__(self, name="test_pregel"): + self.name = name + self.graph_name = name + + def invoke(self, state, config=None): + return {"messages": [MockMessage("Pregel response")]} + + async def ainvoke(self, state, config=None): + return {"messages": [MockMessage("Async Pregel response")]} + + +def test_langgraph_integration_init(): + """Test LanggraphIntegration initialization with different parameters.""" + integration = LanggraphIntegration() + assert integration.include_prompts is True + assert integration.identifier == "langgraph" + assert integration.origin == "auto.ai.langgraph" + + integration = LanggraphIntegration(include_prompts=False) + assert integration.include_prompts is False + assert integration.identifier == "langgraph" + assert integration.origin == "auto.ai.langgraph" + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_state_graph_compile( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test StateGraph.compile() wrapper creates proper create_agent span.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + graph = MockStateGraph() + + def original_compile(self, *args, **kwargs): + return MockCompiledGraph(self.name) + + with patch("sentry_sdk.integrations.langgraph.StateGraph"): + with start_transaction(): + wrapped_compile = _wrap_state_graph_compile(original_compile) + compiled_graph = wrapped_compile( + graph, model="test-model", checkpointer=None + ) + + assert compiled_graph is not None + assert compiled_graph.name == "test_graph" + + tx = events[0] + assert tx["type"] == "transaction" + + agent_spans = [span for span in tx["spans"] if span["op"] == OP.GEN_AI_CREATE_AGENT] + assert len(agent_spans) == 1 + + agent_span = agent_spans[0] + assert agent_span["description"] == "create_agent test_graph" + assert agent_span["origin"] == "auto.ai.langgraph" + assert agent_span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "create_agent" + assert agent_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == "test_graph" + assert agent_span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "test-model" + assert SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS in agent_span["data"] + + tools_data = agent_span["data"][SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS] + assert tools_data == ["search_tool", "calculator"] + assert len(tools_data) == 2 + assert "search_tool" in tools_data + assert "calculator" in tools_data + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_pregel_invoke(sentry_init, capture_events, send_default_pii, include_prompts): + """Test Pregel.invoke() wrapper creates proper invoke_agent span.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + test_state = { + "messages": [ + MockMessage("Hello, can you help me?", name="user"), + MockMessage("Of course! How can I assist you?", name="assistant"), + ] + } + + pregel = MockPregelInstance("test_graph") + + expected_assistant_response = "I'll help you with that task!" + expected_tool_calls = [ + { + "id": "call_test_123", + "type": "function", + "function": {"name": "search_tool", "arguments": '{"query": "help"}'}, + } + ] + + def original_invoke(self, *args, **kwargs): + input_messages = args[0].get("messages", []) + new_messages = input_messages + [ + MockMessage( + content=expected_assistant_response, + name="assistant", + tool_calls=expected_tool_calls, + ) + ] + return {"messages": new_messages} + + with start_transaction(): + wrapped_invoke = _wrap_pregel_invoke(original_invoke) + result = wrapped_invoke(pregel, test_state) + + assert result is not None + + tx = events[0] + assert tx["type"] == "transaction" + + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + assert invoke_span["description"] == "invoke_agent test_graph" + assert invoke_span["origin"] == "auto.ai.langgraph" + assert invoke_span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "invoke_agent" + assert invoke_span["data"][SPANDATA.GEN_AI_PIPELINE_NAME] == "test_graph" + assert invoke_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == "test_graph" + + if send_default_pii and include_prompts: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in invoke_span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT in invoke_span["data"] + + request_messages = invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + + if isinstance(request_messages, str): + import json + + request_messages = json.loads(request_messages) + assert len(request_messages) == 2 + assert request_messages[0]["content"] == "Hello, can you help me?" + assert request_messages[1]["content"] == "Of course! How can I assist you?" + + response_text = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + assert response_text == expected_assistant_response + + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in invoke_span["data"] + tool_calls_data = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS] + if isinstance(tool_calls_data, str): + import json + + tool_calls_data = json.loads(tool_calls_data) + + assert len(tool_calls_data) == 1 + assert tool_calls_data[0]["id"] == "call_test_123" + assert tool_calls_data[0]["function"]["name"] == "search_tool" + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in invoke_span.get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in invoke_span.get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS not in invoke_span.get("data", {}) + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_pregel_ainvoke(sentry_init, capture_events, send_default_pii, include_prompts): + """Test Pregel.ainvoke() async wrapper creates proper invoke_agent span.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + test_state = {"messages": [MockMessage("What's the weather like?", name="user")]} + pregel = MockPregelInstance("async_graph") + + expected_assistant_response = "It's sunny and 72°F today!" + expected_tool_calls = [ + { + "id": "call_weather_456", + "type": "function", + "function": {"name": "get_weather", "arguments": '{"location": "current"}'}, + } + ] + + async def original_ainvoke(self, *args, **kwargs): + input_messages = args[0].get("messages", []) + new_messages = input_messages + [ + MockMessage( + content=expected_assistant_response, + name="assistant", + tool_calls=expected_tool_calls, + ) + ] + return {"messages": new_messages} + + async def run_test(): + with start_transaction(): + + wrapped_ainvoke = _wrap_pregel_ainvoke(original_ainvoke) + result = await wrapped_ainvoke(pregel, test_state) + return result + + result = asyncio.run(run_test()) + assert result is not None + + tx = events[0] + assert tx["type"] == "transaction" + + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + assert invoke_span["description"] == "invoke_agent async_graph" + assert invoke_span["origin"] == "auto.ai.langgraph" + assert invoke_span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "invoke_agent" + assert invoke_span["data"][SPANDATA.GEN_AI_PIPELINE_NAME] == "async_graph" + assert invoke_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == "async_graph" + + if send_default_pii and include_prompts: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in invoke_span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT in invoke_span["data"] + + response_text = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + assert response_text == expected_assistant_response + + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in invoke_span["data"] + tool_calls_data = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS] + if isinstance(tool_calls_data, str): + import json + + tool_calls_data = json.loads(tool_calls_data) + + assert len(tool_calls_data) == 1 + assert tool_calls_data[0]["id"] == "call_weather_456" + assert tool_calls_data[0]["function"]["name"] == "get_weather" + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in invoke_span.get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in invoke_span.get("data", {}) + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS not in invoke_span.get("data", {}) + + +def test_pregel_invoke_error(sentry_init, capture_events): + """Test error handling during graph execution.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + test_state = {"messages": [MockMessage("This will fail")]} + pregel = MockPregelInstance("error_graph") + + def original_invoke(self, *args, **kwargs): + raise Exception("Graph execution failed") + + with start_transaction(), pytest.raises(Exception, match="Graph execution failed"): + + wrapped_invoke = _wrap_pregel_invoke(original_invoke) + wrapped_invoke(pregel, test_state) + + tx = events[0] + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + assert invoke_span.get("tags", {}).get("status") == "internal_error" + + +def test_pregel_ainvoke_error(sentry_init, capture_events): + """Test error handling during async graph execution.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + test_state = {"messages": [MockMessage("This will fail async")]} + pregel = MockPregelInstance("async_error_graph") + + async def original_ainvoke(self, *args, **kwargs): + raise Exception("Async graph execution failed") + + async def run_error_test(): + with start_transaction(), pytest.raises( + Exception, match="Async graph execution failed" + ): + + wrapped_ainvoke = _wrap_pregel_ainvoke(original_ainvoke) + await wrapped_ainvoke(pregel, test_state) + + asyncio.run(run_error_test()) + + tx = events[0] + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + assert invoke_span.get("tags", {}).get("status") == "internal_error" + + +def test_span_origin(sentry_init, capture_events): + """Test that span origins are correctly set.""" + sentry_init( + integrations=[LanggraphIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + graph = MockStateGraph() + + def original_compile(self, *args, **kwargs): + return MockCompiledGraph(self.name) + + with start_transaction(): + from sentry_sdk.integrations.langgraph import _wrap_state_graph_compile + + wrapped_compile = _wrap_state_graph_compile(original_compile) + wrapped_compile(graph) + + tx = events[0] + assert tx["contexts"]["trace"]["origin"] == "manual" + + for span in tx["spans"]: + assert span["origin"] == "auto.ai.langgraph" + + +@pytest.mark.parametrize("graph_name", ["my_graph", None, ""]) +def test_pregel_invoke_with_different_graph_names( + sentry_init, capture_events, graph_name +): + """Test Pregel.invoke() with different graph name scenarios.""" + sentry_init( + integrations=[LanggraphIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + pregel = MockPregelInstance(graph_name) if graph_name else MockPregelInstance() + if not graph_name: + + delattr(pregel, "name") + delattr(pregel, "graph_name") + + def original_invoke(self, *args, **kwargs): + return {"result": "test"} + + with start_transaction(): + + wrapped_invoke = _wrap_pregel_invoke(original_invoke) + wrapped_invoke(pregel, {"messages": []}) + + tx = events[0] + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + + if graph_name and graph_name.strip(): + assert invoke_span["description"] == "invoke_agent my_graph" + assert invoke_span["data"][SPANDATA.GEN_AI_PIPELINE_NAME] == graph_name + assert invoke_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == graph_name + else: + assert invoke_span["description"] == "invoke_agent" + assert SPANDATA.GEN_AI_PIPELINE_NAME not in invoke_span.get("data", {}) + assert SPANDATA.GEN_AI_AGENT_NAME not in invoke_span.get("data", {}) + + +def test_complex_message_parsing(): + """Test message parsing with complex message structures.""" + messages = [ + MockMessage(content="User query", name="user"), + MockMessage( + content="Assistant response with tools", + name="assistant", + tool_calls=[ + { + "id": "call_1", + "type": "function", + "function": {"name": "search", "arguments": "{}"}, + }, + { + "id": "call_2", + "type": "function", + "function": {"name": "calculate", "arguments": '{"x": 5}'}, + }, + ], + ), + MockMessage( + content="Function call response", + name="function", + function_call={"name": "search", "arguments": '{"query": "test"}'}, + ), + ] + + state = {"messages": messages} + result = _parse_langgraph_messages(state) + + assert result is not None + assert len(result) == 3 + + assert result[0]["content"] == "User query" + assert result[0]["name"] == "user" + assert "tool_calls" not in result[0] + assert "function_call" not in result[0] + + assert result[1]["content"] == "Assistant response with tools" + assert result[1]["name"] == "assistant" + assert len(result[1]["tool_calls"]) == 2 + + assert result[2]["content"] == "Function call response" + assert result[2]["name"] == "function" + assert result[2]["function_call"]["name"] == "search" + + +def test_extraction_functions_complex_scenario(sentry_init, capture_events): + """Test extraction functions with complex scenarios including multiple messages and edge cases.""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + pregel = MockPregelInstance("complex_graph") + test_state = {"messages": [MockMessage("Complex request", name="user")]} + + def original_invoke(self, *args, **kwargs): + input_messages = args[0].get("messages", []) + new_messages = input_messages + [ + MockMessage( + content="I'll help with multiple tasks", + name="assistant", + tool_calls=[ + { + "id": "call_multi_1", + "type": "function", + "function": { + "name": "search", + "arguments": '{"query": "complex"}', + }, + }, + { + "id": "call_multi_2", + "type": "function", + "function": { + "name": "calculate", + "arguments": '{"expr": "2+2"}', + }, + }, + ], + ), + MockMessage("", name="assistant"), + MockMessage("Final response", name="ai", type="ai"), + ] + return {"messages": new_messages} + + with start_transaction(): + wrapped_invoke = _wrap_pregel_invoke(original_invoke) + result = wrapped_invoke(pregel, test_state) + + assert result is not None + + tx = events[0] + invoke_spans = [ + span for span in tx["spans"] if span["op"] == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) == 1 + + invoke_span = invoke_spans[0] + assert SPANDATA.GEN_AI_RESPONSE_TEXT in invoke_span["data"] + response_text = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + assert response_text == "Final response" + + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in invoke_span["data"] + import json + + tool_calls_data = invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS] + if isinstance(tool_calls_data, str): + tool_calls_data = json.loads(tool_calls_data) + + assert len(tool_calls_data) == 2 + assert tool_calls_data[0]["id"] == "call_multi_1" + assert tool_calls_data[0]["function"]["name"] == "search" + assert tool_calls_data[1]["id"] == "call_multi_2" + assert tool_calls_data[1]["function"]["name"] == "calculate" diff --git a/tox.ini b/tox.ini index fd633654be..40afc2a6a7 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-04T10:35:13.756355+00:00 +# Last generated: 2025-09-04T12:59:44.328902+00:00 [tox] requires = @@ -142,6 +142,9 @@ envlist = {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.105.0 + {py3.9,py3.12,py3.13}-langgraph-v0.6.6 + {py3.10,py3.12,py3.13}-langgraph-v1.0.0a2 + {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 @@ -525,6 +528,9 @@ deps = openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 + langgraph-v0.6.6: langgraph==0.6.6 + langgraph-v1.0.0a2: langgraph==1.0.0a2 + openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 @@ -841,6 +847,7 @@ setenv = huggingface_hub: TESTPATH=tests/integrations/huggingface_hub langchain-base: TESTPATH=tests/integrations/langchain langchain-notiktoken: TESTPATH=tests/integrations/langchain + langgraph: TESTPATH=tests/integrations/langgraph launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru From 9711b3be884263bb34f1ae5a0f719cb9acb4b0ca Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 4 Sep 2025 16:08:26 +0200 Subject: [PATCH 442/591] tests: Move asyncpg under toxgen (#4757) - remove hardcoded asyncpg config, let toxgen take care of generating it - isolate DB so that multiple asyncpg test suites for different envs can run at the same time without touching the same DB - update instructions on running the test suite locally Ref https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-dbs.yml | 2 +- scripts/populate_tox/config.py | 7 ++++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 9 ----- tests/integrations/asyncpg/test_asyncpg.py | 38 ++++++++++++++++----- tox.ini | 30 ++++++++-------- 6 files changed, 53 insertions(+), 34 deletions(-) diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 5fc0be029b..2d6af43bc3 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.8","3.11","3.12","3.13"] + python-version: ["3.7","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 6795e36303..1dbc78ccf0 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -36,6 +36,13 @@ "<=0.23": ["pydantic<2"], }, }, + "asyncpg": { + "package": "asyncpg", + "deps": { + "*": ["pytest-asyncio"], + }, + "python": ">=3.7", + }, "beam": { "package": "apache-beam", "python": ">=3.7", diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 3d9ef23b66..076a8358f7 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "asyncpg", "boto3", "chalice", "gcp", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 241e0ca288..0ad9af8321 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -39,10 +39,6 @@ envlist = # Asgi {py3.7,py3.12,py3.13}-asgi - # asyncpg - {py3.7,py3.10}-asyncpg-v{0.23} - {py3.8,py3.11,py3.12}-asyncpg-latest - # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -160,11 +156,6 @@ deps = asgi: pytest-asyncio asgi: async-asgi-testclient - # Asyncpg - asyncpg-v0.23: asyncpg~=0.23.0 - asyncpg-latest: asyncpg - asyncpg: pytest-asyncio - # AWS Lambda aws_lambda: aws-cdk-lib aws_lambda: aws-sam-cli diff --git a/tests/integrations/asyncpg/test_asyncpg.py b/tests/integrations/asyncpg/test_asyncpg.py index e36d15c5d2..e23612c055 100644 --- a/tests/integrations/asyncpg/test_asyncpg.py +++ b/tests/integrations/asyncpg/test_asyncpg.py @@ -3,21 +3,13 @@ Tests need a local postgresql instance running, this can best be done using ```sh -docker run --rm --name some-postgres -e POSTGRES_USER=foo -e POSTGRES_PASSWORD=bar -d -p 5432:5432 postgres +docker run --rm --name some-postgres -e POSTGRES_USER=postgres -e POSTGRES_PASSWORD=sentry -d -p 5432:5432 postgres ``` The tests use the following credentials to establish a database connection. """ import os - - -PG_HOST = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_HOST", "localhost") -PG_PORT = int(os.getenv("SENTRY_PYTHON_TEST_POSTGRES_PORT", "5432")) -PG_USER = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_USER", "postgres") -PG_PASSWORD = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_PASSWORD", "sentry") -PG_NAME = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_NAME", "postgres") - import datetime from contextlib import contextmanager from unittest import mock @@ -33,6 +25,19 @@ from sentry_sdk.tracing_utils import record_sql_queries from tests.conftest import ApproxDict +PG_HOST = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_HOST", "localhost") +PG_PORT = int(os.getenv("SENTRY_PYTHON_TEST_POSTGRES_PORT", "5432")) +PG_USER = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_USER", "postgres") +PG_PASSWORD = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_PASSWORD", "sentry") +PG_NAME_BASE = os.getenv("SENTRY_PYTHON_TEST_POSTGRES_NAME", "postgres") + + +def _get_db_name(): + pid = os.getpid() + return f"{PG_NAME_BASE}_{pid}" + + +PG_NAME = _get_db_name() PG_CONNECTION_URI = "postgresql://{}:{}@{}/{}".format( PG_USER, PG_PASSWORD, PG_HOST, PG_NAME @@ -55,6 +60,21 @@ @pytest_asyncio.fixture(autouse=True) async def _clean_pg(): + # Create the test database if it doesn't exist + default_conn = await connect( + "postgresql://{}:{}@{}".format(PG_USER, PG_PASSWORD, PG_HOST) + ) + try: + # Check if database exists, create if not + result = await default_conn.fetchval( + "SELECT 1 FROM pg_database WHERE datname = $1", PG_NAME + ) + if not result: + await default_conn.execute(f'CREATE DATABASE "{PG_NAME}"') + finally: + await default_conn.close() + + # Now connect to our test database and set up the table conn = await connect(PG_CONNECTION_URI) await conn.execute("DROP TABLE IF EXISTS users") await conn.execute( diff --git a/tox.ini b/tox.ini index 40afc2a6a7..1627cf2458 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-04T12:59:44.328902+00:00 +# Last generated: 2025-09-04T13:56:54.117272+00:00 [tox] requires = @@ -39,10 +39,6 @@ envlist = # Asgi {py3.7,py3.12,py3.13}-asgi - # asyncpg - {py3.7,py3.10}-asyncpg-v{0.23} - {py3.8,py3.11,py3.12}-asyncpg-latest - # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -135,12 +131,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.36.1 {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.105.0 + {py3.8,py3.12,py3.13}-openai-base-v1.106.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.105.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.106.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.6 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a2 @@ -157,6 +153,11 @@ envlist = # ~~~ DBs ~~~ + {py3.7,py3.8,py3.9}-asyncpg-v0.23.0 + {py3.7,py3.9,py3.10}-asyncpg-v0.25.0 + {py3.7,py3.9,py3.10}-asyncpg-v0.27.0 + {py3.8,py3.11,py3.12}-asyncpg-v0.30.0 + {py3.7,py3.11,py3.12}-clickhouse_driver-v0.2.9 {py3.6}-pymongo-v3.5.1 @@ -362,11 +363,6 @@ deps = asgi: pytest-asyncio asgi: async-asgi-testclient - # Asyncpg - asyncpg-v0.23: asyncpg~=0.23.0 - asyncpg-latest: asyncpg - asyncpg: pytest-asyncio - # AWS Lambda aws_lambda: aws-cdk-lib aws_lambda: aws-sam-cli @@ -514,7 +510,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.36.1: openai==1.36.1 openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.105.0: openai==1.105.0 + openai-base-v1.106.0: openai==1.106.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -523,7 +519,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.36.1: openai==1.36.1 openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.105.0: openai==1.105.0 + openai-notiktoken-v1.106.0: openai==1.106.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 @@ -544,6 +540,12 @@ deps = # ~~~ DBs ~~~ + asyncpg-v0.23.0: asyncpg==0.23.0 + asyncpg-v0.25.0: asyncpg==0.25.0 + asyncpg-v0.27.0: asyncpg==0.27.0 + asyncpg-v0.30.0: asyncpg==0.30.0 + asyncpg: pytest-asyncio + clickhouse_driver-v0.2.9: clickhouse-driver==0.2.9 pymongo-v3.5.1: pymongo==3.5.1 From b50f7e4a68c69caccdf29bc6b645f51c215e7ada Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 5 Sep 2025 08:46:08 +0200 Subject: [PATCH 443/591] Format span attributes in AI integrations (#4762) The AI Agents integrations render stringified json-like data in a nice way (make the sub nodes of the data structure collapsible) In Javascript it comes down to having double quotes in a string: - Good: `'{"role": "system", "content": "some context"}'` - Bad: `"{'role': 'system', 'content': 'some context'}"` Also pydantics `model_dump()` sometimes returns `function` or `class` objects that can not be json serialized so I updated `_normalize_data()` to make sure everything is converted to a primitive data type, always. --- sentry_sdk/ai/utils.py | 15 +++++++++------ tests/integrations/cohere/test_cohere.py | 8 ++++---- 2 files changed, 13 insertions(+), 10 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index cf52cba6e8..2dc0de4ef3 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -1,30 +1,33 @@ +import json + from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Any + from sentry_sdk.tracing import Span -from sentry_sdk.tracing import Span from sentry_sdk.utils import logger def _normalize_data(data, unpack=True): # type: (Any, bool) -> Any - # convert pydantic data (e.g. OpenAI v1+) to json compatible format if hasattr(data, "model_dump"): try: - return data.model_dump() + return _normalize_data(data.model_dump(), unpack=unpack) except Exception as e: logger.warning("Could not convert pydantic data to JSON: %s", e) - return data + return data if isinstance(data, (int, float, bool, str)) else str(data) + if isinstance(data, list): if unpack and len(data) == 1: return _normalize_data(data[0], unpack=unpack) # remove empty dimensions return list(_normalize_data(x, unpack=unpack) for x in data) + if isinstance(data, dict): return {k: _normalize_data(v, unpack=unpack) for (k, v) in data.items()} - return data + return data if isinstance(data, (int, float, bool, str)) else str(data) def set_data_normalized(span, key, value, unpack=True): @@ -33,4 +36,4 @@ def set_data_normalized(span, key, value, unpack=True): if isinstance(normalized, (int, float, bool, str)): span.set_data(key, normalized) else: - span.set_data(key, str(normalized)) + span.set_data(key, json.dumps(normalized)) diff --git a/tests/integrations/cohere/test_cohere.py b/tests/integrations/cohere/test_cohere.py index b8b6067625..ee876172d1 100644 --- a/tests/integrations/cohere/test_cohere.py +++ b/tests/integrations/cohere/test_cohere.py @@ -58,11 +58,11 @@ def test_nonstreaming_chat( if send_default_pii and include_prompts: assert ( - "{'role': 'system', 'content': 'some context'}" + '{"role": "system", "content": "some context"}' in span["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert ( - "{'role': 'user', 'content': 'hello'}" + '{"role": "user", "content": "hello"}' in span["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] @@ -135,11 +135,11 @@ def test_streaming_chat(sentry_init, capture_events, send_default_pii, include_p if send_default_pii and include_prompts: assert ( - "{'role': 'system', 'content': 'some context'}" + '{"role": "system", "content": "some context"}' in span["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert ( - "{'role': 'user', 'content': 'hello'}" + '{"role": "user", "content": "hello"}' in span["data"][SPANDATA.AI_INPUT_MESSAGES] ) assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] From 0c0a8d8497647e40ae8b285f5a53069394b084ad Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 5 Sep 2025 09:06:19 +0200 Subject: [PATCH 444/591] ci: Fix celery (#4765) We have some newrelic interference/compatibility tests that were failing with the newest newrelic release. Looking at that release, newrelic completely [rehauled](https://github.com/newrelic/newrelic-python-agent/commit/3cfce55a51ec0cf81919ebd475765707d39c90e0) their celery instrumentation, so I'm pinning our tests to only test against older newrelic versions where we had the problem in the first place. Rerunning toxgen on the updated config also pulled in a new openai release. --- scripts/populate_tox/config.py | 2 +- tox.ini | 12 ++++++------ 2 files changed, 7 insertions(+), 7 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 1dbc78ccf0..921098e7e6 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -56,7 +56,7 @@ "celery": { "package": "celery", "deps": { - "*": ["newrelic", "redis"], + "*": ["newrelic<10.17.0", "redis"], "py3.7": ["importlib-metadata<5.0"], }, }, diff --git a/tox.ini b/tox.ini index 1627cf2458..994ad22314 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-04T13:56:54.117272+00:00 +# Last generated: 2025-09-05T06:53:57.545461+00:00 [tox] requires = @@ -131,12 +131,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.36.1 {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.106.0 + {py3.8,py3.12,py3.13}-openai-base-v1.106.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.106.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.106.1 {py3.9,py3.12,py3.13}-langgraph-v0.6.6 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a2 @@ -510,7 +510,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.36.1: openai==1.36.1 openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.106.0: openai==1.106.0 + openai-base-v1.106.1: openai==1.106.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -519,7 +519,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.36.1: openai==1.36.1 openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.106.0: openai==1.106.0 + openai-notiktoken-v1.106.1: openai==1.106.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 @@ -646,7 +646,7 @@ deps = celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.5.3: celery==5.5.3 - celery: newrelic + celery: newrelic<10.17.0 celery: redis py3.7-celery: importlib-metadata<5.0 From ad3c435398d78949eda68dff66ef8eb8b4928679 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 5 Sep 2025 09:44:32 +0200 Subject: [PATCH 445/591] tests: Move boto3 tests under toxgen (#4761) Move boto3 under toxgen. Also, update how Python version constraints are generated. Ref https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-cloud.yml | 4 +- scripts/populate_tox/config.py | 6 +++ scripts/populate_tox/populate_tox.py | 3 +- scripts/populate_tox/tox.jinja | 12 ------ tox.ini | 39 ++++++++++--------- 5 files changed, 30 insertions(+), 34 deletions(-) diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index a04d57497a..8688a1d48e 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.11","3.12","3.13"] + python-version: ["3.8","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -108,7 +108,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 921098e7e6..5aba82b11b 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -47,6 +47,12 @@ "package": "apache-beam", "python": ">=3.7", }, + "boto3": { + "package": "boto3", + "deps": { + "py3.7,py3.8": ["urllib3<2.0.0"], + }, + }, "bottle": { "package": "bottle", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 076a8358f7..b8cc988fda 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "boto3", "chalice", "gcp", "httpx", @@ -439,7 +438,7 @@ def _render_dependencies(integration: str, releases: list[Version]) -> list[str] rendered.append(f"{integration}: {dep}") elif constraint.startswith("py3"): for dep in deps: - rendered.append(f"{constraint}-{integration}: {dep}") + rendered.append(f"{{{constraint}}}-{integration}: {dep}") else: restriction = SpecifierSet(constraint) for release in releases: diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 0ad9af8321..7f23d1fbc7 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -42,12 +42,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Boto3 - {py3.6,py3.7}-boto3-v{1.12} - {py3.7,py3.11,py3.12}-boto3-v{1.23} - {py3.11,py3.12}-boto3-v{1.34} - {py3.11,py3.12,py3.13}-boto3-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -164,12 +158,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Boto3 - boto3-v1.12: boto3~=1.12.0 - boto3-v1.23: boto3~=1.23.0 - boto3-v1.34: boto3~=1.34.0 - boto3-latest: boto3 - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 diff --git a/tox.ini b/tox.ini index 994ad22314..948887f1dd 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-05T06:53:57.545461+00:00 +# Last generated: 2025-09-05T07:14:50.663886+00:00 [tox] requires = @@ -42,12 +42,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Boto3 - {py3.6,py3.7}-boto3-v{1.12} - {py3.7,py3.11,py3.12}-boto3-v{1.23} - {py3.11,py3.12}-boto3-v{1.34} - {py3.11,py3.12,py3.13}-boto3-latest - # Chalice {py3.6,py3.9}-chalice-v{1.16} {py3.8,py3.12,py3.13}-chalice-latest @@ -152,6 +146,13 @@ envlist = {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 + # ~~~ Cloud ~~~ + {py3.6,py3.7}-boto3-v1.12.49 + {py3.6,py3.9,py3.10}-boto3-v1.20.54 + {py3.7,py3.11,py3.12}-boto3-v1.28.85 + {py3.9,py3.12,py3.13}-boto3-v1.40.24 + + # ~~~ DBs ~~~ {py3.7,py3.8,py3.9}-asyncpg-v0.23.0 {py3.7,py3.9,py3.10}-asyncpg-v0.25.0 @@ -371,12 +372,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Boto3 - boto3-v1.12: boto3~=1.12.0 - boto3-v1.23: boto3~=1.23.0 - boto3-v1.34: boto3~=1.34.0 - boto3-latest: boto3 - # Chalice chalice: pytest-chalice==0.0.5 chalice-v1.16: chalice~=1.16.0 @@ -539,6 +534,14 @@ deps = huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 + # ~~~ Cloud ~~~ + boto3-v1.12.49: boto3==1.12.49 + boto3-v1.20.54: boto3==1.20.54 + boto3-v1.28.85: boto3==1.28.85 + boto3-v1.40.24: boto3==1.40.24 + {py3.7,py3.8}-boto3: urllib3<2.0.0 + + # ~~~ DBs ~~~ asyncpg-v0.23.0: asyncpg==0.23.0 asyncpg-v0.25.0: asyncpg==0.25.0 @@ -604,7 +607,7 @@ deps = graphene: fastapi graphene: flask graphene: httpx - py3.6-graphene: aiocontextvars + {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.233.3: strawberry-graphql[fastapi,flask]==0.233.3 @@ -648,7 +651,7 @@ deps = celery-v5.5.3: celery==5.5.3 celery: newrelic<10.17.0 celery: redis - py3.7-celery: importlib-metadata<5.0 + {py3.7}-celery: importlib-metadata<5.0 dramatiq-v1.9.0: dramatiq==1.9.0 dramatiq-v1.12.3: dramatiq==1.12.3 @@ -717,7 +720,7 @@ deps = starlette-v0.16.0: httpx<0.28.0 starlette-v0.26.1: httpx<0.28.0 starlette-v0.36.3: httpx<0.28.0 - py3.6-starlette: aiocontextvars + {py3.6}-starlette: aiocontextvars fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.91.0: fastapi==0.91.0 @@ -731,7 +734,7 @@ deps = fastapi-v0.79.1: httpx<0.28.0 fastapi-v0.91.0: httpx<0.28.0 fastapi-v0.103.2: httpx<0.28.0 - py3.6-fastapi: aiocontextvars + {py3.6}-fastapi: aiocontextvars # ~~~ Web 2 ~~~ @@ -787,7 +790,7 @@ deps = tornado: pytest tornado-v6.0.4: pytest<8.2 tornado-v6.2: pytest<8.2 - py3.6-tornado: aiocontextvars + {py3.6}-tornado: aiocontextvars # ~~~ Misc ~~~ From dee6de1579ba37acb46af622e2892d862e9c70ef Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Fri, 5 Sep 2025 13:13:16 +0200 Subject: [PATCH 446/591] feat(agents): improve instrumentation of input messages (#4750) - Improve the instrumentation of input messages in the AI agents instrumentations. Before: Screenshot 2025-09-03 at 16 30 05 After: Screenshot 2025-09-03 at 16 30 08 Closes TET-1058 --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/langchain.py | 138 +++++++++++++----- sentry_sdk/integrations/langgraph.py | 12 +- sentry_sdk/integrations/openai.py | 43 ++++-- .../integrations/openai_agents/utils.py | 9 +- .../integrations/langchain/test_langchain.py | 73 +++++++++ tests/integrations/openai/test_openai.py | 16 +- .../openai_agents/test_openai_agents.py | 5 +- 7 files changed, 234 insertions(+), 62 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 7e04a740ed..a53115a2a9 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -51,7 +51,6 @@ "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, "tool_calls": SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, - "tools": SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, "top_k": SPANDATA.GEN_AI_REQUEST_TOP_K, "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, } @@ -203,8 +202,12 @@ def on_llm_start( if key in all_params and all_params[key] is not None: set_data_normalized(span, attribute, all_params[key], unpack=False) + _set_tools_on_span(span, all_params.get("tools")) + if should_send_default_pii() and self.include_prompts: - set_data_normalized(span, SPANDATA.GEN_AI_REQUEST_MESSAGES, prompts) + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, prompts, unpack=False + ) def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): # type: (SentryLangchainCallback, Dict[str, Any], List[List[BaseMessage]], UUID, Any) -> Any @@ -246,14 +249,20 @@ def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): if key in all_params and all_params[key] is not None: set_data_normalized(span, attribute, all_params[key], unpack=False) + _set_tools_on_span(span, all_params.get("tools")) + if should_send_default_pii() and self.include_prompts: + normalized_messages = [] + for list_ in messages: + for message in list_: + normalized_messages.append( + self._normalize_langchain_message(message) + ) set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, - [ - [self._normalize_langchain_message(x) for x in list_] - for list_ in messages - ], + normalized_messages, + unpack=False, ) def on_chat_model_end(self, response, *, run_id, **kwargs): @@ -351,9 +360,7 @@ def on_agent_finish(self, finish, *, run_id, **kwargs): if should_send_default_pii() and self.include_prompts: set_data_normalized( - span, - SPANDATA.GEN_AI_RESPONSE_TEXT, - finish.return_values.items(), + span, SPANDATA.GEN_AI_RESPONSE_TEXT, finish.return_values.items() ) self._exit_span(span_data, run_id) @@ -473,13 +480,11 @@ def _get_token_usage(obj): if usage is not None: return usage - # check for usage in the object itself for name in possible_names: usage = _get_value(obj, name) if usage is not None: return usage - # no usage found anywhere return None @@ -531,6 +536,87 @@ def _get_request_data(obj, args, kwargs): return (agent_name, tools) +def _simplify_langchain_tools(tools): + # type: (Any) -> Optional[List[Any]] + """Parse and simplify tools into a cleaner format.""" + if not tools: + return None + + if not isinstance(tools, (list, tuple)): + return None + + simplified_tools = [] + for tool in tools: + try: + if isinstance(tool, dict): + + if "function" in tool and isinstance(tool["function"], dict): + func = tool["function"] + simplified_tool = { + "name": func.get("name"), + "description": func.get("description"), + } + if simplified_tool["name"]: + simplified_tools.append(simplified_tool) + elif "name" in tool: + simplified_tool = { + "name": tool.get("name"), + "description": tool.get("description"), + } + simplified_tools.append(simplified_tool) + else: + name = ( + tool.get("name") + or tool.get("tool_name") + or tool.get("function_name") + ) + if name: + simplified_tools.append( + { + "name": name, + "description": tool.get("description") + or tool.get("desc"), + } + ) + elif hasattr(tool, "name"): + simplified_tool = { + "name": getattr(tool, "name", None), + "description": getattr(tool, "description", None) + or getattr(tool, "desc", None), + } + if simplified_tool["name"]: + simplified_tools.append(simplified_tool) + elif hasattr(tool, "__name__"): + simplified_tools.append( + { + "name": tool.__name__, + "description": getattr(tool, "__doc__", None), + } + ) + else: + tool_str = str(tool) + if tool_str and tool_str != "": + simplified_tools.append({"name": tool_str, "description": None}) + except Exception: + continue + + return simplified_tools if simplified_tools else None + + +def _set_tools_on_span(span, tools): + # type: (Span, Any) -> None + """Set available tools data on a span if tools are provided.""" + if tools is not None: + simplified_tools = _simplify_langchain_tools(tools) + if simplified_tools: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, + simplified_tools, + unpack=False, + ) + + def _wrap_configure(f): # type: (Callable[..., Any]) -> Callable[..., Any] @@ -601,7 +687,7 @@ def new_configure( ] elif isinstance(local_callbacks, BaseCallbackHandler): local_callbacks = [local_callbacks, sentry_handler] - else: # local_callbacks is a list + else: local_callbacks = [*local_callbacks, sentry_handler] return f( @@ -638,10 +724,7 @@ def new_invoke(self, *args, **kwargs): span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, False) - if tools: - set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, tools, unpack=False - ) + _set_tools_on_span(span, tools) # Run the agent result = f(self, *args, **kwargs) @@ -653,11 +736,7 @@ def new_invoke(self, *args, **kwargs): and integration.include_prompts ): set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - [ - input, - ], + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, [input], unpack=False ) output = result.get("output") @@ -666,7 +745,7 @@ def new_invoke(self, *args, **kwargs): and should_send_default_pii() and integration.include_prompts ): - span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, output) return result @@ -698,10 +777,7 @@ def new_stream(self, *args, **kwargs): span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) - if tools: - set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, tools, unpack=False - ) + _set_tools_on_span(span, tools) input = args[0].get("input") if len(args) >= 1 else None if ( @@ -710,11 +786,7 @@ def new_stream(self, *args, **kwargs): and integration.include_prompts ): set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - [ - input, - ], + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, [input], unpack=False ) # Run the agent @@ -737,7 +809,7 @@ def new_iterator(): and should_send_default_pii() and integration.include_prompts ): - span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, output) span.__exit__(None, None, None) @@ -756,7 +828,7 @@ async def new_iterator_async(): and should_send_default_pii() and integration.include_prompts ): - span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, output) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, output) span.__exit__(None, None, None) diff --git a/sentry_sdk/integrations/langgraph.py b/sentry_sdk/integrations/langgraph.py index 4b241fe895..df3941bb13 100644 --- a/sentry_sdk/integrations/langgraph.py +++ b/sentry_sdk/integrations/langgraph.py @@ -183,7 +183,8 @@ def new_invoke(self, *args, **kwargs): set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, - safe_serialize(input_messages), + input_messages, + unpack=False, ) result = f(self, *args, **kwargs) @@ -232,7 +233,8 @@ async def new_ainvoke(self, *args, **kwargs): set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, - safe_serialize(input_messages), + input_messages, + unpack=False, ) result = await f(self, *args, **kwargs) @@ -305,11 +307,9 @@ def _set_response_attributes(span, input_messages, result, integration): if llm_response_text: set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, llm_response_text) elif new_messages: - set_data_normalized( - span, SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(new_messages) - ) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, new_messages) else: - set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(result)) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, result) tool_calls = _extract_tool_calls(new_messages) if tool_calls: diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 6ea545322c..467116c8f4 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -179,7 +179,9 @@ def _set_input_data(span, kwargs, operation, integration): and should_send_default_pii() and integration.include_prompts ): - set_data_normalized(span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages) + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + ) # Input attributes: Common set_data_normalized(span, SPANDATA.GEN_AI_SYSTEM, "openai") @@ -227,25 +229,46 @@ def _set_output_data(span, response, kwargs, integration, finish_span=True): if should_send_default_pii() and integration.include_prompts: response_text = [choice.message.dict() for choice in response.choices] if len(response_text) > 0: - set_data_normalized( - span, - SPANDATA.GEN_AI_RESPONSE_TEXT, - safe_serialize(response_text), - ) + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, response_text) + _calculate_token_usage(messages, response, span, None, integration.count_tokens) + if finish_span: span.__exit__(None, None, None) elif hasattr(response, "output"): if should_send_default_pii() and integration.include_prompts: - response_text = [item.to_dict() for item in response.output] - if len(response_text) > 0: + output_messages = { + "response": [], + "tool": [], + } # type: (dict[str, list[Any]]) + + for output in response.output: + if output.type == "function_call": + output_messages["tool"].append(output.dict()) + elif output.type == "message": + for output_message in output.content: + try: + output_messages["response"].append(output_message.text) + except AttributeError: + # Unknown output message type, just return the json + output_messages["response"].append(output_message.dict()) + + if len(output_messages["tool"]) > 0: set_data_normalized( span, - SPANDATA.GEN_AI_RESPONSE_TEXT, - safe_serialize(response_text), + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + output_messages["tool"], + unpack=False, + ) + + if len(output_messages["response"]) > 0: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, output_messages["response"] ) + _calculate_token_usage(messages, response, span, None, integration.count_tokens) + if finish_span: span.__exit__(None, None, None) diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index 1525346726..44b260d4bc 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -1,4 +1,5 @@ import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii @@ -127,7 +128,9 @@ def _set_input_data(span, get_response_kwargs): if len(messages) > 0: request_messages.append({"role": role, "content": messages}) - span.set_data(SPANDATA.GEN_AI_REQUEST_MESSAGES, safe_serialize(request_messages)) + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, request_messages, unpack=False + ) def _set_output_data(span, result): @@ -157,6 +160,6 @@ def _set_output_data(span, result): ) if len(output_messages["response"]) > 0: - span.set_data( - SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(output_messages["response"]) + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, output_messages["response"] ) diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 9a06ac05d4..99dc5f4e37 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -589,3 +589,76 @@ def test_langchain_callback_list_existing_callback(sentry_init): [handler] = passed_callbacks assert handler is sentry_callback + + +def test_tools_integration_in_spans(sentry_init, capture_events): + """Test that tools are properly set on spans in actual LangChain integration.""" + global llm_type + llm_type = "openai-chat" + + sentry_init( + integrations=[LangchainIntegration(include_prompts=False)], + traces_sample_rate=1.0, + ) + events = capture_events() + + prompt = ChatPromptTemplate.from_messages( + [ + ("system", "You are a helpful assistant"), + ("user", "{input}"), + MessagesPlaceholder(variable_name="agent_scratchpad"), + ] + ) + + global stream_result_mock + stream_result_mock = Mock( + side_effect=[ + [ + ChatGenerationChunk( + type="ChatGenerationChunk", + message=AIMessageChunk(content="Simple response"), + ), + ] + ] + ) + + llm = MockOpenAI( + model_name="gpt-3.5-turbo", + temperature=0, + openai_api_key="badkey", + ) + agent = create_openai_tools_agent(llm, [get_word_length], prompt) + agent_executor = AgentExecutor(agent=agent, tools=[get_word_length], verbose=True) + + with start_transaction(): + list(agent_executor.stream({"input": "Hello"})) + + # Check that events were captured and contain tools data + if events: + tx = events[0] + spans = tx.get("spans", []) + + # Look for spans that should have tools data + tools_found = False + for span in spans: + span_data = span.get("data", {}) + if SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS in span_data: + tools_found = True + tools_data = span_data[SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS] + # Verify tools are in the expected format + assert isinstance(tools_data, (str, list)) # Could be serialized + if isinstance(tools_data, str): + # If serialized as string, should contain tool name + assert "get_word_length" in tools_data + else: + # If still a list, verify structure + assert len(tools_data) >= 1 + names = [ + tool.get("name") + for tool in tools_data + if isinstance(tool, dict) + ] + assert "get_word_length" in names + + # Ensure we found at least one span with tools data + assert tools_found, "No spans found with tools data" diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index a3c7bdd9d9..18968fb36a 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1036,7 +1036,7 @@ def test_ai_client_span_responses_api(sentry_init, capture_events): assert spans[0]["origin"] == "auto.ai.openai" assert spans[0]["data"] == { "gen_ai.operation.name": "responses", - "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.messages": '["How do I check if a Python object is an instance of a class?"]', "gen_ai.request.model": "gpt-4o", "gen_ai.system": "openai", "gen_ai.response.model": "response-model-id", @@ -1045,7 +1045,7 @@ def test_ai_client_span_responses_api(sentry_init, capture_events): "gen_ai.usage.output_tokens": 10, "gen_ai.usage.output_tokens.reasoning": 8, "gen_ai.usage.total_tokens": 30, - "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "gen_ai.response.text": "the model response", "thread.id": mock.ANY, "thread.name": mock.ANY, } @@ -1116,7 +1116,7 @@ async def test_ai_client_span_responses_async_api(sentry_init, capture_events): assert spans[0]["origin"] == "auto.ai.openai" assert spans[0]["data"] == { "gen_ai.operation.name": "responses", - "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.messages": '["How do I check if a Python object is an instance of a class?"]', "gen_ai.request.model": "gpt-4o", "gen_ai.response.model": "response-model-id", "gen_ai.system": "openai", @@ -1125,7 +1125,7 @@ async def test_ai_client_span_responses_async_api(sentry_init, capture_events): "gen_ai.usage.output_tokens": 10, "gen_ai.usage.output_tokens.reasoning": 8, "gen_ai.usage.total_tokens": 30, - "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "gen_ai.response.text": "the model response", "thread.id": mock.ANY, "thread.name": mock.ANY, } @@ -1162,7 +1162,7 @@ async def test_ai_client_span_streaming_responses_async_api( assert spans[0]["origin"] == "auto.ai.openai" assert spans[0]["data"] == { "gen_ai.operation.name": "responses", - "gen_ai.request.messages": "How do I check if a Python object is an instance of a class?", + "gen_ai.request.messages": '["How do I check if a Python object is an instance of a class?"]', "gen_ai.request.model": "gpt-4o", "gen_ai.response.model": "response-model-id", "gen_ai.response.streaming": True, @@ -1172,7 +1172,7 @@ async def test_ai_client_span_streaming_responses_async_api( "gen_ai.usage.output_tokens": 10, "gen_ai.usage.output_tokens.reasoning": 8, "gen_ai.usage.total_tokens": 30, - "gen_ai.response.text": '[{"id": "message-id", "content": [{"annotations": [], "text": "the model response", "type": "output_text"}], "role": "assistant", "status": "completed", "type": "message"}]', + "gen_ai.response.text": "the model response", "thread.id": mock.ANY, "thread.name": mock.ANY, } @@ -1332,7 +1332,7 @@ def test_streaming_responses_api( assert span["op"] == "gen_ai.responses" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == "hello" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '["hello"]' assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "hello world" else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] @@ -1387,7 +1387,7 @@ async def test_streaming_responses_api_async( assert span["op"] == "gen_ai.responses" if send_default_pii and include_prompts: - assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == "hello" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '["hello"]' assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "hello world" else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index a3075e6415..fab8d9e13f 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -582,8 +582,9 @@ def simple_test_tool(message: str) -> str: assert ai_client_span2["data"]["gen_ai.request.model"] == "gpt-4" assert ai_client_span2["data"]["gen_ai.request.temperature"] == 0.7 assert ai_client_span2["data"]["gen_ai.request.top_p"] == 1.0 - assert ai_client_span2["data"]["gen_ai.response.text"] == safe_serialize( - ["Task completed using the tool"] + assert ( + ai_client_span2["data"]["gen_ai.response.text"] + == "Task completed using the tool" ) assert ai_client_span2["data"]["gen_ai.system"] == "openai" assert ai_client_span2["data"]["gen_ai.usage.input_tokens.cached"] == 0 From f78552480e894b7a5a152530c589d9677f81bc14 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Fri, 5 Sep 2025 11:25:19 +0000 Subject: [PATCH 447/591] release: 2.37.0 --- CHANGELOG.md | 13 +++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 16 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2c10ef8b7f..29dfdfff07 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,18 @@ # Changelog +## 2.37.0 + +### Various fixes & improvements + +- feat(agents): improve instrumentation of input messages (#4750) by @shellmayr +- tests: Move boto3 tests under toxgen (#4761) by @sentrivana +- ci: Fix celery (#4765) by @sentrivana +- Format span attributes in AI integrations (#4762) by @antonpirker +- tests: Move asyncpg under toxgen (#4757) by @sentrivana +- feat: Add LangGraph integration (#4727) by @shellmayr +- tests: Move beam under toxgen (#4759) by @sentrivana +- tests: Remove openai pin and update tox (#4748) by @sentrivana + ## 2.36.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 835c20b112..935f45f6af 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.36.0" +release = "2.37.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 5480ef5dce..68a44fe88f 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1330,4 +1330,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.36.0" +VERSION = "2.37.0" diff --git a/setup.py b/setup.py index ca6e7ec534..8c4ea96ab9 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.36.0", + version="2.37.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 7d7c8ea0a50252151b05eeaeaa9c8f87cbe1b0c6 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 5 Sep 2025 13:28:41 +0200 Subject: [PATCH 448/591] tests: Move chalice under toxgen (#4766) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit chalice was the last test suite in the Cloud group that was still hardcoded, so moving it under toxgen also gets rid of the whole `latest` group 🎉 Ref https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-cloud.yml | 79 ------------------- scripts/populate_tox/config.py | 6 ++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 9 --- tests/integrations/chalice/test_chalice.py | 4 +- tox.ini | 22 +++--- 6 files changed, 20 insertions(+), 101 deletions(-) diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 8688a1d48e..62e70d759d 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -22,85 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-cloud-latest: - name: Cloud (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.8","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - services: - docker: - image: docker:dind # Required for Docker network management - options: --privileged # Required for Docker-in-Docker operations - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test aws_lambda latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-aws_lambda-latest" - - name: Test boto3 latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-boto3-latest" - - name: Test chalice latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-chalice-latest" - - name: Test cloud_resource_context latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-cloud_resource_context-latest" - - name: Test gcp latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-gcp-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-cloud-pinned: name: Cloud (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 5aba82b11b..b05c4297f1 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -66,6 +66,12 @@ "py3.7": ["importlib-metadata<5.0"], }, }, + "chalice": { + "package": "chalice", + "deps": { + "*": ["pytest-chalice"], + }, + }, "clickhouse_driver": { "package": "clickhouse-driver", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index b8cc988fda..9aed9fa718 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,7 +67,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "chalice", "gcp", "httpx", "pure_eval", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 7f23d1fbc7..c243b5752e 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -42,10 +42,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Chalice - {py3.6,py3.9}-chalice-v{1.16} - {py3.8,py3.12,py3.13}-chalice-latest - # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context @@ -158,11 +154,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Chalice - chalice: pytest-chalice==0.0.5 - chalice-v1.16: chalice~=1.16.0 - chalice-latest: chalice - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 diff --git a/tests/integrations/chalice/test_chalice.py b/tests/integrations/chalice/test_chalice.py index fbd4be4e59..ec8106eb5f 100644 --- a/tests/integrations/chalice/test_chalice.py +++ b/tests/integrations/chalice/test_chalice.py @@ -110,7 +110,7 @@ def every_hour(event): @pytest.mark.skipif( - parse_version(CHALICE_VERSION) >= (1, 28), + parse_version(CHALICE_VERSION) >= (1, 26, 0), reason="different behavior based on chalice version", ) def test_bad_request_old(client: RequestHandler) -> None: @@ -124,7 +124,7 @@ def test_bad_request_old(client: RequestHandler) -> None: @pytest.mark.skipif( - parse_version(CHALICE_VERSION) < (1, 28), + parse_version(CHALICE_VERSION) < (1, 26, 0), reason="different behavior based on chalice version", ) def test_bad_request(client: RequestHandler) -> None: diff --git a/tox.ini b/tox.ini index 948887f1dd..335007664a 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-05T07:14:50.663886+00:00 +# Last generated: 2025-09-05T07:52:27.350774+00:00 [tox] requires = @@ -42,10 +42,6 @@ envlist = # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda - # Chalice - {py3.6,py3.9}-chalice-v{1.16} - {py3.8,py3.12,py3.13}-chalice-latest - # Cloud Resource Context {py3.6,py3.12,py3.13}-cloud_resource_context @@ -152,6 +148,11 @@ envlist = {py3.7,py3.11,py3.12}-boto3-v1.28.85 {py3.9,py3.12,py3.13}-boto3-v1.40.24 + {py3.6,py3.7,py3.8}-chalice-v1.16.0 + {py3.6,py3.7,py3.8}-chalice-v1.21.9 + {py3.6,py3.8,py3.9}-chalice-v1.26.6 + {py3.9,py3.12,py3.13}-chalice-v1.32.0 + # ~~~ DBs ~~~ {py3.7,py3.8,py3.9}-asyncpg-v0.23.0 @@ -372,11 +373,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # Chalice - chalice: pytest-chalice==0.0.5 - chalice-v1.16: chalice~=1.16.0 - chalice-latest: chalice - # HTTPX httpx-v0.16: pytest-httpx==0.10.0 httpx-v0.18: pytest-httpx==0.12.0 @@ -541,6 +537,12 @@ deps = boto3-v1.40.24: boto3==1.40.24 {py3.7,py3.8}-boto3: urllib3<2.0.0 + chalice-v1.16.0: chalice==1.16.0 + chalice-v1.21.9: chalice==1.21.9 + chalice-v1.26.6: chalice==1.26.6 + chalice-v1.32.0: chalice==1.32.0 + chalice: pytest-chalice + # ~~~ DBs ~~~ asyncpg-v0.23.0: asyncpg==0.23.0 From 75ef769d494c45e6f8b133da22fe75dcf9da713e Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 5 Sep 2025 13:31:56 +0200 Subject: [PATCH 449/591] Updated changelog --- CHANGELOG.md | 21 ++++++++++++--------- 1 file changed, 12 insertions(+), 9 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 29dfdfff07..52478dd4dd 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,16 +2,19 @@ ## 2.37.0 -### Various fixes & improvements +- **New Integration (BETA):** Add support for `langgraph` (#4727) by @shellmayr + + We can now instrument AI agents that are created with [LangGraph](https://www.langchain.com/langgraph) out of the box. + + For more information see the [LangGraph integrations documentation](https://docs.sentry.io/platforms/python/integrations/langgraph/). -- feat(agents): improve instrumentation of input messages (#4750) by @shellmayr -- tests: Move boto3 tests under toxgen (#4761) by @sentrivana -- ci: Fix celery (#4765) by @sentrivana -- Format span attributes in AI integrations (#4762) by @antonpirker -- tests: Move asyncpg under toxgen (#4757) by @sentrivana -- feat: Add LangGraph integration (#4727) by @shellmayr -- tests: Move beam under toxgen (#4759) by @sentrivana -- tests: Remove openai pin and update tox (#4748) by @sentrivana +- AI Agents: Improve rendering of input and output messages in AI agents integrations. (#4750) by @shellmayr +- AI Agents: Format span attributes in AI integrations (#4762) by @antonpirker +- CI: Fix celery (#4765) by @sentrivana +- Tests: Move asyncpg under toxgen (#4757) by @sentrivana +- Tests: Move beam under toxgen (#4759) by @sentrivana +- Tests: Move boto3 tests under toxgen (#4761) by @sentrivana +- Tests: Remove openai pin and update tox (#4748) by @sentrivana ## 2.36.0 From bdf3e6d51dd5007e37ba133a24b1e986502a6bd1 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 8 Sep 2025 11:22:35 +0200 Subject: [PATCH 450/591] tests: Update tox.ini (#4777) --- tox.ini | 30 +++++++++++++++--------------- 1 file changed, 15 insertions(+), 15 deletions(-) diff --git a/tox.ini b/tox.ini index 335007664a..2c1c2382f5 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-05T07:52:27.350774+00:00 +# Last generated: 2025-09-08T07:44:56.804943+00:00 [tox] requires = @@ -128,8 +128,8 @@ envlist = {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.106.1 - {py3.9,py3.12,py3.13}-langgraph-v0.6.6 - {py3.10,py3.12,py3.13}-langgraph-v1.0.0a2 + {py3.9,py3.12,py3.13}-langgraph-v0.6.7 + {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 @@ -146,7 +146,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.24 + {py3.9,py3.12,py3.13}-boto3-v1.40.25 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -205,7 +205,7 @@ envlist = {py3.6,py3.9,py3.10}-gql-v3.4.1 {py3.7,py3.11,py3.12}-gql-v3.5.3 {py3.9,py3.12,py3.13}-gql-v4.0.0 - {py3.9,py3.12,py3.13}-gql-v4.1.0b0 + {py3.9,py3.12,py3.13}-gql-v4.2.0b0 {py3.6,py3.9,py3.10}-graphene-v3.3 {py3.8,py3.12,py3.13}-graphene-v3.4.3 @@ -213,7 +213,7 @@ envlist = {py3.8,py3.10,py3.11}-strawberry-v0.209.8 {py3.8,py3.11,py3.12}-strawberry-v0.233.3 {py3.9,py3.12,py3.13}-strawberry-v0.257.0 - {py3.9,py3.12,py3.13}-strawberry-v0.281.0 + {py3.9,py3.12,py3.13}-strawberry-v0.282.0 # ~~~ Network ~~~ @@ -251,7 +251,7 @@ envlist = {py3.8,py3.9}-spark-v3.0.3 {py3.8,py3.10,py3.11}-spark-v3.5.6 - {py3.9,py3.12,py3.13}-spark-v4.0.0 + {py3.9,py3.12,py3.13}-spark-v4.0.1 # ~~~ Web 1 ~~~ @@ -325,7 +325,7 @@ envlist = {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 - {py3.7,py3.12,py3.13}-typer-v0.17.3 + {py3.7,py3.12,py3.13}-typer-v0.17.4 @@ -515,8 +515,8 @@ deps = openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 - langgraph-v0.6.6: langgraph==0.6.6 - langgraph-v1.0.0a2: langgraph==1.0.0a2 + langgraph-v0.6.7: langgraph==0.6.7 + langgraph-v1.0.0a3: langgraph==1.0.0a3 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 @@ -534,7 +534,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.24: boto3==1.40.24 + boto3-v1.40.25: boto3==1.40.25 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -601,7 +601,7 @@ deps = gql-v3.4.1: gql[all]==3.4.1 gql-v3.5.3: gql[all]==3.5.3 gql-v4.0.0: gql[all]==4.0.0 - gql-v4.1.0b0: gql[all]==4.1.0b0 + gql-v4.2.0b0: gql[all]==4.2.0b0 graphene-v3.3: graphene==3.3 graphene-v3.4.3: graphene==3.4.3 @@ -614,7 +614,7 @@ deps = strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 strawberry-v0.233.3: strawberry-graphql[fastapi,flask]==0.233.3 strawberry-v0.257.0: strawberry-graphql[fastapi,flask]==0.257.0 - strawberry-v0.281.0: strawberry-graphql[fastapi,flask]==0.281.0 + strawberry-v0.282.0: strawberry-graphql[fastapi,flask]==0.282.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 strawberry-v0.233.3: pydantic<2.11 @@ -667,7 +667,7 @@ deps = spark-v3.0.3: pyspark==3.0.3 spark-v3.5.6: pyspark==3.5.6 - spark-v4.0.0: pyspark==4.0.0 + spark-v4.0.1: pyspark==4.0.1 # ~~~ Web 1 ~~~ @@ -810,7 +810,7 @@ deps = typer-v0.15.4: typer==0.15.4 typer-v0.16.1: typer==0.16.1 - typer-v0.17.3: typer==0.17.3 + typer-v0.17.4: typer==0.17.4 From 0eede69d8eded6785c62461d81a3b4585421d9eb Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 8 Sep 2025 13:46:08 +0200 Subject: [PATCH 451/591] tests: Move quart under toxgen (#4775) --- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/populate_tox/config.py | 14 ++++++ scripts/populate_tox/populate_tox.py | 4 +- scripts/populate_tox/tox.jinja | 19 -------- tox.ini | 48 +++++++++++-------- 5 files changed, 45 insertions(+), 42 deletions(-) diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index e79a54ef67..22200f8ae1 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.12","3.13"] + python-version: ["3.9","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index b05c4297f1..679ffddf2c 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -232,6 +232,20 @@ "*": ["werkzeug<2.1.0"], }, }, + "quart": { + "package": "quart", + "deps": { + "*": ["quart-auth", "pytest-asyncio", "Werkzeug"], + ">=0.19": ["quart-flask-patch"], + "<0.19": [ + "blinker<1.6", + "jinja2<3.1.0", + "Werkzeug<2.3.0", + "hypercorn<0.15.0", + ], + "py3.8": ["taskgroup==0.0.0a4"], + }, + }, "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 9aed9fa718..e08c2d4b95 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -57,7 +57,8 @@ # pypi package to install in different versions). # # Test suites that will have to remain hardcoded since they don't fit the - # toxgen usecase + # toxgen usecase (there is no one package that should be tested in different + # versions) "asgi", "aws_lambda", "cloud_resource_context", @@ -70,7 +71,6 @@ "gcp", "httpx", "pure_eval", - "quart", "ray", "redis", "requests", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index c243b5752e..ef2e89c88c 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -64,11 +64,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # Quart - {py3.7,py3.11}-quart-v{0.16} - {py3.8,py3.11,py3.12}-quart-v{0.19} - {py3.8,py3.12,py3.13}-quart-latest - # Ray {py3.10,py3.11}-ray-v{2.34} {py3.10,py3.11}-ray-latest @@ -184,20 +179,6 @@ deps = # pure_eval pure_eval: pure_eval - # Quart - quart: quart-auth - quart: pytest-asyncio - quart-{v0.19,latest}: quart-flask-patch - quart-v0.16: blinker<1.6 - quart-v0.16: jinja2<3.1.0 - quart-v0.16: Werkzeug<2.1.0 - quart-v0.16: hypercorn<0.15.0 - quart-v0.16: quart~=0.16.0 - quart-v0.19: Werkzeug>=3.0.0 - quart-v0.19: quart~=0.19.0 - {py3.8}-quart: taskgroup==0.0.0a4 - quart-latest: quart - # Ray ray-v2.34: ray~=2.34.0 ray-latest: ray diff --git a/tox.ini b/tox.ini index 2c1c2382f5..ff2403f515 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-08T07:44:56.804943+00:00 +# Last generated: 2025-09-08T11:35:09.849536+00:00 [tox] requires = @@ -64,11 +64,6 @@ envlist = # pure_eval {py3.6,py3.12,py3.13}-pure_eval - # Quart - {py3.7,py3.11}-quart-v{0.16} - {py3.8,py3.11,py3.12}-quart-v{0.19} - {py3.8,py3.12,py3.13}-quart-latest - # Ray {py3.10,py3.11}-ray-v{2.34} {py3.10,py3.11}-ray-latest @@ -302,6 +297,11 @@ envlist = {py3.6,py3.8,py3.9}-pyramid-v1.10.8 {py3.6,py3.10,py3.11}-pyramid-v2.0.2 + {py3.7,py3.9,py3.10}-quart-v0.16.3 + {py3.7,py3.9,py3.10}-quart-v0.17.0 + {py3.7,py3.10,py3.11}-quart-v0.18.4 + {py3.9,py3.12,py3.13}-quart-v0.20.0 + {py3.8,py3.10,py3.11}-starlite-v1.48.1 {py3.8,py3.10,py3.11}-starlite-v1.49.0 {py3.8,py3.10,py3.11}-starlite-v1.50.2 @@ -403,20 +403,6 @@ deps = # pure_eval pure_eval: pure_eval - # Quart - quart: quart-auth - quart: pytest-asyncio - quart-{v0.19,latest}: quart-flask-patch - quart-v0.16: blinker<1.6 - quart-v0.16: jinja2<3.1.0 - quart-v0.16: Werkzeug<2.1.0 - quart-v0.16: hypercorn<0.15.0 - quart-v0.16: quart~=0.16.0 - quart-v0.19: Werkzeug>=3.0.0 - quart-v0.19: quart~=0.19.0 - {py3.8}-quart: taskgroup==0.0.0a4 - quart-latest: quart - # Ray ray-v2.34: ray~=2.34.0 ray-latest: ray @@ -774,6 +760,28 @@ deps = pyramid-v2.0.2: pyramid==2.0.2 pyramid: werkzeug<2.1.0 + quart-v0.16.3: quart==0.16.3 + quart-v0.17.0: quart==0.17.0 + quart-v0.18.4: quart==0.18.4 + quart-v0.20.0: quart==0.20.0 + quart: quart-auth + quart: pytest-asyncio + quart: Werkzeug + quart-v0.20.0: quart-flask-patch + quart-v0.16.3: blinker<1.6 + quart-v0.16.3: jinja2<3.1.0 + quart-v0.16.3: Werkzeug<2.3.0 + quart-v0.16.3: hypercorn<0.15.0 + quart-v0.17.0: blinker<1.6 + quart-v0.17.0: jinja2<3.1.0 + quart-v0.17.0: Werkzeug<2.3.0 + quart-v0.17.0: hypercorn<0.15.0 + quart-v0.18.4: blinker<1.6 + quart-v0.18.4: jinja2<3.1.0 + quart-v0.18.4: Werkzeug<2.3.0 + quart-v0.18.4: hypercorn<0.15.0 + {py3.8}-quart: taskgroup==0.0.0a4 + starlite-v1.48.1: starlite==1.48.1 starlite-v1.49.0: starlite==1.49.0 starlite-v1.50.2: starlite==1.50.2 From 20f0f848fb7b2520c0b6082597adcce7fbdd8cee Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Tue, 9 Sep 2025 10:14:44 +0200 Subject: [PATCH 452/591] fix(langchain): make new langchain integration work with just langchain-core (#4783) - Catch ImportError for `langchain.agents` if langchain is not present and set AgentExecutor to None so the rest of the logic still keeps working - Add test for recording new OTEL-compliant data for `langchain-core` - Verified AI Agents functionality (span generation, token accounting, etc) with `langchain-core` and `langchain-openai` Closes TET-1126 Closes PY-1833 Closes https://github.com/getsentry/sentry-python/issues/4776 --- sentry_sdk/integrations/langchain.py | 6 +- .../integrations/langchain/test_langchain.py | 78 ++++++++++++++++++- 2 files changed, 82 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index a53115a2a9..e14dd619fe 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -29,7 +29,6 @@ try: - from langchain.agents import AgentExecutor from langchain_core.agents import AgentFinish from langchain_core.callbacks import ( BaseCallbackHandler, @@ -44,6 +43,11 @@ raise DidNotEnable("langchain not installed") +try: + from langchain.agents import AgentExecutor +except ImportError: + AgentExecutor = None + DATA_FIELDS = { "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, "function_call": SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 99dc5f4e37..b6b432c523 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -1,6 +1,6 @@ from typing import List, Optional, Any, Iterator from unittest import mock -from unittest.mock import Mock +from unittest.mock import Mock, patch import pytest @@ -662,3 +662,79 @@ def test_tools_integration_in_spans(sentry_init, capture_events): # Ensure we found at least one span with tools data assert tools_found, "No spans found with tools data" + + +def test_langchain_integration_with_langchain_core_only(sentry_init, capture_events): + """Test that the langchain integration works when langchain.agents.AgentExecutor + is not available or langchain is not installed, but langchain-core is. + """ + + from langchain_core.outputs import LLMResult, Generation + + with patch("sentry_sdk.integrations.langchain.AgentExecutor", None): + from sentry_sdk.integrations.langchain import ( + LangchainIntegration, + SentryLangchainCallback, + ) + + sentry_init( + integrations=[LangchainIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + try: + LangchainIntegration.setup_once() + except Exception as e: + pytest.fail(f"setup_once() failed when AgentExecutor is None: {e}") + + callback = SentryLangchainCallback(max_span_map_size=100, include_prompts=True) + + run_id = "12345678-1234-1234-1234-123456789012" + serialized = {"_type": "openai-chat", "model_name": "gpt-3.5-turbo"} + prompts = ["What is the capital of France?"] + + with start_transaction(): + callback.on_llm_start( + serialized=serialized, + prompts=prompts, + run_id=run_id, + invocation_params={ + "temperature": 0.7, + "max_tokens": 100, + "model": "gpt-3.5-turbo", + }, + ) + + response = LLMResult( + generations=[[Generation(text="The capital of France is Paris.")]], + llm_output={ + "token_usage": { + "total_tokens": 25, + "prompt_tokens": 10, + "completion_tokens": 15, + } + }, + ) + callback.on_llm_end(response=response, run_id=run_id) + + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + llm_spans = [ + span for span in tx.get("spans", []) if span.get("op") == "gen_ai.pipeline" + ] + assert len(llm_spans) > 0 + + llm_span = llm_spans[0] + assert llm_span["description"] == "Langchain LLM call" + assert llm_span["data"]["gen_ai.request.model"] == "gpt-3.5-turbo" + assert ( + llm_span["data"]["gen_ai.response.text"] + == "The capital of France is Paris." + ) + assert llm_span["data"]["gen_ai.usage.total_tokens"] == 25 + assert llm_span["data"]["gen_ai.usage.input_tokens"] == 10 + assert llm_span["data"]["gen_ai.usage.output_tokens"] == 15 From 2d9c428d43e03ffb1789cfb9948684de8df7a551 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Tue, 9 Sep 2025 13:24:33 +0000 Subject: [PATCH 453/591] release: 2.37.1 --- CHANGELOG.md | 9 +++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 12 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 52478dd4dd..7491ea63d2 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,14 @@ # Changelog +## 2.37.1 + +### Various fixes & improvements + +- fix(langchain): make new langchain integration work with just langchain-core (#4783) by @shellmayr +- tests: Move quart under toxgen (#4775) by @sentrivana +- tests: Update tox.ini (#4777) by @sentrivana +- tests: Move chalice under toxgen (#4766) by @sentrivana + ## 2.37.0 - **New Integration (BETA):** Add support for `langgraph` (#4727) by @shellmayr diff --git a/docs/conf.py b/docs/conf.py index 935f45f6af..28a49b7fa7 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.37.0" +release = "2.37.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 68a44fe88f..4f015643d4 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1330,4 +1330,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.37.0" +VERSION = "2.37.1" diff --git a/setup.py b/setup.py index 8c4ea96ab9..1b4d0063e4 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.37.0", + version="2.37.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From cd23041494a7cf98350c983d69e528a772e5cd6d Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Tue, 9 Sep 2025 15:25:57 +0200 Subject: [PATCH 454/591] updated changelog --- CHANGELOG.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 7491ea63d2..28c4882414 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,10 +4,10 @@ ### Various fixes & improvements -- fix(langchain): make new langchain integration work with just langchain-core (#4783) by @shellmayr -- tests: Move quart under toxgen (#4775) by @sentrivana -- tests: Update tox.ini (#4777) by @sentrivana -- tests: Move chalice under toxgen (#4766) by @sentrivana +- Fix(langchain): Make Langchain integration work with just langchain-core (#4783) by @shellmayr +- Tests: Move quart under toxgen (#4775) by @sentrivana +- Tests: Update tox.ini (#4777) by @sentrivana +- Tests: Move chalice under toxgen (#4766) by @sentrivana ## 2.37.0 From 6463f73e48abd3fc30d26ff07ae60fb65dc38a2a Mon Sep 17 00:00:00 2001 From: Tony Xiao Date: Tue, 9 Sep 2025 11:35:46 -0400 Subject: [PATCH 455/591] fix(profiling): Re-init continuous profiler (#4772) Re-initializing the continuous profiler should use new settings. --- sentry_sdk/profiler/continuous_profiler.py | 16 +++++++++++++--- tests/profiler/test_continuous_profiler.py | 13 ++++++++++--- 2 files changed, 23 insertions(+), 6 deletions(-) diff --git a/sentry_sdk/profiler/continuous_profiler.py b/sentry_sdk/profiler/continuous_profiler.py index 00dd29e36c..165bd13837 100644 --- a/sentry_sdk/profiler/continuous_profiler.py +++ b/sentry_sdk/profiler/continuous_profiler.py @@ -75,9 +75,11 @@ def setup_continuous_profiler(options, sdk_info, capture_func): # type: (Dict[str, Any], SDKInfo, Callable[[Envelope], None]) -> bool global _scheduler - if _scheduler is not None: + already_initialized = _scheduler is not None + + if already_initialized: logger.debug("[Profiling] Continuous Profiler is already setup") - return False + teardown_continuous_profiler() if is_gevent(): # If gevent has patched the threading modules then we cannot rely on @@ -117,11 +119,19 @@ def setup_continuous_profiler(options, sdk_info, capture_func): ) ) - atexit.register(teardown_continuous_profiler) + if not already_initialized: + atexit.register(teardown_continuous_profiler) return True +def is_profile_session_sampled(): + # type: () -> bool + if _scheduler is None: + return False + return _scheduler.sampled + + def try_autostart_continuous_profiler(): # type: () -> None diff --git a/tests/profiler/test_continuous_profiler.py b/tests/profiler/test_continuous_profiler.py index 7283ec7164..e4f5cb5e25 100644 --- a/tests/profiler/test_continuous_profiler.py +++ b/tests/profiler/test_continuous_profiler.py @@ -8,6 +8,7 @@ import sentry_sdk from sentry_sdk.consts import VERSION from sentry_sdk.profiler.continuous_profiler import ( + is_profile_session_sampled, get_profiler_id, setup_continuous_profiler, start_profiler, @@ -113,19 +114,25 @@ def test_continuous_profiler_valid_mode(mode, make_options, teardown_profiling): ], ) def test_continuous_profiler_setup_twice(mode, make_options, teardown_profiling): - options = make_options(mode=mode) + assert not is_profile_session_sampled() + # setting up the first time should return True to indicate success + options = make_options(mode=mode, profile_session_sample_rate=1.0) assert setup_continuous_profiler( options, mock_sdk_info, lambda envelope: None, ) - # setting up the second time should return False to indicate no-op - assert not setup_continuous_profiler( + assert is_profile_session_sampled() + + # setting up the second time should return True to indicate re-init + options = make_options(mode=mode, profile_session_sample_rate=0.0) + assert setup_continuous_profiler( options, mock_sdk_info, lambda envelope: None, ) + assert not is_profile_session_sampled() def assert_single_transaction_with_profile_chunks( From 90011260a435f5dba48785c76cb0869c9c99d111 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 10 Sep 2025 08:48:00 +0200 Subject: [PATCH 456/591] Update HuggingFace Hub integration (#4746) Make our existing `huggingface_hub` integration compatible with the new AI Agents insights module. All spans created by the integrations should should create all spans applicable form the [AI Insights ](https://develop.sentry.dev/sdk/telemetry/traces/modules/ai-agents/)documentation. All spans must have the correct span.op, span.name/span.description and span.data/span.attributes set. This makes our SDK and data compatible with v1.36.0 of the Semantic conventions for generative AI systems of OpenTelementry. There are some cases where our AI Insights documentation diverges from Otels semantic conventions. Details for those attributes can be found in the [Sentry conventions](https://getsentry.github.io/sentry-conventions/generated/attributes/gen_ai.html). --- scripts/populate_tox/config.py | 3 + sentry_sdk/consts.py | 1 + sentry_sdk/integrations/__init__.py | 2 +- sentry_sdk/integrations/huggingface_hub.py | 358 +++++-- .../huggingface_hub/test_huggingface_hub.py | 883 +++++++++++++++--- tox.ini | 23 +- 6 files changed, 1050 insertions(+), 220 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 679ffddf2c..bc20d531b3 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -155,6 +155,9 @@ }, "huggingface_hub": { "package": "huggingface_hub", + "deps": { + "*": ["responses"], + }, }, "langchain-base": { "package": "langchain", diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 4f015643d4..cc3c9b1612 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -795,6 +795,7 @@ class OP: GEN_AI_CREATE_AGENT = "gen_ai.create_agent" GEN_AI_EMBEDDINGS = "gen_ai.embeddings" GEN_AI_EXECUTE_TOOL = "gen_ai.execute_tool" + GEN_AI_GENERATE_TEXT = "gen_ai.generate_text" GEN_AI_HANDOFF = "gen_ai.handoff" GEN_AI_PIPELINE = "gen_ai.pipeline" GEN_AI_INVOKE_AGENT = "gen_ai.invoke_agent" diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 7f202221a7..2f5a1f397e 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -141,7 +141,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "gql": (3, 4, 1), "graphene": (3, 3), "grpc": (1, 32, 0), # grpcio - "huggingface_hub": (0, 22), + "huggingface_hub": (0, 24, 7), "langchain": (0, 1, 0), "langgraph": (0, 6, 6), "launchdarkly": (9, 8, 0), diff --git a/sentry_sdk/integrations/huggingface_hub.py b/sentry_sdk/integrations/huggingface_hub.py index 2dfcb5925a..cb76ccf507 100644 --- a/sentry_sdk/integrations/huggingface_hub.py +++ b/sentry_sdk/integrations/huggingface_hub.py @@ -1,24 +1,24 @@ +import inspect from functools import wraps -from sentry_sdk import consts +import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage from sentry_sdk.ai.utils import set_data_normalized -from sentry_sdk.consts import SPANDATA - -from typing import Any, Iterable, Callable - -import sentry_sdk -from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, ) +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable, Iterable + try: import huggingface_hub.inference._client - - from huggingface_hub import ChatCompletionStreamOutput, TextGenerationOutput except ImportError: raise DidNotEnable("Huggingface not installed") @@ -34,9 +34,18 @@ def __init__(self, include_prompts=True): @staticmethod def setup_once(): # type: () -> None + + # Other tasks that can be called: https://huggingface.co/docs/huggingface_hub/guides/inference#supported-providers-and-tasks huggingface_hub.inference._client.InferenceClient.text_generation = ( - _wrap_text_generation( - huggingface_hub.inference._client.InferenceClient.text_generation + _wrap_huggingface_task( + huggingface_hub.inference._client.InferenceClient.text_generation, + OP.GEN_AI_GENERATE_TEXT, + ) + ) + huggingface_hub.inference._client.InferenceClient.chat_completion = ( + _wrap_huggingface_task( + huggingface_hub.inference._client.InferenceClient.chat_completion, + OP.GEN_AI_CHAT, ) ) @@ -51,131 +60,318 @@ def _capture_exception(exc): sentry_sdk.capture_event(event, hint=hint) -def _wrap_text_generation(f): - # type: (Callable[..., Any]) -> Callable[..., Any] +def _wrap_huggingface_task(f, op): + # type: (Callable[..., Any], str) -> Callable[..., Any] @wraps(f) - def new_text_generation(*args, **kwargs): + def new_huggingface_task(*args, **kwargs): # type: (*Any, **Any) -> Any integration = sentry_sdk.get_client().get_integration(HuggingfaceHubIntegration) if integration is None: return f(*args, **kwargs) + prompt = None if "prompt" in kwargs: prompt = kwargs["prompt"] + elif "messages" in kwargs: + prompt = kwargs["messages"] elif len(args) >= 2: - kwargs["prompt"] = args[1] - prompt = kwargs["prompt"] - args = (args[0],) + args[2:] - else: - # invalid call, let it return error + if isinstance(args[1], str) or isinstance(args[1], list): + prompt = args[1] + + if prompt is None: + # invalid call, dont instrument, let it return error return f(*args, **kwargs) - model = kwargs.get("model") - streaming = kwargs.get("stream") + client = args[0] + model = client.model or kwargs.get("model") or "" + operation_name = op.split(".")[-1] span = sentry_sdk.start_span( - op=consts.OP.HUGGINGFACE_HUB_CHAT_COMPLETIONS_CREATE, - name="Text Generation", + op=op, + name=f"{operation_name} {model}", origin=HuggingfaceHubIntegration.origin, ) span.__enter__() + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, operation_name) + + if model: + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model) + + # Input attributes + if should_send_default_pii() and integration.include_prompts: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, prompt, unpack=False + ) + + attribute_mapping = { + "tools": SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, + "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, + "top_k": SPANDATA.GEN_AI_REQUEST_TOP_K, + "stream": SPANDATA.GEN_AI_RESPONSE_STREAMING, + } + + for attribute, span_attribute in attribute_mapping.items(): + value = kwargs.get(attribute, None) + if value is not None: + if isinstance(value, (int, float, bool, str)): + span.set_data(span_attribute, value) + else: + set_data_normalized(span, span_attribute, value, unpack=False) + + # LLM Execution try: res = f(*args, **kwargs) except Exception as e: + # Error Handling + span.set_status("error") _capture_exception(e) span.__exit__(None, None, None) raise e from None + # Output attributes + finish_reason = None + response_model = None + response_text_buffer: list[str] = [] + tokens_used = 0 + tool_calls = None + usage = None + with capture_internal_exceptions(): - if should_send_default_pii() and integration.include_prompts: - set_data_normalized(span, SPANDATA.AI_INPUT_MESSAGES, prompt) + if isinstance(res, str) and res is not None: + response_text_buffer.append(res) - set_data_normalized(span, SPANDATA.AI_MODEL_ID, model) - set_data_normalized(span, SPANDATA.AI_STREAMING, streaming) + if hasattr(res, "generated_text") and res.generated_text is not None: + response_text_buffer.append(res.generated_text) - if isinstance(res, str): - if should_send_default_pii() and integration.include_prompts: - set_data_normalized( - span, - SPANDATA.AI_RESPONSES, - [res], - ) - span.__exit__(None, None, None) - return res + if hasattr(res, "model") and res.model is not None: + response_model = res.model + + if hasattr(res, "details") and hasattr(res.details, "finish_reason"): + finish_reason = res.details.finish_reason + + if ( + hasattr(res, "details") + and hasattr(res.details, "generated_tokens") + and res.details.generated_tokens is not None + ): + tokens_used = res.details.generated_tokens + + if hasattr(res, "usage") and res.usage is not None: + usage = res.usage + + if hasattr(res, "choices") and res.choices is not None: + for choice in res.choices: + if hasattr(choice, "finish_reason"): + finish_reason = choice.finish_reason + if hasattr(choice, "message") and hasattr( + choice.message, "tool_calls" + ): + tool_calls = choice.message.tool_calls + if ( + hasattr(choice, "message") + and hasattr(choice.message, "content") + and choice.message.content is not None + ): + response_text_buffer.append(choice.message.content) - if isinstance(res, TextGenerationOutput): - if should_send_default_pii() and integration.include_prompts: + if response_model is not None: + span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, response_model) + + if finish_reason is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, + finish_reason, + ) + + if should_send_default_pii() and integration.include_prompts: + if tool_calls is not None and len(tool_calls) > 0: set_data_normalized( span, - SPANDATA.AI_RESPONSES, - [res.generated_text], - ) - if res.details is not None and res.details.generated_tokens > 0: - record_token_usage( - span, - total_tokens=res.details.generated_tokens, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + tool_calls, + unpack=False, ) - span.__exit__(None, None, None) - return res - if not isinstance(res, Iterable): - # we only know how to deal with strings and iterables, ignore - set_data_normalized(span, "unknown_response", True) + if len(response_text_buffer) > 0: + text_response = "".join(response_text_buffer) + if text_response: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + text_response, + ) + + if usage is not None: + record_token_usage( + span, + input_tokens=usage.prompt_tokens, + output_tokens=usage.completion_tokens, + total_tokens=usage.total_tokens, + ) + elif tokens_used > 0: + record_token_usage( + span, + total_tokens=tokens_used, + ) + + # If the response is not a generator (meaning a streaming response) + # we are done and can return the response + if not inspect.isgenerator(res): span.__exit__(None, None, None) return res if kwargs.get("details", False): - # res is Iterable[TextGenerationStreamOutput] + # text-generation stream output def new_details_iterator(): - # type: () -> Iterable[ChatCompletionStreamOutput] + # type: () -> Iterable[Any] + finish_reason = None + response_text_buffer: list[str] = [] + tokens_used = 0 + with capture_internal_exceptions(): - tokens_used = 0 - data_buf: list[str] = [] - for x in res: - if hasattr(x, "token") and hasattr(x.token, "text"): - data_buf.append(x.token.text) - if hasattr(x, "details") and hasattr( - x.details, "generated_tokens" + for chunk in res: + if ( + hasattr(chunk, "token") + and hasattr(chunk.token, "text") + and chunk.token.text is not None + ): + response_text_buffer.append(chunk.token.text) + + if hasattr(chunk, "details") and hasattr( + chunk.details, "finish_reason" + ): + finish_reason = chunk.details.finish_reason + + if ( + hasattr(chunk, "details") + and hasattr(chunk.details, "generated_tokens") + and chunk.details.generated_tokens is not None ): - tokens_used = x.details.generated_tokens - yield x - if ( - len(data_buf) > 0 - and should_send_default_pii() - and integration.include_prompts - ): + tokens_used = chunk.details.generated_tokens + + yield chunk + + if finish_reason is not None: set_data_normalized( - span, SPANDATA.AI_RESPONSES, "".join(data_buf) + span, + SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, + finish_reason, ) + + if should_send_default_pii() and integration.include_prompts: + if len(response_text_buffer) > 0: + text_response = "".join(response_text_buffer) + if text_response: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + text_response, + ) + if tokens_used > 0: record_token_usage( span, total_tokens=tokens_used, ) + span.__exit__(None, None, None) return new_details_iterator() - else: - # res is Iterable[str] + else: + # chat-completion stream output def new_iterator(): # type: () -> Iterable[str] - data_buf: list[str] = [] + finish_reason = None + response_model = None + response_text_buffer: list[str] = [] + tool_calls = None + usage = None + with capture_internal_exceptions(): - for s in res: - if isinstance(s, str): - data_buf.append(s) - yield s - if ( - len(data_buf) > 0 - and should_send_default_pii() - and integration.include_prompts - ): + for chunk in res: + if hasattr(chunk, "model") and chunk.model is not None: + response_model = chunk.model + + if hasattr(chunk, "usage") and chunk.usage is not None: + usage = chunk.usage + + if isinstance(chunk, str): + if chunk is not None: + response_text_buffer.append(chunk) + + if hasattr(chunk, "choices") and chunk.choices is not None: + for choice in chunk.choices: + if ( + hasattr(choice, "delta") + and hasattr(choice.delta, "content") + and choice.delta.content is not None + ): + response_text_buffer.append( + choice.delta.content + ) + + if ( + hasattr(choice, "finish_reason") + and choice.finish_reason is not None + ): + finish_reason = choice.finish_reason + + if ( + hasattr(choice, "delta") + and hasattr(choice.delta, "tool_calls") + and choice.delta.tool_calls is not None + ): + tool_calls = choice.delta.tool_calls + + yield chunk + + if response_model is not None: + span.set_data( + SPANDATA.GEN_AI_RESPONSE_MODEL, response_model + ) + + if finish_reason is not None: set_data_normalized( - span, SPANDATA.AI_RESPONSES, "".join(data_buf) + span, + SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, + finish_reason, ) + + if should_send_default_pii() and integration.include_prompts: + if tool_calls is not None and len(tool_calls) > 0: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + tool_calls, + unpack=False, + ) + + if len(response_text_buffer) > 0: + text_response = "".join(response_text_buffer) + if text_response: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TEXT, + text_response, + ) + + if usage is not None: + record_token_usage( + span, + input_tokens=usage.prompt_tokens, + output_tokens=usage.completion_tokens, + total_tokens=usage.total_tokens, + ) + span.__exit__(None, None, None) return new_iterator() - return new_text_generation + return new_huggingface_task diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index df0c6c6d76..86f9c10109 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -1,186 +1,815 @@ -import itertools from unittest import mock - import pytest -from huggingface_hub import ( - InferenceClient, -) -from huggingface_hub.errors import OverloadedError +import responses + +from huggingface_hub import InferenceClient -from sentry_sdk import start_transaction -from sentry_sdk.consts import SPANDATA +import sentry_sdk +from sentry_sdk.utils import package_version from sentry_sdk.integrations.huggingface_hub import HuggingfaceHubIntegration +from typing import TYPE_CHECKING -def mock_client_post(client, post_mock): - # huggingface-hub==0.28.0 deprecates the `post` method - # so patch `_inner_post` instead - if hasattr(client, "post"): - client.post = post_mock - if hasattr(client, "_inner_post"): - client._inner_post = post_mock +try: + from huggingface_hub.utils._errors import HfHubHTTPError +except ImportError: + from huggingface_hub.errors import HfHubHTTPError -@pytest.mark.parametrize( - "send_default_pii, include_prompts, details_arg", - itertools.product([True, False], repeat=3), -) -def test_nonstreaming_chat_completion( - sentry_init, capture_events, send_default_pii, include_prompts, details_arg -): - sentry_init( - integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], - traces_sample_rate=1.0, - send_default_pii=send_default_pii, +if TYPE_CHECKING: + from typing import Any + + +HF_VERSION = package_version("huggingface-hub") + +if HF_VERSION and HF_VERSION < (0, 30, 0): + MODEL_ENDPOINT = "https://api-inference.huggingface.co/models/%7Bmodel_name%7D" + INFERENCE_ENDPOINT = "https://api-inference.huggingface.co/models/%7Bmodel_name%7D" +else: + MODEL_ENDPOINT = "https://huggingface.co/api/models/%7Bmodel_name%7D" + INFERENCE_ENDPOINT = ( + "https://router.huggingface.co/hf-inference/models/%7Bmodel_name%7D" ) - events = capture_events() - client = InferenceClient(model="https://") - if details_arg: - post_mock = mock.Mock( - return_value=b"""[{ - "generated_text": "the model response", +@pytest.fixture +def mock_hf_text_generation_api(): + # type: () -> Any + """Mock HuggingFace text generation API""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "text-generation", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "text-generation", + } + }, + }, + status=200, + ) + + # Mock text generation endpoint + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name), + json={ + "generated_text": "[mocked] Hello! How can i help you?", "details": { "finish_reason": "length", "generated_tokens": 10, "prefill": [], - "tokens": [] - } - }]""" + "tokens": [], + }, + }, + status=200, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_api_with_errors(): + # type: () -> Any + """Mock HuggingFace API that always raises errors for any request""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint with error + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={"error": "Model not found"}, + status=404, + ) + + # Mock text generation endpoint with error + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name), + json={"error": "Internal server error", "message": "Something went wrong"}, + status=500, + ) + + # Mock chat completion endpoint with error + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", + json={"error": "Internal server error", "message": "Something went wrong"}, + status=500, + ) + + # Catch-all pattern for any other model requests + rsps.add( + responses.GET, + "https://huggingface.co/api/models/test-model-error", + json={"error": "Generic model error"}, + status=500, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_text_generation_api_streaming(): + # type: () -> Any + """Mock streaming HuggingFace text generation API""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "text-generation", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "text-generation", + } + }, + }, + status=200, + ) + + # Mock text generation endpoint for streaming + streaming_response = b'data:{"token":{"id":1, "special": false, "text": "the mocked "}}\n\ndata:{"token":{"id":2, "special": false, "text": "model response"}, "details":{"finish_reason": "length", "generated_tokens": 10, "seed": 0}}\n\n' + + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name), + body=streaming_response, + status=200, + headers={ + "Content-Type": "text/event-stream", + "Cache-Control": "no-cache", + "Connection": "keep-alive", + }, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_chat_completion_api(): + # type: () -> Any + """Mock HuggingFace chat completion API""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "conversational", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "conversational", + } + }, + }, + status=200, + ) + + # Mock chat completion endpoint + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", + json={ + "id": "xyz-123", + "created": 1234567890, + "model": f"{model_name}-123", + "system_fingerprint": "fp_123", + "choices": [ + { + "index": 0, + "finish_reason": "stop", + "message": { + "role": "assistant", + "content": "[mocked] Hello! How can I help you today?", + }, + } + ], + "usage": { + "completion_tokens": 8, + "prompt_tokens": 10, + "total_tokens": 18, + }, + }, + status=200, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_chat_completion_api_tools(): + # type: () -> Any + """Mock HuggingFace chat completion API with tool calls.""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "conversational", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "conversational", + } + }, + }, + status=200, + ) + + # Mock chat completion endpoint + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", + json={ + "id": "xyz-123", + "created": 1234567890, + "model": f"{model_name}-123", + "system_fingerprint": "fp_123", + "choices": [ + { + "index": 0, + "finish_reason": "tool_calls", + "message": { + "role": "assistant", + "tool_calls": [ + { + "id": "call_123", + "type": "function", + "function": { + "name": "get_weather", + "arguments": {"location": "Paris"}, + }, + } + ], + }, + } + ], + "usage": { + "completion_tokens": 8, + "prompt_tokens": 10, + "total_tokens": 18, + }, + }, + status=200, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_chat_completion_api_streaming(): + # type: () -> Any + """Mock streaming HuggingFace chat completion API""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "conversational", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "conversational", + } + }, + }, + status=200, + ) + + # Mock chat completion streaming endpoint + streaming_chat_response = ( + b'data:{"id":"xyz-123","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","content":"the mocked "},"index":0,"finish_reason":null}],"usage":null}\n\n' + b'data:{"id":"xyz-124","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","content":"model response"},"index":0,"finish_reason":"stop"}],"usage":{"prompt_tokens":183,"completion_tokens":14,"total_tokens":197}}\n\n' + ) + + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", + body=streaming_chat_response, + status=200, + headers={ + "Content-Type": "text/event-stream", + "Cache-Control": "no-cache", + "Connection": "keep-alive", + }, + ) + + yield rsps + + +@pytest.fixture +def mock_hf_chat_completion_api_streaming_tools(): + # type: () -> Any + """Mock streaming HuggingFace chat completion API with tool calls.""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: + model_name = "test-model" + + # Mock model info endpoint + rsps.add( + responses.GET, + MODEL_ENDPOINT.format(model_name=model_name), + json={ + "id": model_name, + "pipeline_tag": "conversational", + "inferenceProviderMapping": { + "hf-inference": { + "status": "live", + "providerId": model_name, + "task": "conversational", + } + }, + }, + status=200, + ) + + # Mock chat completion streaming endpoint + streaming_chat_response = ( + b'data:{"id":"xyz-123","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","content":"response with tool calls follows"},"index":0,"finish_reason":null}],"usage":null}\n\n' + b'data:{"id":"xyz-124","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","tool_calls": [{"id": "call_123","type": "function","function": {"name": "get_weather", "arguments": {"location": "Paris"}}}]},"index":0,"finish_reason":"tool_calls"}],"usage":{"prompt_tokens":183,"completion_tokens":14,"total_tokens":197}}\n\n' ) - else: - post_mock = mock.Mock( - return_value=b'[{"generated_text": "the model response"}]' + + rsps.add( + responses.POST, + INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", + body=streaming_chat_response, + status=200, + headers={ + "Content-Type": "text/event-stream", + "Cache-Control": "no-cache", + "Connection": "keep-alive", + }, ) - mock_client_post(client, post_mock) - with start_transaction(name="huggingface_hub tx"): - response = client.text_generation( - prompt="hello", - details=details_arg, + yield rsps + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_text_generation( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_text_generation_api, +): + # type: (Any, Any, Any, Any, Any) -> None + sentry_init( + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], + ) + events = capture_events() + + client = InferenceClient(model="test-model") + + with sentry_sdk.start_transaction(name="test"): + client.text_generation( + "Hello", stream=False, + details=True, ) - if details_arg: - assert response.generated_text == "the model response" - else: - assert response == "the model response" - tx = events[0] - assert tx["type"] == "transaction" - span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.huggingface_hub" + + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.generate_text" + assert span["description"] == "generate_text test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "generate_text", + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "length", + "gen_ai.response.streaming": False, + "gen_ai.usage.total_tokens": 10, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] - assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] - else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] - - if details_arg: - assert span["data"]["gen_ai.usage.total_tokens"] == 10 - - -@pytest.mark.parametrize( - "send_default_pii, include_prompts, details_arg", - itertools.product([True, False], repeat=3), -) -def test_streaming_chat_completion( - sentry_init, capture_events, send_default_pii, include_prompts, details_arg + expected_data["gen_ai.request.messages"] = "Hello" + expected_data["gen_ai.response.text"] = "[mocked] Hello! How can i help you?" + + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + + assert span["data"] == expected_data + + # text generation does not set the response model + assert "gen_ai.response.model" not in span["data"] + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_text_generation_streaming( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_text_generation_api_streaming, ): + # type: (Any, Any, Any, Any, Any) -> None sentry_init( + traces_sample_rate=1.0, + send_default_pii=send_default_pii, integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], + ) + events = capture_events() + + client = InferenceClient(model="test-model") + + with sentry_sdk.start_transaction(name="test"): + for _ in client.text_generation( + prompt="Hello", + stream=True, + details=True, + ): + pass + + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.generate_text" + assert span["description"] == "generate_text test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "generate_text", + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "length", + "gen_ai.response.streaming": True, + "gen_ai.usage.total_tokens": 10, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + if send_default_pii and include_prompts: + expected_data["gen_ai.request.messages"] = "Hello" + expected_data["gen_ai.response.text"] = "the mocked model response" + + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + + assert span["data"] == expected_data + + # text generation does not set the response model + assert "gen_ai.response.model" not in span["data"] + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_chat_completion( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_chat_completion_api, +): + # type: (Any, Any, Any, Any, Any) -> None + sentry_init( traces_sample_rate=1.0, send_default_pii=send_default_pii, + integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], ) events = capture_events() - client = InferenceClient(model="https://") - - post_mock = mock.Mock( - return_value=[ - b"""data:{ - "token":{"id":1, "special": false, "text": "the model "} - }""", - b"""data:{ - "token":{"id":2, "special": false, "text": "response"}, - "details":{"finish_reason": "length", "generated_tokens": 10, "seed": 0} - }""", - ] + client = InferenceClient(model="test-model") + + with sentry_sdk.start_transaction(name="test"): + client.chat_completion( + messages=[{"role": "user", "content": "Hello!"}], + stream=False, + ) + + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.chat" + assert span["description"] == "chat test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "chat", + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "stop", + "gen_ai.response.model": "test-model-123", + "gen_ai.response.streaming": False, + "gen_ai.usage.input_tokens": 10, + "gen_ai.usage.output_tokens": 8, + "gen_ai.usage.total_tokens": 18, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + if send_default_pii and include_prompts: + expected_data["gen_ai.request.messages"] = ( + '[{"role": "user", "content": "Hello!"}]' + ) + expected_data["gen_ai.response.text"] = ( + "[mocked] Hello! How can I help you today?" + ) + + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + + assert span["data"] == expected_data + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_chat_completion_streaming( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_chat_completion_api_streaming, +): + # type: (Any, Any, Any, Any, Any) -> None + sentry_init( + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], ) - mock_client_post(client, post_mock) + events = capture_events() + + client = InferenceClient(model="test-model") - with start_transaction(name="huggingface_hub tx"): - response = list( - client.text_generation( - prompt="hello", - details=details_arg, + with sentry_sdk.start_transaction(name="test"): + _ = list( + client.chat_completion( + [{"role": "user", "content": "Hello!"}], stream=True, ) ) - assert len(response) == 2 - if details_arg: - assert response[0].token.text + response[1].token.text == "the model response" - else: - assert response[0] + response[1] == "the model response" - tx = events[0] - assert tx["type"] == "transaction" - span = tx["spans"][0] - assert span["op"] == "ai.chat_completions.create.huggingface_hub" + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.chat" + assert span["description"] == "chat test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "chat", + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "stop", + "gen_ai.response.model": "test-model-123", + "gen_ai.response.streaming": True, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + # usage is not available in older versions of the library + if HF_VERSION and HF_VERSION >= (0, 26, 0): + expected_data["gen_ai.usage.input_tokens"] = 183 + expected_data["gen_ai.usage.output_tokens"] = 14 + expected_data["gen_ai.usage.total_tokens"] = 197 if send_default_pii and include_prompts: - assert "hello" in span["data"][SPANDATA.AI_INPUT_MESSAGES] - assert "the model response" in span["data"][SPANDATA.AI_RESPONSES] - else: - assert SPANDATA.AI_INPUT_MESSAGES not in span["data"] - assert SPANDATA.AI_RESPONSES not in span["data"] + expected_data["gen_ai.request.messages"] = ( + '[{"role": "user", "content": "Hello!"}]' + ) + expected_data["gen_ai.response.text"] = "the mocked model response" - if details_arg: - assert span["data"]["gen_ai.usage.total_tokens"] == 10 + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + assert span["data"] == expected_data -def test_bad_chat_completion(sentry_init, capture_events): - sentry_init(integrations=[HuggingfaceHubIntegration()], traces_sample_rate=1.0) + +def test_chat_completion_api_error( + sentry_init, capture_events, mock_hf_api_with_errors +): + # type: (Any, Any, Any) -> None + sentry_init(traces_sample_rate=1.0) events = capture_events() - client = InferenceClient(model="https://") - post_mock = mock.Mock(side_effect=OverloadedError("The server is overloaded")) - mock_client_post(client, post_mock) + client = InferenceClient(model="test-model") + + with sentry_sdk.start_transaction(name="test"): + with pytest.raises(HfHubHTTPError): + client.chat_completion( + messages=[{"role": "user", "content": "Hello!"}], + ) + + ( + error, + transaction, + ) = events - with pytest.raises(OverloadedError): - client.text_generation(prompt="hello") + assert error["exception"]["values"][0]["mechanism"]["type"] == "huggingface_hub" + assert not error["exception"]["values"][0]["mechanism"]["handled"] - (event,) = events - assert event["level"] == "error" + (span,) = transaction["spans"] + assert span["op"] == "gen_ai.chat" + assert span["description"] == "chat test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + assert span.get("tags", {}).get("status") == "error" -def test_span_origin(sentry_init, capture_events): + assert ( + error["contexts"]["trace"]["trace_id"] + == transaction["contexts"]["trace"]["trace_id"] + ) + expected_data = { + "gen_ai.operation.name": "chat", + "gen_ai.request.model": "test-model", + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + assert span["data"] == expected_data + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_chat_completion_with_tools( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_chat_completion_api_tools, +): + # type: (Any, Any, Any, Any, Any) -> None sentry_init( - integrations=[HuggingfaceHubIntegration()], traces_sample_rate=1.0, + send_default_pii=send_default_pii, + integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], ) events = capture_events() - client = InferenceClient(model="https://") - post_mock = mock.Mock( - return_value=[ - b"""data:{ - "token":{"id":1, "special": false, "text": "the model "} - }""", - ] + client = InferenceClient(model="test-model") + + tools = [ + { + "type": "function", + "function": { + "name": "get_weather", + "description": "Get current weather", + "parameters": { + "type": "object", + "properties": {"location": {"type": "string"}}, + "required": ["location"], + }, + }, + } + ] + + with sentry_sdk.start_transaction(name="test"): + client.chat_completion( + messages=[{"role": "user", "content": "What is the weather in Paris?"}], + tools=tools, + tool_choice="auto", + ) + + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.chat" + assert span["description"] == "chat test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "chat", + "gen_ai.request.available_tools": '[{"type": "function", "function": {"name": "get_weather", "description": "Get current weather", "parameters": {"type": "object", "properties": {"location": {"type": "string"}}, "required": ["location"]}}}]', + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "tool_calls", + "gen_ai.response.model": "test-model-123", + "gen_ai.usage.input_tokens": 10, + "gen_ai.usage.output_tokens": 8, + "gen_ai.usage.total_tokens": 18, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + if send_default_pii and include_prompts: + expected_data["gen_ai.request.messages"] = ( + '[{"role": "user", "content": "What is the weather in Paris?"}]' + ) + expected_data["gen_ai.response.tool_calls"] = ( + '[{"function": {"arguments": {"location": "Paris"}, "name": "get_weather", "description": "None"}, "id": "call_123", "type": "function"}]' + ) + + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + assert "gen_ai.response.tool_calls" not in expected_data + + assert span["data"] == expected_data + + +@pytest.mark.parametrize("send_default_pii", [True, False]) +@pytest.mark.parametrize("include_prompts", [True, False]) +def test_chat_completion_streaming_with_tools( + sentry_init, + capture_events, + send_default_pii, + include_prompts, + mock_hf_chat_completion_api_streaming_tools, +): + # type: (Any, Any, Any, Any, Any) -> None + sentry_init( + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + integrations=[HuggingfaceHubIntegration(include_prompts=include_prompts)], ) - mock_client_post(client, post_mock) + events = capture_events() - with start_transaction(name="huggingface_hub tx"): - list( - client.text_generation( - prompt="hello", + client = InferenceClient(model="test-model") + + tools = [ + { + "type": "function", + "function": { + "name": "get_weather", + "description": "Get current weather", + "parameters": { + "type": "object", + "properties": {"location": {"type": "string"}}, + "required": ["location"], + }, + }, + } + ] + + with sentry_sdk.start_transaction(name="test"): + _ = list( + client.chat_completion( + messages=[{"role": "user", "content": "What is the weather in Paris?"}], stream=True, + tools=tools, + tool_choice="auto", ) ) - (event,) = events + (transaction,) = events + (span,) = transaction["spans"] + + assert span["op"] == "gen_ai.chat" + assert span["description"] == "chat test-model" + assert span["origin"] == "auto.ai.huggingface_hub" + + expected_data = { + "gen_ai.operation.name": "chat", + "gen_ai.request.available_tools": '[{"type": "function", "function": {"name": "get_weather", "description": "Get current weather", "parameters": {"type": "object", "properties": {"location": {"type": "string"}}, "required": ["location"]}}}]', + "gen_ai.request.model": "test-model", + "gen_ai.response.finish_reasons": "tool_calls", + "gen_ai.response.model": "test-model-123", + "gen_ai.response.streaming": True, + "thread.id": mock.ANY, + "thread.name": mock.ANY, + } + + if HF_VERSION and HF_VERSION >= (0, 26, 0): + expected_data["gen_ai.usage.input_tokens"] = 183 + expected_data["gen_ai.usage.output_tokens"] = 14 + expected_data["gen_ai.usage.total_tokens"] = 197 + + if send_default_pii and include_prompts: + expected_data["gen_ai.request.messages"] = ( + '[{"role": "user", "content": "What is the weather in Paris?"}]' + ) + expected_data["gen_ai.response.text"] = "response with tool calls follows" + expected_data["gen_ai.response.tool_calls"] = ( + '[{"function": {"arguments": {"location": "Paris"}, "name": "get_weather"}, "id": "call_123", "type": "function", "index": "None"}]' + ) + + if not send_default_pii or not include_prompts: + assert "gen_ai.request.messages" not in expected_data + assert "gen_ai.response.text" not in expected_data + assert "gen_ai.response.tool_calls" not in expected_data - assert event["contexts"]["trace"]["origin"] == "manual" - assert event["spans"][0]["origin"] == "auto.ai.huggingface_hub" + assert span["data"] == expected_data diff --git a/tox.ini b/tox.ini index ff2403f515..1bc9757b9a 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-08T11:35:09.849536+00:00 +# Last generated: 2025-09-09T08:24:12.875177+00:00 [tox] requires = @@ -116,12 +116,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.36.1 {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.106.1 + {py3.8,py3.12,py3.13}-openai-base-v1.107.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.106.1 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -130,8 +130,8 @@ envlist = {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.8,py3.10,py3.11}-huggingface_hub-v0.22.2 - {py3.8,py3.11,py3.12}-huggingface_hub-v0.26.5 + {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.27.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.4 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 @@ -141,7 +141,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.25 + {py3.9,py3.12,py3.13}-boto3-v1.40.26 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -487,7 +487,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.36.1: openai==1.36.1 openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.106.1: openai==1.106.1 + openai-base-v1.107.0: openai==1.107.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -496,7 +496,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.36.1: openai==1.36.1 openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.106.1: openai==1.106.1 + openai-notiktoken-v1.107.0: openai==1.107.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 @@ -509,18 +509,19 @@ deps = openai_agents-v0.2.11: openai-agents==0.2.11 openai_agents: pytest-asyncio - huggingface_hub-v0.22.2: huggingface_hub==0.22.2 - huggingface_hub-v0.26.5: huggingface_hub==0.26.5 + huggingface_hub-v0.24.7: huggingface_hub==0.24.7 + huggingface_hub-v0.27.1: huggingface_hub==0.27.1 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 huggingface_hub-v0.34.4: huggingface_hub==0.34.4 huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 + huggingface_hub: responses # ~~~ Cloud ~~~ boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.25: boto3==1.40.25 + boto3-v1.40.26: boto3==1.40.26 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 From 94a92d8b6475905a247472ced68911320a4126c4 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 10 Sep 2025 15:02:02 +0200 Subject: [PATCH 457/591] Add input and output to `invoke_agent` span. (#4785) Add the original input for the agent and the final output of the agent to the `invoke_agent` span. fixes #4512 fixes PY-1740 --- .../openai_agents/patches/agent_run.py | 8 +-- .../openai_agents/spans/invoke_agent.py | 53 +++++++++++++++++-- .../openai_agents/test_openai_agents.py | 16 ++++++ 3 files changed, 68 insertions(+), 9 deletions(-) diff --git a/sentry_sdk/integrations/openai_agents/patches/agent_run.py b/sentry_sdk/integrations/openai_agents/patches/agent_run.py index 29002f6619..5473915b48 100644 --- a/sentry_sdk/integrations/openai_agents/patches/agent_run.py +++ b/sentry_sdk/integrations/openai_agents/patches/agent_run.py @@ -26,12 +26,12 @@ def _patch_agent_run(): original_execute_handoffs = agents._run_impl.RunImpl.execute_handoffs original_execute_final_output = agents._run_impl.RunImpl.execute_final_output - def _start_invoke_agent_span(context_wrapper, agent): - # type: (agents.RunContextWrapper, agents.Agent) -> None + def _start_invoke_agent_span(context_wrapper, agent, kwargs): + # type: (agents.RunContextWrapper, agents.Agent, dict[str, Any]) -> None """Start an agent invocation span""" # Store the agent on the context wrapper so we can access it later context_wrapper._sentry_current_agent = agent - invoke_agent_span(context_wrapper, agent) + invoke_agent_span(context_wrapper, agent, kwargs) def _end_invoke_agent_span(context_wrapper, agent, output=None): # type: (agents.RunContextWrapper, agents.Agent, Optional[Any]) -> None @@ -72,7 +72,7 @@ async def patched_run_single_turn(cls, *args, **kwargs): if current_agent and current_agent != agent: _end_invoke_agent_span(context_wrapper, current_agent) - _start_invoke_agent_span(context_wrapper, agent) + _start_invoke_agent_span(context_wrapper, agent, kwargs) # Call original method with all the correct parameters result = await original_run_single_turn(*args, **kwargs) diff --git a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py index 549ade1246..d76d39f338 100644 --- a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py +++ b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py @@ -1,5 +1,8 @@ import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import safe_serialize from ..consts import SPAN_ORIGIN from ..utils import _set_agent_data @@ -11,8 +14,8 @@ from typing import Any -def invoke_agent_span(context, agent): - # type: (agents.RunContextWrapper, agents.Agent) -> sentry_sdk.tracing.Span +def invoke_agent_span(context, agent, kwargs): + # type: (agents.RunContextWrapper, agents.Agent, dict[str, Any]) -> sentry_sdk.tracing.Span span = sentry_sdk.start_span( op=OP.GEN_AI_INVOKE_AGENT, name=f"invoke_agent {agent.name}", @@ -22,6 +25,40 @@ def invoke_agent_span(context, agent): span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + if should_send_default_pii(): + messages = [] + if agent.instructions: + message = ( + agent.instructions + if isinstance(agent.instructions, str) + else safe_serialize(agent.instructions) + ) + messages.append( + { + "content": [{"text": message, "type": "text"}], + "role": "system", + } + ) + + original_input = kwargs.get("original_input") + if original_input is not None: + message = ( + original_input + if isinstance(original_input, str) + else safe_serialize(original_input) + ) + messages.append( + { + "content": [{"text": message, "type": "text"}], + "role": "user", + } + ) + + if len(messages) > 0: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + ) + _set_agent_data(span, agent) return span @@ -29,6 +66,12 @@ def invoke_agent_span(context, agent): def update_invoke_agent_span(context, agent, output): # type: (agents.RunContextWrapper, agents.Agent, Any) -> None - current_span = sentry_sdk.get_current_span() - if current_span: - current_span.__exit__(None, None, None) + span = sentry_sdk.get_current_span() + + if span: + if should_send_default_pii(): + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, output, unpack=False + ) + + span.__exit__(None, None, None) diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index fab8d9e13f..047b919213 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -115,6 +115,7 @@ async def test_agent_invocation_span( sentry_init( integrations=[OpenAIAgentsIntegration()], traces_sample_rate=1.0, + send_default_pii=True, ) events = capture_events() @@ -134,6 +135,21 @@ async def test_agent_invocation_span( assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" assert invoke_agent_span["description"] == "invoke_agent test_agent" + assert invoke_agent_span["data"]["gen_ai.request.messages"] == safe_serialize( + [ + { + "content": [ + {"text": "You are a helpful test assistant.", "type": "text"} + ], + "role": "system", + }, + {"content": [{"text": "Test input", "type": "text"}], "role": "user"}, + ] + ) + assert ( + invoke_agent_span["data"]["gen_ai.response.text"] + == "Hello, how can I help you?" + ) assert invoke_agent_span["data"]["gen_ai.operation.name"] == "invoke_agent" assert invoke_agent_span["data"]["gen_ai.system"] == "openai" assert invoke_agent_span["data"]["gen_ai.agent.name"] == "test_agent" From a6a2f930ff95807d64ff5bfb892435b3d06cdd2e Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 11 Sep 2025 15:57:58 +0200 Subject: [PATCH 458/591] Add log message when `DedupeIntegration` is dropping an error. (#4788) Make it clearer for the user what is happening. resolves: PY-1840 --- sentry_sdk/integrations/dedupe.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/dedupe.py b/sentry_sdk/integrations/dedupe.py index a115e35292..eab2764fcd 100644 --- a/sentry_sdk/integrations/dedupe.py +++ b/sentry_sdk/integrations/dedupe.py @@ -1,5 +1,5 @@ import sentry_sdk -from sentry_sdk.utils import ContextVar +from sentry_sdk.utils import ContextVar, logger from sentry_sdk.integrations import Integration from sentry_sdk.scope import add_global_event_processor @@ -37,7 +37,9 @@ def processor(event, hint): exc = exc_info[1] if integration._last_seen.get(None) is exc: + logger.info("DedupeIntegration dropped duplicated error event %s", exc) return None + integration._last_seen.set(exc) return event From bd311b20c686778a7e83394dfa06970af2bcc90f Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 11 Sep 2025 14:12:12 +0000 Subject: [PATCH 459/591] build(deps): bump actions/setup-python from 5 to 6 (#4774) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/setup-python](https://github.com/actions/setup-python) from 5 to 6.
Release notes

Sourced from actions/setup-python's releases.

v6.0.0

What's Changed

Breaking Changes

Make sure your runner is on version v2.327.1 or later to ensure compatibility with this release. See Release Notes

Enhancements:

Bug fixes:

Dependency updates:

New Contributors

Full Changelog: https://github.com/actions/setup-python/compare/v5...v6.0.0

v5.6.0

What's Changed

Full Changelog: https://github.com/actions/setup-python/compare/v5...v5.6.0

v5.5.0

What's Changed

Enhancements:

Bug fixes:

... (truncated)

Commits
  • e797f83 Upgrade to node 24 (#1164)
  • 3d1e2d2 Revert "Enhance cache-dependency-path handling to support files outside the w...
  • 65b0712 Clarify pythonLocation behavior for PyPy and GraalPy in environment variables...
  • 5b668cf Bump actions/checkout from 4 to 5 (#1181)
  • f62a0e2 Change missing cache directory error to warning (#1182)
  • 9322b3c Upgrade setuptools to 78.1.1 to fix path traversal vulnerability in PackageIn...
  • fbeb884 Bump form-data to fix critical vulnerabilities #182 & #183 (#1163)
  • 03bb615 Bump idna from 2.9 to 3.7 in /tests/data (#843)
  • 36da51d Add version parsing from Pipfile (#1067)
  • 3c6f142 update documentation (#1156)
  • Additional commits viewable in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/setup-python&package-manager=github_actions&previous-version=5&new-version=6)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/ci.yml | 8 ++++---- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 14 files changed, 21 insertions(+), 21 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index ffc0a741fc..67b4fd3546 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -25,7 +25,7 @@ jobs: steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 with: python-version: 3.12 @@ -40,7 +40,7 @@ jobs: steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 with: python-version: 3.12 @@ -59,7 +59,7 @@ jobs: steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 with: python-version: 3.12 - name: Setup build cache @@ -90,7 +90,7 @@ jobs: steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 with: python-version: 3.12 diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 26a8bdb8bb..f65ee87ec3 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 62e70d759d..92c7d40ff4 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -43,7 +43,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 1c0c9b80d2..ef1fab573c 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index 2d6af43bc3..f22487eb54 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -57,7 +57,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} @@ -156,7 +156,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index f744f514ee..d7acf0670d 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 382e6a5f15..c32102df8c 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 93675fb4fe..578b7d65bf 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index e8937708bc..c2673350b2 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 867681d3a3..9520d8ef4d 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} @@ -106,7 +106,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index f842683285..051567b92b 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} @@ -133,7 +133,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index ba802faa01..6131ff4250 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -57,7 +57,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 22200f8ae1..c59553a88a 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -39,7 +39,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} @@ -134,7 +134,7 @@ jobs: container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 if: ${{ matrix.python-version != '3.6' }} with: python-version: ${{ matrix.python-version }} diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 4ac0d03eb2..28e18c501b 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -43,7 +43,7 @@ {% raw %}container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }}{% endraw %} steps: - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v5 + - uses: actions/setup-python@v6 {% raw %}if: ${{ matrix.python-version != '3.6' }}{% endraw %} with: python-version: {% raw %}${{ matrix.python-version }}{% endraw %} From 18d38996c7fba2ddde885d98a5cd808eae1cae26 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 11 Sep 2025 14:15:24 +0000 Subject: [PATCH 460/591] build(deps): bump codecov/codecov-action from 5.5.0 to 5.5.1 (#4773) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [codecov/codecov-action](https://github.com/codecov/codecov-action) from 5.5.0 to 5.5.1.
Release notes

Sourced from codecov/codecov-action's releases.

v5.5.1

What's Changed

New Contributors

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.5.0...v5.5.1

Changelog

Sourced from codecov/codecov-action's changelog.

v5.5.1

What's Changed

Full Changelog: https://github.com/codecov/codecov-action/compare/v5.5.0..v5.5.1

Commits

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=codecov/codecov-action&package-manager=github_actions&previous-version=5.5.0&new-version=5.5.1)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Anton Pirker --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 4 ++-- .github/workflows/test-integrations-flags.yml | 2 +- .github/workflows/test-integrations-gevent.yml | 2 +- .github/workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-network.yml | 4 ++-- .github/workflows/test-integrations-tasks.yml | 4 ++-- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 4 ++-- scripts/split_tox_gh_actions/templates/test_group.jinja | 2 +- 13 files changed, 17 insertions(+), 17 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index f65ee87ec3..972df704e0 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -99,7 +99,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 92c7d40ff4..6aeaea8c3a 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index ef1fab573c..b682428dd1 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index f22487eb54..efa9f8db39 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -107,7 +107,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -206,7 +206,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index d7acf0670d..d7baeeb870 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index c32102df8c..9af6b4d7af 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -67,7 +67,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 578b7d65bf..5c306dff3f 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -79,7 +79,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index c2673350b2..005e8395a2 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -87,7 +87,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 9520d8ef4d..e34706ff09 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -75,7 +75,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -142,7 +142,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 051567b92b..0038f1d050 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -102,7 +102,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -196,7 +196,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 6131ff4250..4b22db6155 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -97,7 +97,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index c59553a88a..6b7fe58815 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -103,7 +103,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml @@ -198,7 +198,7 @@ jobs: coverage xml - name: Upload coverage to Codecov if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: ${{ secrets.CODECOV_TOKEN }} files: coverage.xml diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index 28e18c501b..f020a44b84 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -100,7 +100,7 @@ - name: Upload coverage to Codecov if: {% raw %}${{ !cancelled() }}{% endraw %} - uses: codecov/codecov-action@v5.5.0 + uses: codecov/codecov-action@v5.5.1 with: token: {% raw %}${{ secrets.CODECOV_TOKEN }}{% endraw %} files: coverage.xml From 73e2b7e2d4a5c1bd833314a7d40606e98debae13 Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Thu, 11 Sep 2025 16:30:06 +0200 Subject: [PATCH 461/591] feat(ai): Create transaction in AI agents framworks, when no transaction is running. (#4758) This includes: - nthropic: `message.create` - langchain: `invoke_agent` spans - openai_agent: `invoke_agent` spans + agent workflow (which was already like that) Closes https://linear.app/getsentry/issue/TET-1048/auto-wrap-gen-aiinvoke-agent-if-no-transaction-in-scope Co-authored-by: Anton Pirker --- sentry_sdk/ai/utils.py | 12 +++++++++++- sentry_sdk/integrations/anthropic.py | 4 ++-- sentry_sdk/integrations/langchain.py | 8 +++++--- .../openai_agents/spans/agent_workflow.py | 4 ++-- .../integrations/openai_agents/spans/invoke_agent.py | 5 +++-- sentry_sdk/integrations/openai_agents/utils.py | 10 ---------- 6 files changed, 23 insertions(+), 20 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index 2dc0de4ef3..41f89f5623 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -3,9 +3,10 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Any + from typing import Any, Callable from sentry_sdk.tracing import Span +import sentry_sdk from sentry_sdk.utils import logger @@ -37,3 +38,12 @@ def set_data_normalized(span, key, value, unpack=True): span.set_data(key, normalized) else: span.set_data(key, json.dumps(normalized)) + + +def get_start_span_function(): + # type: () -> Callable[..., Any] + current_span = sentry_sdk.get_current_span() + transaction_exists = ( + current_span is not None and current_span.containing_transaction == current_span + ) + return sentry_sdk.start_span if transaction_exists else sentry_sdk.start_transaction diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 05d45ef62f..ff3d9a3388 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import set_data_normalized, get_start_span_function from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -194,7 +194,7 @@ def _sentry_patched_create_common(f, *args, **kwargs): model = kwargs.get("model", "") - span = sentry_sdk.start_span( + span = get_start_span_function()( op=OP.GEN_AI_CHAT, name=f"chat {model}".strip(), origin=AnthropicIntegration.origin, diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index e14dd619fe..1401be06e1 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -4,7 +4,7 @@ import sentry_sdk from sentry_sdk.ai.monitoring import set_ai_pipeline_name -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import set_data_normalized, get_start_span_function from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -716,8 +716,9 @@ def new_invoke(self, *args, **kwargs): return f(self, *args, **kwargs) agent_name, tools = _get_request_data(self, args, kwargs) + start_span_function = get_start_span_function() - with sentry_sdk.start_span( + with start_span_function( op=OP.GEN_AI_INVOKE_AGENT, name=f"invoke_agent {agent_name}" if agent_name else "invoke_agent", origin=LangchainIntegration.origin, @@ -767,8 +768,9 @@ def new_stream(self, *args, **kwargs): return f(self, *args, **kwargs) agent_name, tools = _get_request_data(self, args, kwargs) + start_span_function = get_start_span_function() - span = sentry_sdk.start_span( + span = start_span_function( op=OP.GEN_AI_INVOKE_AGENT, name=f"invoke_agent {agent_name}".strip(), origin=LangchainIntegration.origin, diff --git a/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py b/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py index de2f28d41e..ef69b856e3 100644 --- a/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py +++ b/sentry_sdk/integrations/openai_agents/spans/agent_workflow.py @@ -1,7 +1,7 @@ import sentry_sdk +from sentry_sdk.ai.utils import get_start_span_function from ..consts import SPAN_ORIGIN -from ..utils import _get_start_span_function from typing import TYPE_CHECKING @@ -13,7 +13,7 @@ def agent_workflow_span(agent): # type: (agents.Agent) -> sentry_sdk.tracing.Span # Create a transaction or a span if an transaction is already active - span = _get_start_span_function()( + span = get_start_span_function()( name=f"{agent.name} workflow", origin=SPAN_ORIGIN, ) diff --git a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py index d76d39f338..cf06120625 100644 --- a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py +++ b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py @@ -1,5 +1,5 @@ import sentry_sdk -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import get_start_span_function, set_data_normalized from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import safe_serialize @@ -16,7 +16,8 @@ def invoke_agent_span(context, agent, kwargs): # type: (agents.RunContextWrapper, agents.Agent, dict[str, Any]) -> sentry_sdk.tracing.Span - span = sentry_sdk.start_span( + start_span_function = get_start_span_function() + span = start_span_function( op=OP.GEN_AI_INVOKE_AGENT, name=f"invoke_agent {agent.name}", origin=SPAN_ORIGIN, diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index 44b260d4bc..a0487e0e3a 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -9,7 +9,6 @@ if TYPE_CHECKING: from typing import Any - from typing import Callable from agents import Usage try: @@ -29,15 +28,6 @@ def _capture_exception(exc): sentry_sdk.capture_event(event, hint=hint) -def _get_start_span_function(): - # type: () -> Callable[..., Any] - current_span = sentry_sdk.get_current_span() - transaction_exists = ( - current_span is not None and current_span.containing_transaction == current_span - ) - return sentry_sdk.start_span if transaction_exists else sentry_sdk.start_transaction - - def _set_agent_data(span, agent): # type: (sentry_sdk.tracing.Span, agents.Agent) -> None span.set_data( From dcefe3840181e0aa534040b3932657f0ca1c36b8 Mon Sep 17 00:00:00 2001 From: Vadim Markovtsev Date: Thu, 11 Sep 2025 16:36:12 +0200 Subject: [PATCH 462/591] Avoid reporting false-positive StopAsyncIteration in the asyncio integration (#4741) If a coroutine exits an async loop by raising `AsyncStopIteration`, Sentry reports it as an error. There is no error in that case. Co-authored-by: Anton Pirker --- sentry_sdk/integrations/asyncio.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/sentry_sdk/integrations/asyncio.py b/sentry_sdk/integrations/asyncio.py index ae580ca038..66742fe6e4 100644 --- a/sentry_sdk/integrations/asyncio.py +++ b/sentry_sdk/integrations/asyncio.py @@ -51,6 +51,8 @@ async def _task_with_sentry_span_creation(): ): try: result = await coro + except StopAsyncIteration as e: + raise e from None except Exception: reraise(*_capture_exception()) From 007058d0e66a6156c258d2acef76d43efad49223 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 12 Sep 2025 09:10:03 +0200 Subject: [PATCH 463/591] ref(tracing): Use float for sample rand (#4677) Closes https://github.com/getsentry/sentry-python/issues/4270 --------- Co-authored-by: Anton Pirker --- sentry_sdk/tracing.py | 3 +- sentry_sdk/tracing_utils.py | 40 +++++++++---------- sentry_sdk/utils.py | 6 +++ tests/integrations/aiohttp/test_aiohttp.py | 2 +- tests/integrations/celery/test_celery.py | 4 +- tests/integrations/httpx/test_httpx.py | 4 +- tests/integrations/stdlib/test_httplib.py | 2 +- tests/test_dsc.py | 2 +- tests/test_monitor.py | 2 +- tests/test_propagationcontext.py | 17 ++++---- tests/tracing/test_integration_tests.py | 2 +- tests/tracing/test_sample_rand.py | 37 +---------------- tests/tracing/test_sample_rand_propagation.py | 6 +-- 13 files changed, 49 insertions(+), 78 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 0d1fcc45da..fc43a33dc7 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -1,4 +1,3 @@ -from decimal import Decimal import uuid import warnings from datetime import datetime, timedelta, timezone @@ -1251,7 +1250,7 @@ def _set_initial_sampling_decision(self, sampling_context): return # Now we roll the dice. - self.sampled = self._sample_rand < Decimal.from_float(self.sample_rate) + self.sampled = self._sample_rand < self.sample_rate if self.sampled: logger.debug( diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index b31d3d85c5..c1cfde293b 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -6,7 +6,6 @@ import sys from collections.abc import Mapping from datetime import timedelta -from decimal import ROUND_DOWN, Decimal, DefaultContext, localcontext from random import Random from urllib.parse import quote, unquote import uuid @@ -502,7 +501,7 @@ def _fill_sample_rand(self): return sample_rand = try_convert( - Decimal, self.dynamic_sampling_context.get("sample_rand") + float, self.dynamic_sampling_context.get("sample_rand") ) if sample_rand is not None and 0 <= sample_rand < 1: # sample_rand is present and valid, so don't overwrite it @@ -650,7 +649,7 @@ def populate_from_transaction(cls, transaction): options = client.options or {} sentry_items["trace_id"] = transaction.trace_id - sentry_items["sample_rand"] = str(transaction._sample_rand) + sentry_items["sample_rand"] = f"{transaction._sample_rand:.6f}" # noqa: E231 if options.get("environment"): sentry_items["environment"] = options["environment"] @@ -724,15 +723,15 @@ def strip_sentry_baggage(header): ) def _sample_rand(self): - # type: () -> Optional[Decimal] + # type: () -> Optional[float] """Convenience method to get the sample_rand value from the sentry_items. - We validate the value and parse it as a Decimal before returning it. The value is considered - valid if it is a Decimal in the range [0, 1). + We validate the value and parse it as a float before returning it. The value is considered + valid if it is a float in the range [0, 1). """ - sample_rand = try_convert(Decimal, self.sentry_items.get("sample_rand")) + sample_rand = try_convert(float, self.sentry_items.get("sample_rand")) - if sample_rand is not None and Decimal(0) <= sample_rand < Decimal(1): + if sample_rand is not None and 0.0 <= sample_rand < 1.0: return sample_rand return None @@ -898,7 +897,7 @@ def _generate_sample_rand( *, interval=(0.0, 1.0), # type: tuple[float, float] ): - # type: (...) -> Decimal + # type: (...) -> float """Generate a sample_rand value from a trace ID. The generated value will be pseudorandomly chosen from the provided @@ -913,19 +912,16 @@ def _generate_sample_rand( raise ValueError("Invalid interval: lower must be less than upper") rng = Random(trace_id) - sample_rand = upper - while sample_rand >= upper: - sample_rand = rng.uniform(lower, upper) - - # Round down to exactly six decimal-digit precision. - # Setting the context is needed to avoid an InvalidOperation exception - # in case the user has changed the default precision or set traps. - with localcontext(DefaultContext) as ctx: - ctx.prec = 6 - return Decimal(sample_rand).quantize( - Decimal("0.000001"), - rounding=ROUND_DOWN, - ) + lower_scaled = int(lower * 1_000_000) + upper_scaled = int(upper * 1_000_000) + try: + sample_rand_scaled = rng.randrange(lower_scaled, upper_scaled) + except ValueError: + # In some corner cases it might happen that the range is too small + # In that case, just take the lower bound + sample_rand_scaled = lower_scaled + + return sample_rand_scaled / 1_000_000 def _sample_rand_range(parent_sampled, sample_rate): diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index b0f3fa4a4c..3fe3ac3eec 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1934,6 +1934,12 @@ def try_convert(convert_func, value): given function. Return None if the conversion fails, i.e. if the function raises an exception. """ + try: + if isinstance(value, convert_func): # type: ignore + return value + except TypeError: + pass + try: return convert_func(value) except Exception: diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index dbb4286370..267ce08fdd 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -618,7 +618,7 @@ async def handler(request): raw_server = await aiohttp_raw_server(handler) - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=500000): with start_transaction( name="/interactions/other-dogs/new-dog", op="greeting.sniff", diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index ce2e693143..80b4a423cb 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -518,8 +518,8 @@ def test_baggage_propagation(init_celery): def dummy_task(self, x, y): return _get_headers(self) - # patch random.uniform to return a predictable sample_rand value - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + # patch random.randrange to return a predictable sample_rand value + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=500000): with start_transaction() as transaction: result = dummy_task.apply_async( args=(1, 0), diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index 5a35b68076..ba2575ce59 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -170,8 +170,8 @@ def test_outgoing_trace_headers_append_to_baggage( url = "http://example.com/" - # patch random.uniform to return a predictable sample_rand value - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.5): + # patch random.randrange to return a predictable sample_rand value + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=500000): with start_transaction( name="/interactions/other-dogs/new-dog", op="greeting.sniff", diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index f6735d0e74..b8d46d0558 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -236,7 +236,7 @@ def test_outgoing_trace_headers_head_sdk(sentry_init, monkeypatch): monkeypatch.setattr(HTTPSConnection, "send", mock_send) sentry_init(traces_sample_rate=0.5, release="foo") - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.25): + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=250000): transaction = Transaction.continue_from_headers({}) with start_transaction(transaction=transaction, name="Head SDK tx") as transaction: diff --git a/tests/test_dsc.py b/tests/test_dsc.py index 8e549d0cf8..6097af7f95 100644 --- a/tests/test_dsc.py +++ b/tests/test_dsc.py @@ -175,7 +175,7 @@ def my_traces_sampler(sampling_context): } # We continue the incoming trace and start a new transaction - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.125): + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=125000): transaction = sentry_sdk.continue_trace(incoming_http_headers) with sentry_sdk.start_transaction(transaction, name="foo"): pass diff --git a/tests/test_monitor.py b/tests/test_monitor.py index b48d9f6282..9ffc943bed 100644 --- a/tests/test_monitor.py +++ b/tests/test_monitor.py @@ -73,7 +73,7 @@ def test_transaction_uses_downsampled_rate( assert monitor.downsample_factor == 1 # make sure we don't sample the transaction - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.75): + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=750000): with sentry_sdk.start_transaction(name="foobar") as transaction: assert transaction.sampled is False assert transaction.sample_rate == 0.5 diff --git a/tests/test_propagationcontext.py b/tests/test_propagationcontext.py index a0ce1094fa..078a69c72b 100644 --- a/tests/test_propagationcontext.py +++ b/tests/test_propagationcontext.py @@ -136,13 +136,13 @@ def test_sample_rand_filled(parent_sampled, sample_rate, expected_interval): else: sample_rate_str = "" - # for convenience, we'll just return the lower bound of the interval - mock_uniform = mock.Mock(return_value=expected_interval[0]) + # for convenience, we'll just return the lower bound of the interval as an integer + mock_randrange = mock.Mock(return_value=int(expected_interval[0] * 1000000)) def mock_random_class(seed): assert seed == "00000000000000000000000000000000", "seed should be the trace_id" rv = Mock() - rv.uniform = mock_uniform + rv.randrange = mock_randrange return rv with mock.patch("sentry_sdk.tracing_utils.Random", mock_random_class): @@ -158,17 +158,20 @@ def mock_random_class(seed): ctx.dynamic_sampling_context["sample_rand"] == f"{expected_interval[0]:.6f}" # noqa: E231 ) - assert mock_uniform.call_count == 1 - assert mock_uniform.call_args[0] == expected_interval + assert mock_randrange.call_count == 1 + assert mock_randrange.call_args[0] == ( + int(expected_interval[0] * 1000000), + int(expected_interval[1] * 1000000), + ) def test_sample_rand_rounds_down(): # Mock value that should round down to 0.999_999 - mock_uniform = mock.Mock(return_value=0.999_999_9) + mock_randrange = mock.Mock(return_value=999999) def mock_random_class(_): rv = Mock() - rv.uniform = mock_uniform + rv.randrange = mock_randrange return rv with mock.patch("sentry_sdk.tracing_utils.Random", mock_random_class): diff --git a/tests/tracing/test_integration_tests.py b/tests/tracing/test_integration_tests.py index 61ef14b7d0..8b5659b694 100644 --- a/tests/tracing/test_integration_tests.py +++ b/tests/tracing/test_integration_tests.py @@ -169,7 +169,7 @@ def test_dynamic_sampling_head_sdk_creates_dsc( envelopes = capture_envelopes() # make sure transaction is sampled for both cases - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", return_value=0.25): + with mock.patch("sentry_sdk.tracing_utils.Random.randrange", return_value=250000): transaction = Transaction.continue_from_headers({}, name="Head SDK tx") # will create empty mutable baggage diff --git a/tests/tracing/test_sample_rand.py b/tests/tracing/test_sample_rand.py index f9c10aa04e..4a74950b30 100644 --- a/tests/tracing/test_sample_rand.py +++ b/tests/tracing/test_sample_rand.py @@ -1,5 +1,3 @@ -import decimal -from decimal import Inexact, FloatOperation from unittest import mock import pytest @@ -20,7 +18,8 @@ def test_deterministic_sampled(sentry_init, capture_events, sample_rate, sample_ events = capture_events() with mock.patch( - "sentry_sdk.tracing_utils.Random.uniform", return_value=sample_rand + "sentry_sdk.tracing_utils.Random.randrange", + return_value=int(sample_rand * 1000000), ): with sentry_sdk.start_transaction() as transaction: assert ( @@ -55,35 +54,3 @@ def test_transaction_uses_incoming_sample_rand( # Transaction event captured if sample_rand < sample_rate, indicating that # sample_rand is used to make the sampling decision. assert len(events) == int(sample_rand < sample_rate) - - -def test_decimal_context(sentry_init, capture_events): - """ - Ensure that having a user altered decimal context with a precision below 6 - does not cause an InvalidOperation exception. - """ - sentry_init(traces_sample_rate=1.0) - events = capture_events() - - old_prec = decimal.getcontext().prec - old_inexact = decimal.getcontext().traps[Inexact] - old_float_operation = decimal.getcontext().traps[FloatOperation] - - decimal.getcontext().prec = 2 - decimal.getcontext().traps[Inexact] = True - decimal.getcontext().traps[FloatOperation] = True - - try: - with mock.patch( - "sentry_sdk.tracing_utils.Random.uniform", return_value=0.123456789 - ): - with sentry_sdk.start_transaction() as transaction: - assert ( - transaction.get_baggage().sentry_items["sample_rand"] == "0.123456" - ) - finally: - decimal.getcontext().prec = old_prec - decimal.getcontext().traps[Inexact] = old_inexact - decimal.getcontext().traps[FloatOperation] = old_float_operation - - assert len(events) == 1 diff --git a/tests/tracing/test_sample_rand_propagation.py b/tests/tracing/test_sample_rand_propagation.py index ea3ea548ff..e6f3e99510 100644 --- a/tests/tracing/test_sample_rand_propagation.py +++ b/tests/tracing/test_sample_rand_propagation.py @@ -35,9 +35,9 @@ def test_continue_trace_missing_sample_rand(): "baggage": "sentry-placeholder=asdf", } - mock_uniform = Mock(return_value=0.5) - - with mock.patch("sentry_sdk.tracing_utils.Random.uniform", mock_uniform): + with mock.patch( + "sentry_sdk.tracing_utils.Random.randrange", Mock(return_value=500000) + ): transaction = sentry_sdk.continue_trace(headers) assert transaction.get_baggage().sentry_items["sample_rand"] == "0.500000" From ca12bbf547559d21319ac676b2e754df712de0f7 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 12 Sep 2025 10:47:59 +0200 Subject: [PATCH 464/591] ci: Fix mypy, gevent (#4790) - mypy: looks like the error we're ignoring is now called differently - gevent: zope.interface pushed out a new major recently which broke old python tests --- scripts/populate_tox/tox.jinja | 1 + sentry_sdk/integrations/threading.py | 2 +- tox.ini | 1 + 3 files changed, 3 insertions(+), 1 deletion(-) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index ef2e89c88c..4a4bd96c52 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -134,6 +134,7 @@ deps = {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest gevent: pytest-asyncio {py3.10,py3.11}-gevent: zope.event<5.0.0 + {py3.10,py3.11}-gevent: zope.interface<8.0 # === Integrations === diff --git a/sentry_sdk/integrations/threading.py b/sentry_sdk/integrations/threading.py index fc4f539228..c031c51f50 100644 --- a/sentry_sdk/integrations/threading.py +++ b/sentry_sdk/integrations/threading.py @@ -52,7 +52,7 @@ def setup_once(): try: from django import VERSION as django_version # noqa: N811 - import channels # type: ignore[import-not-found] + import channels # type: ignore[import-untyped] channels_version = channels.__version__ except ImportError: diff --git a/tox.ini b/tox.ini index 1bc9757b9a..39ef4785b3 100644 --- a/tox.ini +++ b/tox.ini @@ -358,6 +358,7 @@ deps = {py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest gevent: pytest-asyncio {py3.10,py3.11}-gevent: zope.event<5.0.0 + {py3.10,py3.11}-gevent: zope.interface<8.0 # === Integrations === From b19e08642e093c2f7945849bc26fd0d5f735bed9 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 12 Sep 2025 10:58:01 +0200 Subject: [PATCH 465/591] Correctly check for a running transaction (#4791) Fix check for the existence of a currently transaction. --- sentry_sdk/ai/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index 41f89f5623..d0ccf1bed3 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -44,6 +44,6 @@ def get_start_span_function(): # type: () -> Callable[..., Any] current_span = sentry_sdk.get_current_span() transaction_exists = ( - current_span is not None and current_span.containing_transaction == current_span + current_span is not None and current_span.containing_transaction is not None ) return sentry_sdk.start_span if transaction_exists else sentry_sdk.start_transaction From 398b7c4f4486e52493116f9dc469fbeef14f6bc7 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 12 Sep 2025 11:31:44 +0200 Subject: [PATCH 466/591] feat(anthropic): Add proper tool calling data to Anthropic integration (#4769) - Format the response of the LLM (`gen_ai.response.text`) correctly. Not using the JSON but only use the actual text that was returned. - Add responses for tool calls (`gen_ai.response.tool_calls`) to the LLM spans. - Add results of tool calls to the request (`gen_ai.request.messages`). Before: Screenshot 2025-09-12 at 10 43 32 After: Screenshot 2025-09-12 at 10 45 11 --- sentry_sdk/integrations/anthropic.py | 55 +++++++++++++++---- .../integrations/anthropic/test_anthropic.py | 41 ++++++-------- 2 files changed, 61 insertions(+), 35 deletions(-) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index ff3d9a3388..4f4c0b1a2a 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -1,5 +1,4 @@ from functools import wraps -import json from typing import TYPE_CHECKING import sentry_sdk @@ -117,8 +116,29 @@ def _set_input_data(span, kwargs, integration): and should_send_default_pii() and integration.include_prompts ): + normalized_messages = [] + for message in messages: + if ( + message.get("role") == "user" + and "content" in message + and isinstance(message["content"], (list, tuple)) + ): + for item in message["content"]: + if item.get("type") == "tool_result": + normalized_messages.append( + { + "role": "tool", + "content": { + "tool_use_id": item.get("tool_use_id"), + "output": item.get("content"), + }, + } + ) + else: + normalized_messages.append(message) + set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, safe_serialize(messages) + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, normalized_messages, unpack=False ) set_data_normalized( @@ -159,12 +179,29 @@ def _set_output_data( Set output data for the span based on the AI response.""" span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, model) if should_send_default_pii() and integration.include_prompts: - set_data_normalized( - span, - SPANDATA.GEN_AI_RESPONSE_TEXT, - json.dumps(content_blocks), - unpack=False, - ) + output_messages = { + "response": [], + "tool": [], + } # type: (dict[str, list[Any]]) + + for output in content_blocks: + if output["type"] == "text": + output_messages["response"].append(output["text"]) + elif output["type"] == "tool_use": + output_messages["tool"].append(output) + + if len(output_messages["tool"]) > 0: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + output_messages["tool"], + unpack=False, + ) + + if len(output_messages["response"]) > 0: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, output_messages["response"] + ) record_token_usage( span, @@ -172,8 +209,6 @@ def _set_output_data( output_tokens=output_tokens, ) - # TODO: GEN_AI_RESPONSE_TOOL_CALLS ? - if finish_span: span.__exit__(None, None, None) diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index eba07a1df6..3893626026 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -1,6 +1,6 @@ +import pytest from unittest import mock - try: from unittest.mock import AsyncMock except ImportError: @@ -10,7 +10,6 @@ async def __call__(self, *args, **kwargs): return super(AsyncMock, self).__call__(*args, **kwargs) -import pytest from anthropic import Anthropic, AnthropicError, AsyncAnthropic, AsyncStream, Stream from anthropic.types import MessageDeltaUsage, TextDelta, Usage from anthropic.types.content_block_delta_event import ContentBlockDeltaEvent @@ -20,9 +19,6 @@ async def __call__(self, *args, **kwargs): from anthropic.types.message_delta_event import MessageDeltaEvent from anthropic.types.message_start_event import MessageStartEvent -from sentry_sdk.integrations.anthropic import _set_output_data, _collect_ai_data -from sentry_sdk.utils import package_version - try: from anthropic.types import InputJSONDelta except ImportError: @@ -46,9 +42,16 @@ async def __call__(self, *args, **kwargs): from sentry_sdk import start_transaction, start_span from sentry_sdk.consts import OP, SPANDATA -from sentry_sdk.integrations.anthropic import AnthropicIntegration +from sentry_sdk.integrations.anthropic import ( + AnthropicIntegration, + _set_output_data, + _collect_ai_data, +) +from sentry_sdk.utils import package_version + ANTHROPIC_VERSION = package_version("anthropic") + EXAMPLE_MESSAGE = Message( id="id", model="model", @@ -121,10 +124,7 @@ def test_nonstreaming_create_message( span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '[{"role": "user", "content": "Hello, Claude"}]' ) - assert ( - span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "Hi, I\'m Claude.", "type": "text"}]' - ) + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "Hi, I'm Claude." else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] @@ -193,10 +193,7 @@ async def test_nonstreaming_create_message_async( span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '[{"role": "user", "content": "Hello, Claude"}]' ) - assert ( - span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "Hi, I\'m Claude.", "type": "text"}]' - ) + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "Hi, I'm Claude." else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] @@ -296,10 +293,7 @@ def test_streaming_create_message( span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '[{"role": "user", "content": "Hello, Claude"}]' ) - assert ( - span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "Hi! I\'m Claude!", "type": "text"}]' - ) + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "Hi! I'm Claude!" else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] @@ -403,10 +397,7 @@ async def test_streaming_create_message_async( span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] == '[{"role": "user", "content": "Hello, Claude"}]' ) - assert ( - span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "Hi! I\'m Claude!", "type": "text"}]' - ) + assert span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] == "Hi! I'm Claude!" else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] @@ -539,7 +530,7 @@ def test_streaming_create_message_with_input_json_delta( ) assert ( span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "{\'location\': \'San Francisco, CA\'}", "type": "text"}]' + == "{'location': 'San Francisco, CA'}" ) else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] @@ -679,7 +670,7 @@ async def test_streaming_create_message_with_input_json_delta_async( ) assert ( span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] - == '[{"text": "{\'location\': \'San Francisco, CA\'}", "type": "text"}]' + == "{'location': 'San Francisco, CA'}" ) else: @@ -835,7 +826,7 @@ def test_set_output_data_with_input_json_delta(sentry_init): assert ( span._data.get(SPANDATA.GEN_AI_RESPONSE_TEXT) - == "[{\"text\": \"{'test': 'data','more': 'json'}\", \"type\": \"text\"}]" + == "{'test': 'data','more': 'json'}" ) assert span._data.get(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS) == 10 assert span._data.get(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS) == 20 From 5a122b56fc39b841cf01243a622733bf43133403 Mon Sep 17 00:00:00 2001 From: "ZhengYu, Xu" Date: Fri, 12 Sep 2025 17:34:40 +0800 Subject: [PATCH 467/591] chore: Reexport module `profiler` (#4535) The example provided by sentry causes pylance to report `"profiler" is not a known attribute of module "sentry_sdk"` image --------- Co-authored-by: Anton Pirker Co-authored-by: Anton Pirker --- sentry_sdk/__init__.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index a37b52ff4e..1939be0510 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -1,10 +1,10 @@ +from sentry_sdk import profiler from sentry_sdk.scope import Scope from sentry_sdk.transport import Transport, HttpTransport from sentry_sdk.client import Client from sentry_sdk.api import * # noqa - -from sentry_sdk.consts import VERSION # noqa +from sentry_sdk.consts import VERSION __all__ = [ # noqa "Hub", @@ -12,6 +12,7 @@ "Client", "Transport", "HttpTransport", + "VERSION", "integrations", # From sentry_sdk.api "init", @@ -47,6 +48,7 @@ "trace", "monitor", "logger", + "profiler", "start_session", "end_session", "set_transaction_name", From 16f2c3df628ef1b0e8ecdaac272ab6e94931eec1 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 15 Sep 2025 08:50:39 +0200 Subject: [PATCH 468/591] build(deps): bump actions/create-github-app-token from 2.1.1 to 2.1.4 (#4795) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/create-github-app-token](https://github.com/actions/create-github-app-token) from 2.1.1 to 2.1.4.
Release notes

Sourced from actions/create-github-app-token's releases.

v2.1.4

2.1.4 (2025-09-13)

Bug Fixes

  • deps: bump @​octokit/auth-app from 7.2.1 to 8.0.1 (#257) (bef1eaf)

v2.1.3

2.1.3 (2025-09-13)

Bug Fixes

  • deps: bump undici from 7.8.0 to 7.10.0 in the production-dependencies group (#254) (f3d5ec2)

v2.1.2

2.1.2 (2025-09-12)

Bug Fixes

  • deps: bump @​octokit/request from 9.2.3 to 10.0.2 (#256) (5d7307b)
Commits
  • 6701853 build(release): 2.1.4 [skip ci]
  • bef1eaf fix(deps): bump @​octokit/auth-app from 7.2.1 to 8.0.1 (#257)
  • 1526738 build(release): 2.1.3 [skip ci]
  • f3d5ec2 fix(deps): bump undici from 7.8.0 to 7.10.0 in the production-dependencies gr...
  • def152b build(release): 2.1.2 [skip ci]
  • 5d7307b fix(deps): bump @​octokit/request from 9.2.3 to 10.0.2 (#256)
  • 525760a build(deps): bump stefanzweifel/git-auto-commit-action from 5.2.0 to 6.0.1 (#...
  • 8ab05a8 Add beta branch support for releases (#282)
  • d00315e build(deps): bump actions/checkout from 4 to 5 (#279)
  • fcc6c28 build(deps-dev): bump dotenv from 16.5.0 to 17.2.1 (#269)
  • Additional commits viewable in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/create-github-app-token&package-manager=github_actions&previous-version=2.1.1&new-version=2.1.4)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index f5e952d0de..68aeebf2b7 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Get auth token id: token - uses: actions/create-github-app-token@a8d616148505b5069dccd32f177bb87d7f39123b # v2.1.1 + uses: actions/create-github-app-token@67018539274d69449ef7c02e8e71183d1719ab42 # v2.1.4 with: app-id: ${{ vars.SENTRY_RELEASE_BOT_CLIENT_ID }} private-key: ${{ secrets.SENTRY_RELEASE_BOT_PRIVATE_KEY }} From 5747863128a3bbb382d37ea94d104a7d9b358441 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 15 Sep 2025 09:10:23 +0200 Subject: [PATCH 469/591] feat(integrations): Support gql 4.0-style execute (#4779) gql 4.0 [changed](https://github.com/graphql-python/gql/pull/556/files) the signature of the `execute` function which we were patching. Instead of a `DocumentNode` it now gets a `GraphQLRequest` (which contains the `document` attribute with the `DocumentNode`). This means we need to update the way we're extracting additional data in the event processor. --- sentry_sdk/integrations/gql.py | 27 ++++++++++++++++++++++----- 1 file changed, 22 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/gql.py b/sentry_sdk/integrations/gql.py index 5f4436f5b2..8c378060b7 100644 --- a/sentry_sdk/integrations/gql.py +++ b/sentry_sdk/integrations/gql.py @@ -18,6 +18,13 @@ ) from gql.transport import Transport, AsyncTransport # type: ignore[import-not-found] from gql.transport.exceptions import TransportQueryError # type: ignore[import-not-found] + + try: + # gql 4.0+ + from gql import GraphQLRequest + except ImportError: + GraphQLRequest = None + except ImportError: raise DidNotEnable("gql is not installed") @@ -92,13 +99,13 @@ def _patch_execute(): real_execute = gql.Client.execute @ensure_integration_enabled(GQLIntegration, real_execute) - def sentry_patched_execute(self, document, *args, **kwargs): + def sentry_patched_execute(self, document_or_request, *args, **kwargs): # type: (gql.Client, DocumentNode, Any, Any) -> Any scope = sentry_sdk.get_isolation_scope() - scope.add_event_processor(_make_gql_event_processor(self, document)) + scope.add_event_processor(_make_gql_event_processor(self, document_or_request)) try: - return real_execute(self, document, *args, **kwargs) + return real_execute(self, document_or_request, *args, **kwargs) except TransportQueryError as e: event, hint = event_from_exception( e, @@ -112,8 +119,8 @@ def sentry_patched_execute(self, document, *args, **kwargs): gql.Client.execute = sentry_patched_execute -def _make_gql_event_processor(client, document): - # type: (gql.Client, DocumentNode) -> EventProcessor +def _make_gql_event_processor(client, document_or_request): + # type: (gql.Client, Union[DocumentNode, gql.GraphQLRequest]) -> EventProcessor def processor(event, hint): # type: (Event, dict[str, Any]) -> Event try: @@ -130,6 +137,16 @@ def processor(event, hint): ) if should_send_default_pii(): + if GraphQLRequest is not None and isinstance( + document_or_request, GraphQLRequest + ): + # In v4.0.0, gql moved to using GraphQLRequest instead of + # DocumentNode in execute + # https://github.com/graphql-python/gql/pull/556 + document = document_or_request.document + else: + document = document_or_request + request["data"] = _data_from_document(document) contexts = event.setdefault("contexts", {}) response = contexts.setdefault("response", {}) From 0df7f4508ceb45d146143f2ff95d37c0c54e7b74 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 15 Sep 2025 14:34:43 +0200 Subject: [PATCH 470/591] fix(logs): Expect `log_item` as rate limit category (#4798) The data category for rate limiting logs is `log_item`, not `log`. Closes https://github.com/getsentry/sentry-python/issues/4797 --- sentry_sdk/_types.py | 2 +- sentry_sdk/envelope.py | 2 +- tests/test_envelope.py | 23 ++++++++++++++++++++++- tests/test_transport.py | 39 ++++++++++++++++++++++++++++++++++++++- 4 files changed, 62 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index 8336617a8d..b28c7260ce 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -269,7 +269,7 @@ class SDKInfo(TypedDict): "metric_bucket", "monitor", "span", - "log", + "log_item", ] SessionStatus = Literal["ok", "exited", "crashed", "abnormal"] diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index 5f7220bf21..7dbbdec5c8 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -273,7 +273,7 @@ def data_category(self): elif ty == "event": return "error" elif ty == "log": - return "log" + return "log_item" elif ty == "client_report": return "internal" elif ty == "profile": diff --git a/tests/test_envelope.py b/tests/test_envelope.py index d1bc668f05..06f8971dc3 100644 --- a/tests/test_envelope.py +++ b/tests/test_envelope.py @@ -1,4 +1,4 @@ -from sentry_sdk.envelope import Envelope +from sentry_sdk.envelope import Envelope, Item, PayloadRef from sentry_sdk.session import Session from sentry_sdk import capture_event import sentry_sdk.client @@ -239,3 +239,24 @@ def test_envelope_without_headers(): assert len(items) == 1 assert items[0].payload.get_bytes() == b'{"started": "2020-02-07T14:16:00Z"}' + + +def test_envelope_item_data_category_mapping(): + """Test that envelope items map to correct data categories for rate limiting.""" + test_cases = [ + ("event", "error"), + ("transaction", "transaction"), + ("log", "log_item"), + ("session", "session"), + ("attachment", "attachment"), + ("client_report", "internal"), + ("profile", "profile"), + ("profile_chunk", "profile_chunk"), + ("statsd", "metric_bucket"), + ("check_in", "monitor"), + ("unknown_type", "default"), + ] + + for item_type, expected_category in test_cases: + item = Item(payload=PayloadRef(json={"test": "data"}), type=item_type) + assert item.data_category == expected_category diff --git a/tests/test_transport.py b/tests/test_transport.py index c6a1a0a7a7..e493515e9a 100644 --- a/tests/test_transport.py +++ b/tests/test_transport.py @@ -611,7 +611,7 @@ def test_metric_bucket_limits(capturing_server, response_code, make_client): assert capturing_server.captured[0].path == "/api/132/envelope/" capturing_server.clear_captured() - assert set(client.transport._disabled_until) == set(["metric_bucket"]) + assert set(client.transport._disabled_until) == {"metric_bucket"} client.transport.capture_envelope(envelope) client.capture_event({"type": "transaction"}) @@ -629,6 +629,43 @@ def test_metric_bucket_limits(capturing_server, response_code, make_client): ] +@pytest.mark.parametrize("response_code", [200, 429]) +def test_log_item_limits(capturing_server, response_code, make_client): + client = make_client() + capturing_server.respond_with( + code=response_code, + headers={ + "X-Sentry-Rate-Limits": "4711:log_item:organization:quota_exceeded:custom" + }, + ) + + envelope = Envelope() + envelope.add_item(Item(payload=b"{}", type="log")) + client.transport.capture_envelope(envelope) + client.flush() + + assert len(capturing_server.captured) == 1 + assert capturing_server.captured[0].path == "/api/132/envelope/" + capturing_server.clear_captured() + + assert set(client.transport._disabled_until) == {"log_item"} + + client.transport.capture_envelope(envelope) + client.capture_event({"type": "transaction"}) + client.flush() + + assert len(capturing_server.captured) == 2 + + envelope = capturing_server.captured[0].envelope + assert envelope.items[0].type == "transaction" + envelope = capturing_server.captured[1].envelope + assert envelope.items[0].type == "client_report" + report = parse_json(envelope.items[0].get_bytes()) + assert report["discarded_events"] == [ + {"category": "log_item", "reason": "ratelimit_backoff", "quantity": 1}, + ] + + @pytest.mark.parametrize("response_code", [200, 429]) def test_metric_bucket_limits_with_namespace( capturing_server, response_code, make_client From 36ae7c4386a900874e8d5423fda1793ebaaf0e73 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 15 Sep 2025 15:26:01 +0200 Subject: [PATCH 471/591] tests: Update tox.ini (#4799) Regular update --- tox.ini | 49 ++++++++++++++++++++++++++----------------------- 1 file changed, 26 insertions(+), 23 deletions(-) diff --git a/tox.ini b/tox.ini index 39ef4785b3..5fe52a1e2b 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-09T08:24:12.875177+00:00 +# Last generated: 2025-09-15T12:28:26.599446+00:00 [tox] requires = @@ -98,12 +98,12 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.33.1 {py3.8,py3.11,py3.12}-anthropic-v0.50.0 - {py3.8,py3.12,py3.13}-anthropic-v0.66.0 + {py3.8,py3.12,py3.13}-anthropic-v0.67.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 {py3.9,py3.11,py3.12}-cohere-v5.13.12 - {py3.9,py3.11,py3.12}-cohere-v5.17.0 + {py3.9,py3.11,py3.12}-cohere-v5.18.0 {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.11,py3.12}-langchain-base-v0.2.17 @@ -116,12 +116,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.36.1 {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.107.0 + {py3.8,py3.12,py3.13}-openai-base-v1.107.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.2 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -129,6 +129,7 @@ envlist = {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 + {py3.10,py3.12,py3.13}-openai_agents-v0.3.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.27.1 @@ -141,7 +142,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.26 + {py3.9,py3.12,py3.13}-boto3-v1.40.30 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -160,7 +161,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.14.1 + {py3.9,py3.12,py3.13}-pymongo-v4.15.0 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -183,7 +184,7 @@ envlist = {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.58.4 {py3.7,py3.12,py3.13}-statsig-v0.61.0 - {py3.7,py3.12,py3.13}-statsig-v0.63.0 + {py3.7,py3.12,py3.13}-statsig-v0.64.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.1.0 @@ -233,6 +234,7 @@ envlist = {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.12,py3.13}-celery-v5.5.3 + {py3.8,py3.12,py3.13}-celery-v5.6.0b1 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 @@ -263,9 +265,9 @@ envlist = {py3.9,py3.12,py3.13}-flask-v3.1.2 {py3.6,py3.9,py3.10}-starlette-v0.16.0 - {py3.7,py3.10,py3.11}-starlette-v0.26.1 - {py3.8,py3.11,py3.12}-starlette-v0.36.3 - {py3.9,py3.12,py3.13}-starlette-v0.47.3 + {py3.7,py3.10,py3.11}-starlette-v0.27.0 + {py3.8,py3.12,py3.13}-starlette-v0.38.6 + {py3.9,py3.12,py3.13}-starlette-v0.48.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 @@ -460,7 +462,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.33.1: anthropic==0.33.1 anthropic-v0.50.0: anthropic==0.50.0 - anthropic-v0.66.0: anthropic==0.66.0 + anthropic-v0.67.0: anthropic==0.67.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.33.1: httpx<0.28.0 @@ -468,7 +470,7 @@ deps = cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 cohere-v5.13.12: cohere==5.13.12 - cohere-v5.17.0: cohere==5.17.0 + cohere-v5.18.0: cohere==5.18.0 langchain-base-v0.1.20: langchain==0.1.20 langchain-base-v0.2.17: langchain==0.2.17 @@ -488,7 +490,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.36.1: openai==1.36.1 openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.107.0: openai==1.107.0 + openai-base-v1.107.2: openai==1.107.2 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -497,7 +499,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.36.1: openai==1.36.1 openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.107.0: openai==1.107.0 + openai-notiktoken-v1.107.2: openai==1.107.2 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 @@ -508,6 +510,7 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 + openai_agents-v0.3.0: openai-agents==0.3.0 openai_agents: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 @@ -522,7 +525,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.26: boto3==1.40.26 + boto3-v1.40.30: boto3==1.40.30 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -544,7 +547,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.14.1: pymongo==4.14.1 + pymongo-v4.15.0: pymongo==4.15.0 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -568,7 +571,7 @@ deps = statsig-v0.55.3: statsig==0.55.3 statsig-v0.58.4: statsig==0.58.4 statsig-v0.61.0: statsig==0.61.0 - statsig-v0.63.0: statsig==0.63.0 + statsig-v0.64.0: statsig==0.64.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -639,6 +642,7 @@ deps = celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 celery-v5.5.3: celery==5.5.3 + celery-v5.6.0b1: celery==5.6.0b1 celery: newrelic<10.17.0 celery: redis {py3.7}-celery: importlib-metadata<5.0 @@ -698,9 +702,9 @@ deps = flask-v1.1.4: markupsafe<2.1.0 starlette-v0.16.0: starlette==0.16.0 - starlette-v0.26.1: starlette==0.26.1 - starlette-v0.36.3: starlette==0.36.3 - starlette-v0.47.3: starlette==0.47.3 + starlette-v0.27.0: starlette==0.27.0 + starlette-v0.38.6: starlette==0.38.6 + starlette-v0.48.0: starlette==0.48.0 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -708,8 +712,7 @@ deps = starlette: jinja2 starlette: httpx starlette-v0.16.0: httpx<0.28.0 - starlette-v0.26.1: httpx<0.28.0 - starlette-v0.36.3: httpx<0.28.0 + starlette-v0.27.0: httpx<0.28.0 {py3.6}-starlette: aiocontextvars fastapi-v0.79.1: fastapi==0.79.1 From 7ecb39b486231788ba3f18b547d7cb3ded25952e Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 15 Sep 2025 13:35:20 +0000 Subject: [PATCH 472/591] release: 2.38.0 --- CHANGELOG.md | 22 ++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 25 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 28c4882414..6ceda55626 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,27 @@ # Changelog +## 2.38.0 + +### Various fixes & improvements + +- tests: Update tox.ini (#4799) by @sentrivana +- fix(logs): Expect `log_item` as rate limit category (#4798) by @sentrivana +- feat(integrations): Support gql 4.0-style execute (#4779) by @sentrivana +- build(deps): bump actions/create-github-app-token from 2.1.1 to 2.1.4 (#4795) by @dependabot +- chore: Reexport module `profiler` (#4535) by @zen-xu +- feat(anthropic): Add proper tool calling data to Anthropic integration (#4769) by @antonpirker +- Correctly check for a running transaction (#4791) by @antonpirker +- ci: Fix mypy, gevent (#4790) by @sentrivana +- ref(tracing): Use float for sample rand (#4677) by @sentrivana +- Avoid reporting false-positive StopAsyncIteration in the asyncio integration (#4741) by @vmarkovtsev +- feat(ai): Create transaction in AI agents framworks, when no transaction is running. (#4758) by @constantinius +- build(deps): bump codecov/codecov-action from 5.5.0 to 5.5.1 (#4773) by @dependabot +- build(deps): bump actions/setup-python from 5 to 6 (#4774) by @dependabot +- Add log message when `DedupeIntegration` is dropping an error. (#4788) by @antonpirker +- Add input and output to `invoke_agent` span. (#4785) by @antonpirker +- Update HuggingFace Hub integration (#4746) by @antonpirker +- fix(profiling): Re-init continuous profiler (#4772) by @Zylphrex + ## 2.37.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 28a49b7fa7..061b2bdfc8 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.37.1" +release = "2.38.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index cc3c9b1612..91a1740526 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1331,4 +1331,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.37.1" +VERSION = "2.38.0" diff --git a/setup.py b/setup.py index 1b4d0063e4..58101aa65f 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.37.1", + version="2.38.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 2e37b516aa626984a9b94dc15d9c5ff3459cefb2 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 15 Sep 2025 15:43:47 +0200 Subject: [PATCH 473/591] Updated Changelog --- CHANGELOG.md | 34 +++++++++++++++++----------------- sentry_sdk/tracing.py | 15 ++++++++++++--- 2 files changed, 29 insertions(+), 20 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 6ceda55626..7abbed7218 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,23 +4,23 @@ ### Various fixes & improvements -- tests: Update tox.ini (#4799) by @sentrivana -- fix(logs): Expect `log_item` as rate limit category (#4798) by @sentrivana -- feat(integrations): Support gql 4.0-style execute (#4779) by @sentrivana -- build(deps): bump actions/create-github-app-token from 2.1.1 to 2.1.4 (#4795) by @dependabot -- chore: Reexport module `profiler` (#4535) by @zen-xu -- feat(anthropic): Add proper tool calling data to Anthropic integration (#4769) by @antonpirker -- Correctly check for a running transaction (#4791) by @antonpirker -- ci: Fix mypy, gevent (#4790) by @sentrivana -- ref(tracing): Use float for sample rand (#4677) by @sentrivana -- Avoid reporting false-positive StopAsyncIteration in the asyncio integration (#4741) by @vmarkovtsev -- feat(ai): Create transaction in AI agents framworks, when no transaction is running. (#4758) by @constantinius -- build(deps): bump codecov/codecov-action from 5.5.0 to 5.5.1 (#4773) by @dependabot -- build(deps): bump actions/setup-python from 5 to 6 (#4774) by @dependabot -- Add log message when `DedupeIntegration` is dropping an error. (#4788) by @antonpirker -- Add input and output to `invoke_agent` span. (#4785) by @antonpirker -- Update HuggingFace Hub integration (#4746) by @antonpirker -- fix(profiling): Re-init continuous profiler (#4772) by @Zylphrex +- Feat(huggingface_hub): Update HuggingFace Hub integration (#4746) by @antonpirker +- Feat(Anthropic): Add proper tool calling data to Anthropic integration (#4769) by @antonpirker +- Feat(openai-agents): Add input and output to `invoke_agent` span. (#4785) by @antonpirker +- Feat(AI): Create transaction in AI agents framworks, when no transaction is running. (#4758) by @constantinius +- Feat(GraphQL): Support gql 4.0-style execute (#4779) by @sentrivana +- Fix(logs): Expect `log_item` as rate limit category (#4798) by @sentrivana +- Fix: CI for mypy, gevent (#4790) by @sentrivana +- Fix: Correctly check for a running transaction (#4791) by @antonpirker +- Fix: Use float for sample rand (#4677) by @sentrivana +- Fix: Avoid reporting false-positive StopAsyncIteration in the asyncio integration (#4741) by @vmarkovtsev +- Fix: Add log message when `DedupeIntegration` is dropping an error. (#4788) by @antonpirker +- Fix(profiling): Re-init continuous profiler (#4772) by @Zylphrex +- Chore: Reexport module `profiler` (#4535) by @zen-xu +- Tests: Update tox.ini (#4799) by @sentrivana +- Build(deps): bump actions/create-github-app-token from 2.1.1 to 2.1.4 (#4795) by @dependabot +- Build(deps): bump actions/setup-python from 5 to 6 (#4774) by @dependabot +- Build(deps): bump codecov/codecov-action from 5.5.0 to 5.5.1 (#4773) by @dependabot ## 2.37.1 diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index fc43a33dc7..76f91eebc5 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -708,10 +708,19 @@ def finish(self, scope=None, end_timestamp=None): end_timestamp = datetime.fromtimestamp(end_timestamp, timezone.utc) self.timestamp = end_timestamp else: - elapsed = nanosecond_time() - self._start_timestamp_monotonic_ns - self.timestamp = self.start_timestamp + timedelta( - microseconds=elapsed / 1000 + print(f"In Span.finish for span {self}") + now = nanosecond_time() + elapsed = now - self._start_timestamp_monotonic_ns + print(f"NOW before assigment: {now}") + self.timestamp = ( + self.start_timestamp + + timedelta( # The assignment in this line is taking a lot of time + microseconds=elapsed / 1000 + ) ) + now2 = nanosecond_time() + print(f"NOW2 after assigment: {now2}") + print(f"Assignment duration: {now2 - now}") except AttributeError: self.timestamp = datetime.now(timezone.utc) From d94652a5527cd0e7810266f8cd30d9780e099a46 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Mon, 15 Sep 2025 16:43:25 +0200 Subject: [PATCH 474/591] removed accidental stuff --- sentry_sdk/tracing.py | 15 +++------------ 1 file changed, 3 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 76f91eebc5..fc43a33dc7 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -708,19 +708,10 @@ def finish(self, scope=None, end_timestamp=None): end_timestamp = datetime.fromtimestamp(end_timestamp, timezone.utc) self.timestamp = end_timestamp else: - print(f"In Span.finish for span {self}") - now = nanosecond_time() - elapsed = now - self._start_timestamp_monotonic_ns - print(f"NOW before assigment: {now}") - self.timestamp = ( - self.start_timestamp - + timedelta( # The assignment in this line is taking a lot of time - microseconds=elapsed / 1000 - ) + elapsed = nanosecond_time() - self._start_timestamp_monotonic_ns + self.timestamp = self.start_timestamp + timedelta( + microseconds=elapsed / 1000 ) - now2 = nanosecond_time() - print(f"NOW2 after assigment: {now2}") - print(f"Assignment duration: {now2 - now}") except AttributeError: self.timestamp = datetime.now(timezone.utc) From 83e8d798af0d22626e3104e2c02cb125a3dbde10 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Johannes=20Daxb=C3=B6ck?= <114897+dingsdax@users.noreply.github.com> Date: Tue, 16 Sep 2025 11:20:42 +0200 Subject: [PATCH 475/591] meta: update pull request template (#4803) * resolves: #4796 * resolves: PY-1841 --- .github/PULL_REQUEST_TEMPLATE.md | 17 ++++++++++++----- 1 file changed, 12 insertions(+), 5 deletions(-) diff --git a/.github/PULL_REQUEST_TEMPLATE.md b/.github/PULL_REQUEST_TEMPLATE.md index 12db62315a..dd7ef45832 100644 --- a/.github/PULL_REQUEST_TEMPLATE.md +++ b/.github/PULL_REQUEST_TEMPLATE.md @@ -1,7 +1,14 @@ - +### Description + ---- +#### Issues + -Thank you for contributing to `sentry-python`! Please add tests to validate your changes, and lint your code using `tox -e linters`. - -Running the test suite on your PR might require maintainer approval. \ No newline at end of file +#### Reminders +- Please add tests to validate your changes, and lint your code using `tox -e linters`. +- Add GH Issue ID _&_ Linear ID (if applicable) +- PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) +- For external contributors: [CONTRIBUTING.md](../CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) From 7a905085087d5d11285037d609f6a91cec04b138 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 16 Sep 2025 11:24:56 +0200 Subject: [PATCH 476/591] tests: Another tox.ini update (#4801) More new releases. An initial RC of huggingface_hub v1.0 is also available and breaks our integration. Ignoring it in the tests for now. Follow up here: https://github.com/getsentry/sentry-python/issues/4802 --- scripts/populate_tox/config.py | 1 + tox.ini | 28 +++++++++++++++------------- 2 files changed, 16 insertions(+), 13 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index bc20d531b3..8e35eaba4f 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -158,6 +158,7 @@ "deps": { "*": ["responses"], }, + "include": "<1.0", }, "langchain-base": { "package": "langchain", diff --git a/tox.ini b/tox.ini index 5fe52a1e2b..1fb8e9be40 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-15T12:28:26.599446+00:00 +# Last generated: 2025-09-16T07:18:41.732958+00:00 [tox] requires = @@ -116,12 +116,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.36.1 {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.107.2 + {py3.8,py3.12,py3.13}-openai-base-v1.107.3 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.2 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.3 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -134,7 +134,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.27.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.4 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.5 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 @@ -142,7 +142,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.30 + {py3.9,py3.12,py3.13}-boto3-v1.40.31 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -230,6 +230,7 @@ envlist = {py3.7,py3.8}-beam-v2.32.0 {py3.8,py3.10,py3.11}-beam-v2.50.0 {py3.9,py3.12,py3.13}-beam-v2.67.0 + {py3.9,py3.12,py3.13}-beam-v2.68.0rc2 {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.6,py3.7,py3.8}-celery-v5.0.5 @@ -322,8 +323,8 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.35 - {py3.9,py3.12,py3.13}-trytond-v7.6.6 + {py3.8,py3.11,py3.12}-trytond-v7.0.36 + {py3.9,py3.12,py3.13}-trytond-v7.6.7 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 @@ -490,7 +491,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.36.1: openai==1.36.1 openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.107.2: openai==1.107.2 + openai-base-v1.107.3: openai==1.107.3 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -499,7 +500,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.36.1: openai==1.36.1 openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.107.2: openai==1.107.2 + openai-notiktoken-v1.107.3: openai==1.107.3 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.36.1: httpx<0.28 @@ -516,7 +517,7 @@ deps = huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.27.1: huggingface_hub==0.27.1 huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.34.4: huggingface_hub==0.34.4 + huggingface_hub-v0.34.5: huggingface_hub==0.34.5 huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 huggingface_hub: responses @@ -525,7 +526,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.30: boto3==1.40.30 + boto3-v1.40.31: boto3==1.40.31 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -638,6 +639,7 @@ deps = beam-v2.32.0: apache-beam==2.32.0 beam-v2.50.0: apache-beam==2.50.0 beam-v2.67.0: apache-beam==2.67.0 + beam-v2.68.0rc2: apache-beam==2.68.0rc2 celery-v4.4.7: celery==4.4.7 celery-v5.0.5: celery==5.0.5 @@ -815,8 +817,8 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.35: trytond==7.0.35 - trytond-v7.6.6: trytond==7.6.6 + trytond-v7.0.36: trytond==7.0.36 + trytond-v7.6.7: trytond==7.6.7 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 875ed893cc2aeb61570123e556ac79fd1bed1ced Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 17 Sep 2025 09:17:36 +0200 Subject: [PATCH 477/591] feat: Add action to run toxgen periodically (#4805) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit ### Description Run `scripts/generate-test-files.sh` in a GitHub action every Monday morning and create a PR with the changes. If there are already any toxgen PRs open, close them. Here's an example run of the action: https://github.com/sentrivana/sentry-python/actions/runs/17768312844 (from a fork of the repo so that I can test it) And an example PR: https://github.com/sentrivana/sentry-python/pull/6 Action shamelessly stolen and adapted from sentry-cli. 🥔 #### Issues Closes https://github.com/getsentry/sentry-python/issues/4050 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](../CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --------- Co-authored-by: Daniel Szoke <7881302+szokeasaurusrex@users.noreply.github.com> --- .github/workflows/update-tox.yml | 111 +++++++++++++++++++++++++++++++ 1 file changed, 111 insertions(+) create mode 100644 .github/workflows/update-tox.yml diff --git a/.github/workflows/update-tox.yml b/.github/workflows/update-tox.yml new file mode 100644 index 0000000000..cfe98bbfe0 --- /dev/null +++ b/.github/workflows/update-tox.yml @@ -0,0 +1,111 @@ +name: Update test matrix + +on: + workflow_dispatch: + schedule: + # early Monday morning + - cron: '23 3 * * 1' + +jobs: + update-tox: + name: Update test matrix + runs-on: ubuntu-latest + timeout-minutes: 10 + + permissions: + contents: write + pull-requests: write + + steps: + - name: Setup Python + uses: actions/setup-python@v5 + with: + python-version: 3.13 + + - name: Checkout repo + uses: actions/checkout@v5.0.0 + with: + token: ${{ secrets.GITHUB_TOKEN }} + + - name: Configure git + run: | + git config user.name 'github-actions[bot]' + git config user.email '41898282+github-actions[bot]@users.noreply.github.com' + + - name: Run generate-test-files.sh + run: | + set -e + sh scripts/generate-test-files.sh + + - name: Create branch + id: create-branch + run: | + COMMIT_TITLE="ci: 🤖 Update test matrix with new releases" + DATE=`date +%m/%d` + BRANCH_NAME="toxgen/update" + + git checkout -B "$BRANCH_NAME" + git add --all + git commit -m "$COMMIT_TITLE" + git push origin "$BRANCH_NAME" --force + + echo "branch_name=$BRANCH_NAME" >> $GITHUB_OUTPUT + echo "commit_title=$COMMIT_TITLE" >> $GITHUB_OUTPUT + echo "date=$DATE" >> $GITHUB_OUTPUT + + - name: Create pull request + uses: actions/github-script@v8.0.0 + with: + script: | + const branchName = '${{ steps.create-branch.outputs.branch_name }}'; + const commitTitle = '${{ steps.create-branch.outputs.commit_title }}'; + const date = '${{ steps.create-branch.outputs.date }}'; + const prBody = `Update our test matrix with new releases of integrated frameworks and libraries. + + ## How it works + - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. + - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. + - Update [tox.ini](tox.ini) with the new releases. + + ## Action required + - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. + - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. + - Check what the failures look like and either fix them, or update the [test config](scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](scripts/generate-test-files.sh). See [README.md](scripts/populate_tox/README.md) for what configuration options are available. + + _____________________ + + _🤖 This PR was automatically created using [a GitHub action](.github/workflows/update-tox.yml)._`.replace(/^ {16}/gm, '') + + // Close existing toxgen PRs as they're now obsolete + + const { data: existingPRs } = await github.rest.pulls.list({ + owner: context.repo.owner, + repo: context.repo.repo, + head: `${context.repo.owner}:${branchName}`, + state: 'open' + }); + + for (const pr of existingPRs) { + await github.rest.pulls.update({ + owner: context.repo.owner, + repo: context.repo.repo, + pull_number: pr.number, + state: 'closed' + }) + }; + + const { data: pr } = await github.rest.pulls.create({ + owner: context.repo.owner, + repo: context.repo.repo, + title: commitTitle + ' (' + date + ')', + head: branchName, + base: '${{ github.ref_name }}', + body: prBody, + }); + + await github.rest.issues.addLabels({ + owner: context.repo.owner, + repo: context.repo.repo, + issue_number: pr.number, + labels: ['Component: CI', 'Component: Tests'] + }); From bc990db9f02be3d9446bd5ed80587966bba9c2fb Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Wed, 17 Sep 2025 11:11:05 +0200 Subject: [PATCH 478/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(09/17)=20(#4806)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](scripts/generate-test-files.sh). See [README.md](scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- tox.ini | 42 +++++++++++++++++++----------------------- 1 file changed, 19 insertions(+), 23 deletions(-) diff --git a/tox.ini b/tox.ini index 1fb8e9be40..50ac22e886 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-16T07:18:41.732958+00:00 +# Last generated: 2025-09-17T07:20:17.058541+00:00 [tox] requires = @@ -132,17 +132,16 @@ envlist = {py3.10,py3.12,py3.13}-openai_agents-v0.3.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.27.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.30.2 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.34.5 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0rc0 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.31 + {py3.9,py3.12,py3.13}-boto3-v1.40.32 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -161,7 +160,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.6,py3.9,py3.10}-pymongo-v4.0.2 - {py3.9,py3.12,py3.13}-pymongo-v4.15.0 + {py3.9,py3.12,py3.13}-pymongo-v4.15.1 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 @@ -214,10 +213,9 @@ envlist = # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 - {py3.7,py3.9,py3.10}-grpc-v1.46.5 - {py3.7,py3.11,py3.12}-grpc-v1.60.2 - {py3.9,py3.12,py3.13}-grpc-v1.74.0 - {py3.9,py3.12,py3.13}-grpc-v1.75.0rc1 + {py3.7,py3.9,py3.10}-grpc-v1.47.5 + {py3.7,py3.11,py3.12}-grpc-v1.62.3 + {py3.9,py3.12,py3.13}-grpc-v1.75.0 # ~~~ Tasks ~~~ @@ -273,7 +271,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.91.0 {py3.7,py3.10,py3.11}-fastapi-v0.103.2 - {py3.8,py3.12,py3.13}-fastapi-v0.116.1 + {py3.8,py3.12,py3.13}-fastapi-v0.116.2 # ~~~ Web 2 ~~~ @@ -515,10 +513,9 @@ deps = openai_agents: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 - huggingface_hub-v0.27.1: huggingface_hub==0.27.1 - huggingface_hub-v0.30.2: huggingface_hub==0.30.2 - huggingface_hub-v0.34.5: huggingface_hub==0.34.5 - huggingface_hub-v0.35.0rc0: huggingface_hub==0.35.0rc0 + huggingface_hub-v0.28.1: huggingface_hub==0.28.1 + huggingface_hub-v0.32.6: huggingface_hub==0.32.6 + huggingface_hub-v0.35.0: huggingface_hub==0.35.0 huggingface_hub: responses @@ -526,7 +523,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.31: boto3==1.40.31 + boto3-v1.40.32: boto3==1.40.32 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -548,7 +545,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 pymongo-v4.0.2: pymongo==4.0.2 - pymongo-v4.15.0: pymongo==4.15.0 + pymongo-v4.15.1: pymongo==4.15.1 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 @@ -615,10 +612,9 @@ deps = # ~~~ Network ~~~ grpc-v1.32.0: grpcio==1.32.0 - grpc-v1.46.5: grpcio==1.46.5 - grpc-v1.60.2: grpcio==1.60.2 - grpc-v1.74.0: grpcio==1.74.0 - grpc-v1.75.0rc1: grpcio==1.75.0rc1 + grpc-v1.47.5: grpcio==1.47.5 + grpc-v1.62.3: grpcio==1.62.3 + grpc-v1.75.0: grpcio==1.75.0 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -720,7 +716,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.91.0: fastapi==0.91.0 fastapi-v0.103.2: fastapi==0.103.2 - fastapi-v0.116.1: fastapi==0.116.1 + fastapi-v0.116.2: fastapi==0.116.2 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart From 62ff9ac68672dacaa9fe94df54de1999203d3e5e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 17 Sep 2025 12:13:27 +0200 Subject: [PATCH 479/591] fix: Fix link to CONTRIBUTING.md in PR template (#4808) ### Description The link to CONTRIBUTING.md in the PR template doesn't work (see below), updating it. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](../CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/PULL_REQUEST_TEMPLATE.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/PULL_REQUEST_TEMPLATE.md b/.github/PULL_REQUEST_TEMPLATE.md index dd7ef45832..79f27c30d8 100644 --- a/.github/PULL_REQUEST_TEMPLATE.md +++ b/.github/PULL_REQUEST_TEMPLATE.md @@ -11,4 +11,4 @@ - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) -- For external contributors: [CONTRIBUTING.md](../CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) +- For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) From 6c2a99601a75227f39f36e636d3b3bc24017ae8e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 17 Sep 2025 12:13:37 +0200 Subject: [PATCH 480/591] fix: Fix links to files in toxgen PRs (#4807) ### Description The toxgen PR template contains links to files in the repo, but they were broken, fixing that here. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](../CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/update-tox.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/update-tox.yml b/.github/workflows/update-tox.yml index cfe98bbfe0..11d143a2a9 100644 --- a/.github/workflows/update-tox.yml +++ b/.github/workflows/update-tox.yml @@ -65,16 +65,16 @@ jobs: ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - - Update [tox.ini](tox.ini) with the new releases. + - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - - Check what the failures look like and either fix them, or update the [test config](scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](scripts/generate-test-files.sh). See [README.md](scripts/populate_tox/README.md) for what configuration options are available. + - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ - _🤖 This PR was automatically created using [a GitHub action](.github/workflows/update-tox.yml)._`.replace(/^ {16}/gm, '') + _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._`.replace(/^ {16}/gm, '') // Close existing toxgen PRs as they're now obsolete From 43293832af8fd30f3fd18ef9206ba52c78fc5a52 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 17 Sep 2025 12:42:32 +0200 Subject: [PATCH 481/591] fix(Django): Avoid evaluating complex Django object in span.data/span.attributes (#4804) ### Description When rendering templates of the Django Admin in Django 5.0+ the template context includes also a `QuerySet` with lots of log items. In our `serialize()` function this was not properly serialized leading to evaluating the `QuerySet` (read: running the SQL query). This change updates the `serialize()` to also correctly serialize `QuerySet`s in `span.data` #### Issues * resolves: #4604 * resolves: PY-1781 --- sentry_sdk/serializer.py | 13 ++++++- tests/integrations/django/myapp/urls.py | 1 + tests/integrations/django/myapp/views.py | 23 +++++++++++++ tests/integrations/django/test_basic.py | 43 ++++++++++++++++++++++-- 4 files changed, 77 insertions(+), 3 deletions(-) diff --git a/sentry_sdk/serializer.py b/sentry_sdk/serializer.py index 04df9857bd..6bde5c08bd 100644 --- a/sentry_sdk/serializer.py +++ b/sentry_sdk/serializer.py @@ -187,6 +187,16 @@ def _is_databag(): return False + def _is_span_attribute(): + # type: () -> Optional[bool] + try: + if path[0] == "spans" and path[2] == "data": + return True + except IndexError: + return None + + return False + def _is_request_body(): # type: () -> Optional[bool] try: @@ -282,7 +292,8 @@ def _serialize_node_impl( ) return None - if is_databag and global_repr_processors: + is_span_attribute = _is_span_attribute() + if (is_databag or is_span_attribute) and global_repr_processors: hints = {"memo": memo, "remaining_depth": remaining_depth} for processor in global_repr_processors: result = processor(obj, hints) diff --git a/tests/integrations/django/myapp/urls.py b/tests/integrations/django/myapp/urls.py index 79dd4edd52..fbc9e6032e 100644 --- a/tests/integrations/django/myapp/urls.py +++ b/tests/integrations/django/myapp/urls.py @@ -58,6 +58,7 @@ def path(path, *args, **kwargs): path("template-test", views.template_test, name="template_test"), path("template-test2", views.template_test2, name="template_test2"), path("template-test3", views.template_test3, name="template_test3"), + path("template-test4", views.template_test4, name="template_test4"), path("postgres-select", views.postgres_select, name="postgres_select"), path("postgres-select-slow", views.postgres_select_orm, name="postgres_select_orm"), path( diff --git a/tests/integrations/django/myapp/views.py b/tests/integrations/django/myapp/views.py index 5e8cc39053..9c14bc27d7 100644 --- a/tests/integrations/django/myapp/views.py +++ b/tests/integrations/django/myapp/views.py @@ -208,6 +208,29 @@ def template_test3(request, *args, **kwargs): return render(request, "trace_meta.html", {}) +@csrf_exempt +def template_test4(request, *args, **kwargs): + User.objects.create_user("john", "lennon@thebeatles.com", "johnpassword") + my_queryset = User.objects.all() # noqa + + template_context = { + "user_age": 25, + "complex_context": my_queryset, + "complex_list": [1, 2, 3, my_queryset], + "complex_dict": { + "a": 1, + "d": my_queryset, + }, + "none_context": None, + } + + return TemplateResponse( + request, + "user_name.html", + template_context, + ) + + @csrf_exempt def postgres_select(request, *args, **kwargs): from django.db import connections diff --git a/tests/integrations/django/test_basic.py b/tests/integrations/django/test_basic.py index e96cd09e4f..bbe29c7238 100644 --- a/tests/integrations/django/test_basic.py +++ b/tests/integrations/django/test_basic.py @@ -1,9 +1,10 @@ import inspect import json import os +import pytest import re import sys -import pytest + from functools import partial from unittest.mock import patch @@ -15,8 +16,8 @@ from django.core.management import execute_from_command_line from django.db.utils import OperationalError, ProgrammingError, DataError from django.http.request import RawPostDataException -from django.utils.functional import SimpleLazyObject from django.template.context import make_context +from django.utils.functional import SimpleLazyObject try: from django.urls import reverse @@ -956,6 +957,44 @@ def test_render_spans(sentry_init, client, capture_events, render_span_tree): assert expected_line in render_span_tree(transaction) +@pytest.mark.skipif(DJANGO_VERSION < (1, 9), reason="Requires Django >= 1.9") +@pytest.mark.forked +@pytest_mark_django_db_decorator() +def test_render_spans_queryset_in_data(sentry_init, client, capture_events): + sentry_init( + integrations=[ + DjangoIntegration( + cache_spans=False, + middleware_spans=False, + signals_spans=False, + ) + ], + traces_sample_rate=1.0, + ) + events = capture_events() + + client.get(reverse("template_test4")) + + (transaction,) = events + template_context = transaction["spans"][-1]["data"]["context"] + + assert template_context["user_age"] == 25 + assert template_context["complex_context"].startswith( + "= (1, 10): EXPECTED_MIDDLEWARE_SPANS = """\ - op="http.server": description=null From 092b5cbd5f7e6fd83035ec646aebc54229b59f44 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 18 Sep 2025 08:20:53 +0200 Subject: [PATCH 482/591] tests: Move pure_eval under toxgen (#4815) ### Description - moved pure_eval under toxgen #### Issues Ref https://github.com/getsentry/sentry-python/issues/4506 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/config.py | 3 +++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 6 ------ tox.ini | 16 +++++++++------- 4 files changed, 12 insertions(+), 14 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 8e35eaba4f..634aa62526 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -224,6 +224,9 @@ "openfeature": { "package": "openfeature-sdk", }, + "pure_eval": { + "package": "pure_eval", + }, "pymongo": { "package": "pymongo", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index e08c2d4b95..5734167a6a 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -70,7 +70,6 @@ # of these from the IGNORE list "gcp", "httpx", - "pure_eval", "ray", "redis", "requests", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 4a4bd96c52..a433406bd8 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -61,9 +61,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # pure_eval - {py3.6,py3.12,py3.13}-pure_eval - # Ray {py3.10,py3.11}-ray-v{2.34} {py3.10,py3.11}-ray-latest @@ -177,9 +174,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # pure_eval - pure_eval: pure_eval - # Ray ray-v2.34: ray~=2.34.0 ray-latest: ray diff --git a/tox.ini b/tox.ini index 50ac22e886..d325d6071e 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-17T07:20:17.058541+00:00 +# Last generated: 2025-09-17T14:43:06.969007+00:00 [tox] requires = @@ -61,9 +61,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # pure_eval - {py3.6,py3.12,py3.13}-pure_eval - # Ray {py3.10,py3.11}-ray-v{2.34} {py3.10,py3.11}-ray-latest @@ -317,6 +314,10 @@ envlist = # ~~~ Misc ~~~ {py3.6,py3.12,py3.13}-loguru-v0.7.3 + {py3.6,py3.7,py3.8}-pure_eval-v0.0.3 + {py3.6,py3.8,py3.9}-pure_eval-v0.1.1 + {py3.7,py3.12,py3.13}-pure_eval-v0.2.3 + {py3.6}-trytond-v4.6.22 {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 @@ -402,9 +403,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # pure_eval - pure_eval: pure_eval - # Ray ray-v2.34: ray~=2.34.0 ray-latest: ray @@ -809,6 +807,10 @@ deps = # ~~~ Misc ~~~ loguru-v0.7.3: loguru==0.7.3 + pure_eval-v0.0.3: pure_eval==0.0.3 + pure_eval-v0.1.1: pure_eval==0.1.1 + pure_eval-v0.2.3: pure_eval==0.2.3 + trytond-v4.6.22: trytond==4.6.22 trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 From 7ae6866a650228a67a18abfa52b4d488b50d9be3 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 18 Sep 2025 06:54:02 +0000 Subject: [PATCH 483/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(09/18)=20(#4816)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> --- scripts/populate_tox/populate_tox.py | 2 +- tox.ini | 42 +++++++++++++++------------- 2 files changed, 24 insertions(+), 20 deletions(-) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 5734167a6a..0d5f5043ed 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -438,7 +438,7 @@ def _render_dependencies(integration: str, releases: list[Version]) -> list[str] for dep in deps: rendered.append(f"{{{constraint}}}-{integration}: {dep}") else: - restriction = SpecifierSet(constraint) + restriction = SpecifierSet(constraint, prereleases=True) for release in releases: if release in restriction: for dep in deps: diff --git a/tox.ini b/tox.ini index d325d6071e..dc289b4587 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-17T14:43:06.969007+00:00 +# Last generated: 2025-09-18T06:43:59.191429+00:00 [tox] requires = @@ -95,7 +95,7 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.33.1 {py3.8,py3.11,py3.12}-anthropic-v0.50.0 - {py3.8,py3.12,py3.13}-anthropic-v0.67.0 + {py3.8,py3.12,py3.13}-anthropic-v0.68.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 @@ -111,14 +111,14 @@ envlist = {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 - {py3.8,py3.11,py3.12}-openai-base-v1.36.1 - {py3.8,py3.11,py3.12}-openai-base-v1.71.0 - {py3.8,py3.12,py3.13}-openai-base-v1.107.3 + {py3.8,py3.11,py3.12}-openai-base-v1.37.2 + {py3.8,py3.11,py3.12}-openai-base-v1.73.0 + {py3.8,py3.12,py3.13}-openai-base-v1.108.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.36.1 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.71.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.107.3 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 + {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -138,7 +138,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.32 + {py3.9,py3.12,py3.13}-boto3-v1.40.33 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -254,6 +254,7 @@ envlist = {py3.8,py3.11,py3.12}-django-v4.2.24 {py3.10,py3.11,py3.12}-django-v5.0.14 {py3.10,py3.12,py3.13}-django-v5.2.6 + {py3.12,py3.13}-django-v6.0a1 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 @@ -459,7 +460,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.33.1: anthropic==0.33.1 anthropic-v0.50.0: anthropic==0.50.0 - anthropic-v0.67.0: anthropic==0.67.0 + anthropic-v0.68.0: anthropic==0.68.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.33.1: httpx<0.28.0 @@ -485,21 +486,21 @@ deps = langchain-notiktoken-v0.3.27: langchain-community openai-base-v1.0.1: openai==1.0.1 - openai-base-v1.36.1: openai==1.36.1 - openai-base-v1.71.0: openai==1.71.0 - openai-base-v1.107.3: openai==1.107.3 + openai-base-v1.37.2: openai==1.37.2 + openai-base-v1.73.0: openai==1.73.0 + openai-base-v1.108.0: openai==1.108.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 - openai-base-v1.36.1: httpx<0.28 + openai-base-v1.37.2: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 - openai-notiktoken-v1.36.1: openai==1.36.1 - openai-notiktoken-v1.71.0: openai==1.71.0 - openai-notiktoken-v1.107.3: openai==1.107.3 + openai-notiktoken-v1.37.2: openai==1.37.2 + openai-notiktoken-v1.73.0: openai==1.73.0 + openai-notiktoken-v1.108.0: openai==1.108.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 - openai-notiktoken-v1.36.1: httpx<0.28 + openai-notiktoken-v1.37.2: httpx<0.28 langgraph-v0.6.7: langgraph==0.6.7 langgraph-v1.0.0a3: langgraph==1.0.0a3 @@ -521,7 +522,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.32: boto3==1.40.32 + boto3-v1.40.33: boto3==1.40.33 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -665,6 +666,7 @@ deps = django-v4.2.24: django==4.2.24 django-v5.0.14: django==5.0.14 django-v5.2.6: django==5.2.6 + django-v6.0a1: django==6.0a1 django: psycopg2-binary django: djangorestframework django: pytest-django @@ -674,11 +676,13 @@ deps = django-v4.2.24: channels[daphne] django-v5.0.14: channels[daphne] django-v5.2.6: channels[daphne] + django-v6.0a1: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio django-v4.2.24: pytest-asyncio django-v5.0.14: pytest-asyncio django-v5.2.6: pytest-asyncio + django-v6.0a1: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 From e1101516330033d04c158f02d1405b2d8d795697 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 18 Sep 2025 12:22:05 +0200 Subject: [PATCH 484/591] tests: Move sanic under toxgen & add option to control how many releases to pick (#4767) - Move sanic under toxgen - This gets rid of the `-latest` group for Web 2 as all Web 2 test suites are now governed by toxgen - Add `num_versions` as a new option to optionally slim down the number of tested versions (useful for sanic as it has a lot of "majors" due to calver) - When a package had multiple majors, we used to always pick the lowest as well as the highest release in the last major -- removed the lowest now to not over-test. All majors will still be tested by default, we just won't be testing two releases in the last major. Ref https://github.com/getsentry/sentry-python/issues/4506 --- .github/workflows/test-integrations-web-2.yml | 95 ------------------- scripts/populate_tox/README.md | 9 +- scripts/populate_tox/config.py | 10 ++ scripts/populate_tox/populate_tox.py | 75 +++++++++------ scripts/populate_tox/tox.jinja | 18 ---- tox.ini | 52 ++++------ 6 files changed, 82 insertions(+), 177 deletions(-) diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 6b7fe58815..3dbe2e1168 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -22,101 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_2-latest: - name: Web 2 (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.9","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v6 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test aiohttp latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-aiohttp-latest" - - name: Test asgi latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-asgi-latest" - - name: Test bottle latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-bottle-latest" - - name: Test falcon latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-falcon-latest" - - name: Test litestar latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-litestar-latest" - - name: Test pyramid latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-pyramid-latest" - - name: Test quart latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-quart-latest" - - name: Test sanic latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-sanic-latest" - - name: Test starlite latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-starlite-latest" - - name: Test tornado latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-tornado-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-web_2-pinned: name: Web 2 (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index c48d57734d..4acebd3259 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -18,8 +18,7 @@ then determining which versions make sense to test to get good coverage. The lowest supported and latest version of a framework are always tested, with a number of releases in between: -- If the package has majors, we pick the highest version of each major. For the - latest major, we also pick the lowest version in that major. +- If the package has majors, we pick the highest version of each major. - If the package doesn't have multiple majors, we pick two versions in between lowest and highest. @@ -46,6 +45,8 @@ integration_name: { }, "python": python_version_specifier, "include": package_version_specifier, + "integration_name": integration_name, + "num_versions": int, } ``` @@ -161,6 +162,10 @@ of which are actually testing the `openai` integration. If this is the case, you Linking an integration to a test suite allows the script to access integration configuration like for example the minimum version defined in `sentry_sdk/integrations/__init__.py`. +### `num_versions` + +With this option you can tweak the default version picking behavior by specifying how many package versions should be tested. It accepts an integer equal to or greater than 2, as the oldest and latest supported versions will always be picked. Additionally, if there is a recent prerelease, it'll also always be picked (this doesn't count towards `num_versions`). + ## How-Tos diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 634aa62526..ce9fdefe48 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -259,6 +259,16 @@ "requests": { "package": "requests", }, + "sanic": { + "package": "sanic", + "deps": { + "*": ["websockets<11.0", "aiohttp"], + ">=22": ["sanic-testing"], + "py3.6": ["aiocontextvars==0.2.1"], + "py3.8": ["tracerite<1.1.2"], + }, + "num_versions": 4, + }, "spark": { "package": "pyspark", "python": ">=3.8", diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 0d5f5043ed..a545eed330 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -74,7 +74,6 @@ "redis", "requests", "rq", - "sanic", } @@ -258,47 +257,48 @@ def _supports_lowest(release: Version) -> bool: def pick_releases_to_test( - releases: list[Version], last_prerelease: Optional[Version] + integration: str, releases: list[Version], last_prerelease: Optional[Version] ) -> list[Version]: """Pick a handful of releases to test from a sorted list of supported releases.""" # If the package has majors (or major-like releases, even if they don't do - # semver), we want to make sure we're testing them all. If not, we just pick - # the oldest, the newest, and a couple in between. + # semver), we want to make sure we're testing them all (unless there's too + # many). If not, we just pick the oldest, the newest, and a couple + # in between. # # If there is a relevant prerelease, also test that in addition to the above. - has_majors = len(set([v.major for v in releases])) > 1 + num_versions = TEST_SUITE_CONFIG[integration].get("num_versions") + if num_versions is not None and ( + not isinstance(num_versions, int) or num_versions < 2 + ): + print(" Integration has invalid `num_versions`: must be an int >= 2") + num_versions = None + + has_majors = len({v.major for v in releases}) > 1 filtered_releases = set() if has_majors: # Always check the very first supported release filtered_releases.add(releases[0]) - # Find out the min and max release by each major + # Find out the max release by each major releases_by_major = {} for release in releases: - if release.major not in releases_by_major: - releases_by_major[release.major] = [release, release] - if release < releases_by_major[release.major][0]: - releases_by_major[release.major][0] = release - if release > releases_by_major[release.major][1]: - releases_by_major[release.major][1] = release - - for i, (min_version, max_version) in enumerate(releases_by_major.values()): + if ( + release.major not in releases_by_major + or release > releases_by_major[release.major] + ): + releases_by_major[release.major] = release + + # Add the highest release in each major + for max_version in releases_by_major.values(): filtered_releases.add(max_version) - if i == len(releases_by_major) - 1: - # If this is the latest major release, also check the lowest - # version of this version - filtered_releases.add(min_version) + + # If num_versions was provided, slim down the selection + if num_versions is not None: + filtered_releases = _pick_releases(sorted(filtered_releases), num_versions) else: - filtered_releases = { - releases[0], # oldest version supported - releases[len(releases) // 3], - releases[ - len(releases) // 3 * 2 - ], # two releases in between, roughly evenly spaced - releases[-1], # latest - } + filtered_releases = _pick_releases(releases, num_versions) filtered_releases = sorted(filtered_releases) if last_prerelease is not None: @@ -307,6 +307,25 @@ def pick_releases_to_test( return filtered_releases +def _pick_releases( + releases: list[Version], num_versions: Optional[int] +) -> set[Version]: + num_versions = num_versions or 4 + + versions = { + releases[0], # oldest version supported + releases[-1], # latest + } + + for i in range(1, num_versions - 1): + try: + versions.add(releases[len(releases) // (num_versions - 1) * i]) + except IndexError: + pass + + return versions + + def supported_python_versions( package_python_versions: Union[SpecifierSet, list[Version]], custom_supported_versions: Optional[SpecifierSet] = None, @@ -631,7 +650,9 @@ def main(fail_on_changes: bool = False) -> None: # Pick a handful of the supported releases to actually test against # and fetch the PyPI data for each to determine which Python versions # to test it on - test_releases = pick_releases_to_test(releases, latest_prerelease) + test_releases = pick_releases_to_test( + integration, releases, latest_prerelease + ) for release in test_releases: _add_python_versions_to_release(integration, package, release) diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index a433406bd8..0522c60231 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -81,12 +81,6 @@ envlist = {py3.7,py3.11,py3.12}-rq-v{1.15,1.16} {py3.7,py3.12,py3.13}-rq-latest - # Sanic - {py3.6,py3.7}-sanic-v{0.8} - {py3.6,py3.8}-sanic-v{20} - {py3.8,py3.11,py3.12}-sanic-v{24.6} - {py3.9,py3.12,py3.13}-sanic-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -210,18 +204,6 @@ deps = rq-v1.16: rq~=1.16.0 rq-latest: rq - # Sanic - sanic: websockets<11.0 - sanic: aiohttp - sanic-v{24.6}: sanic_testing - sanic-latest: sanic_testing - {py3.6}-sanic: aiocontextvars==0.2.1 - {py3.8}-sanic: tracerite<1.1.2 - sanic-v0.8: sanic~=0.8.0 - sanic-v20: sanic~=20.0 - sanic-v24.6: sanic~=24.6.0 - sanic-latest: sanic - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. diff --git a/tox.ini b/tox.ini index dc289b4587..e2898ad8f3 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-18T06:43:59.191429+00:00 +# Last generated: 2025-09-18T08:05:49.500134+00:00 [tox] requires = @@ -81,12 +81,6 @@ envlist = {py3.7,py3.11,py3.12}-rq-v{1.15,1.16} {py3.7,py3.12,py3.13}-rq-latest - # Sanic - {py3.6,py3.7}-sanic-v{0.8} - {py3.6,py3.8}-sanic-v{20} - {py3.8,py3.11,py3.12}-sanic-v{24.6} - {py3.9,py3.12,py3.13}-sanic-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -156,11 +150,9 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 - {py3.6,py3.9,py3.10}-pymongo-v4.0.2 {py3.9,py3.12,py3.13}-pymongo-v4.15.1 {py3.6}-redis_py_cluster_legacy-v1.3.6 - {py3.6,py3.7}-redis_py_cluster_legacy-v2.0.0 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 {py3.6,py3.8,py3.9}-sqlalchemy-v1.3.24 @@ -228,7 +220,6 @@ envlist = {py3.9,py3.12,py3.13}-beam-v2.68.0rc2 {py3.6,py3.7,py3.8}-celery-v4.4.7 - {py3.6,py3.7,py3.8}-celery-v5.0.5 {py3.8,py3.12,py3.13}-celery-v5.5.3 {py3.8,py3.12,py3.13}-celery-v5.6.0b1 @@ -252,13 +243,11 @@ envlist = {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 {py3.8,py3.11,py3.12}-django-v4.2.24 - {py3.10,py3.11,py3.12}-django-v5.0.14 {py3.10,py3.12,py3.13}-django-v5.2.6 {py3.12,py3.13}-django-v6.0a1 {py3.6,py3.7,py3.8}-flask-v1.1.4 {py3.8,py3.12,py3.13}-flask-v2.3.3 - {py3.8,py3.12,py3.13}-flask-v3.0.3 {py3.9,py3.12,py3.13}-flask-v3.1.2 {py3.6,py3.9,py3.10}-starlette-v0.16.0 @@ -284,7 +273,6 @@ envlist = {py3.6}-falcon-v1.4.1 {py3.6,py3.7}-falcon-v2.0.0 {py3.6,py3.11,py3.12}-falcon-v3.1.3 - {py3.8,py3.11,py3.12}-falcon-v4.0.2 {py3.8,py3.11,py3.12}-falcon-v4.1.0 {py3.8,py3.10,py3.11}-litestar-v2.0.1 @@ -301,6 +289,11 @@ envlist = {py3.7,py3.10,py3.11}-quart-v0.18.4 {py3.9,py3.12,py3.13}-quart-v0.20.0 + {py3.6}-sanic-v0.8.3 + {py3.6,py3.8,py3.9}-sanic-v20.12.7 + {py3.8,py3.10,py3.11}-sanic-v23.12.2 + {py3.9,py3.12,py3.13}-sanic-v25.3.0 + {py3.8,py3.10,py3.11}-starlite-v1.48.1 {py3.8,py3.10,py3.11}-starlite-v1.49.0 {py3.8,py3.10,py3.11}-starlite-v1.50.2 @@ -323,7 +316,6 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.8,py3.11,py3.12}-trytond-v7.0.36 {py3.9,py3.12,py3.13}-trytond-v7.6.7 {py3.7,py3.12,py3.13}-typer-v0.15.4 @@ -440,18 +432,6 @@ deps = rq-v1.16: rq~=1.16.0 rq-latest: rq - # Sanic - sanic: websockets<11.0 - sanic: aiohttp - sanic-v{24.6}: sanic_testing - sanic-latest: sanic_testing - {py3.6}-sanic: aiocontextvars==0.2.1 - {py3.8}-sanic: tracerite<1.1.2 - sanic-v0.8: sanic~=0.8.0 - sanic-v20: sanic~=20.0 - sanic-v24.6: sanic~=24.6.0 - sanic-latest: sanic - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -543,12 +523,10 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 - pymongo-v4.0.2: pymongo==4.0.2 pymongo-v4.15.1: pymongo==4.15.1 pymongo: mockupdb redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 - redis_py_cluster_legacy-v2.0.0: redis-py-cluster==2.0.0 redis_py_cluster_legacy-v2.1.3: redis-py-cluster==2.1.3 sqlalchemy-v1.3.24: sqlalchemy==1.3.24 @@ -637,7 +615,6 @@ deps = beam-v2.68.0rc2: apache-beam==2.68.0rc2 celery-v4.4.7: celery==4.4.7 - celery-v5.0.5: celery==5.0.5 celery-v5.5.3: celery==5.5.3 celery-v5.6.0b1: celery==5.6.0b1 celery: newrelic<10.17.0 @@ -664,7 +641,6 @@ deps = django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 django-v4.2.24: django==4.2.24 - django-v5.0.14: django==5.0.14 django-v5.2.6: django==5.2.6 django-v6.0a1: django==6.0a1 django: psycopg2-binary @@ -674,13 +650,11 @@ deps = django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] django-v4.2.24: channels[daphne] - django-v5.0.14: channels[daphne] django-v5.2.6: channels[daphne] django-v6.0a1: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio django-v4.2.24: pytest-asyncio - django-v5.0.14: pytest-asyncio django-v5.2.6: pytest-asyncio django-v6.0a1: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 @@ -694,7 +668,6 @@ deps = flask-v1.1.4: flask==1.1.4 flask-v2.3.3: flask==2.3.3 - flask-v3.0.3: flask==3.0.3 flask-v3.1.2: flask==3.1.2 flask: flask-login flask: werkzeug @@ -746,7 +719,6 @@ deps = falcon-v1.4.1: falcon==1.4.1 falcon-v2.0.0: falcon==2.0.0 falcon-v3.1.3: falcon==3.1.3 - falcon-v4.0.2: falcon==4.0.2 falcon-v4.1.0: falcon==4.1.0 litestar-v2.0.1: litestar==2.0.1 @@ -787,6 +759,17 @@ deps = quart-v0.18.4: hypercorn<0.15.0 {py3.8}-quart: taskgroup==0.0.0a4 + sanic-v0.8.3: sanic==0.8.3 + sanic-v20.12.7: sanic==20.12.7 + sanic-v23.12.2: sanic==23.12.2 + sanic-v25.3.0: sanic==25.3.0 + sanic: websockets<11.0 + sanic: aiohttp + sanic-v23.12.2: sanic-testing + sanic-v25.3.0: sanic-testing + {py3.6}-sanic: aiocontextvars==0.2.1 + {py3.8}-sanic: tracerite<1.1.2 + starlite-v1.48.1: starlite==1.48.1 starlite-v1.49.0: starlite==1.49.0 starlite-v1.50.2: starlite==1.50.2 @@ -819,7 +802,6 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.0.36: trytond==7.0.36 trytond-v7.6.7: trytond==7.6.7 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 From 78843dcbd47228adc4dc84788880d09ac81a1e40 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 18 Sep 2025 12:41:11 +0200 Subject: [PATCH 485/591] tests: Move ray under toxgen (#4810) ### Description - moving the Ray test suite under toxgen - removed `@pytest.mark.forked` and instead: - added a fixture to ensure ray is shut down after each test - gave each test that was looking at logs its own temporary directory to ensure logs don't get mixed up - we were only testing Ray on Python 3.10 and 3.11 -- fixing that #### Issues Ref https://github.com/getsentry/sentry-python/issues/4506 Closes https://github.com/getsentry/sentry-python/issues/4811 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-tasks.yml | 2 +- scripts/populate_tox/config.py | 5 + scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 8 - tests/integrations/ray/test_ray.py | 154 +++++++++++------- tox.ini | 16 +- 6 files changed, 105 insertions(+), 81 deletions(-) diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 0038f1d050..d84789b767 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.10","3.11","3.12","3.13"] + python-version: ["3.7","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index ce9fdefe48..346cabc2db 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -253,6 +253,11 @@ "py3.8": ["taskgroup==0.0.0a4"], }, }, + "ray": { + "package": "ray", + "python": ">=3.9", + "num_versions": 2, + }, "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index a545eed330..a3f1ec8fea 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -70,7 +70,6 @@ # of these from the IGNORE list "gcp", "httpx", - "ray", "redis", "requests", "rq", diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 0522c60231..9c511d3f1e 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -61,10 +61,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Ray - {py3.10,py3.11}-ray-v{2.34} - {py3.10,py3.11}-ray-latest - # Redis {py3.6,py3.8}-redis-v{3} {py3.7,py3.8,py3.11}-redis-v{4} @@ -168,10 +164,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Ray - ray-v2.34: ray~=2.34.0 - ray-latest: ray - # Redis redis: fakeredis!=1.7.4 redis: pytest<8.0.0 diff --git a/tests/integrations/ray/test_ray.py b/tests/integrations/ray/test_ray.py index b5bdd473c4..f4e67df038 100644 --- a/tests/integrations/ray/test_ray.py +++ b/tests/integrations/ray/test_ray.py @@ -1,6 +1,8 @@ import json import os import pytest +import shutil +import uuid import ray @@ -10,6 +12,12 @@ from tests.conftest import TestTransport +@pytest.fixture(autouse=True) +def shutdown_ray(tmpdir): + yield + ray.shutdown() + + class RayTestTransport(TestTransport): def __init__(self): self.envelopes = [] @@ -20,9 +28,6 @@ def capture_envelope(self, envelope: Envelope) -> None: class RayLoggingTransport(TestTransport): - def __init__(self): - super().__init__() - def capture_envelope(self, envelope: Envelope) -> None: print(envelope.serialize().decode("utf-8", "replace")) @@ -39,13 +44,24 @@ def setup_sentry(transport=None): ) -def read_error_from_log(job_id): - log_dir = "/tmp/ray/session_latest/logs/" +def read_error_from_log(job_id, ray_temp_dir): + # Find the actual session directory that Ray created + session_dirs = [d for d in os.listdir(ray_temp_dir) if d.startswith("session_")] + if not session_dirs: + raise FileNotFoundError(f"No session directory found in {ray_temp_dir}") + + session_dir = os.path.join(ray_temp_dir, session_dirs[0]) + log_dir = os.path.join(session_dir, "logs") + + if not os.path.exists(log_dir): + raise FileNotFoundError(f"No logs directory found at {log_dir}") + log_file = [ f for f in os.listdir(log_dir) if "worker" in f and job_id in f and f.endswith(".out") ][0] + with open(os.path.join(log_dir, log_file), "r") as file: lines = file.readlines() @@ -58,7 +74,6 @@ def read_error_from_log(job_id): return error -@pytest.mark.forked @pytest.mark.parametrize( "task_options", [{}, {"num_cpus": 0, "memory": 1024 * 1024 * 10}] ) @@ -124,40 +139,47 @@ def example_task(): ) -@pytest.mark.forked def test_errors_in_ray_tasks(): setup_sentry_with_logging_transport() - ray.init( - runtime_env={ - "worker_process_setup_hook": setup_sentry_with_logging_transport, - "working_dir": "./", - } - ) + ray_temp_dir = os.path.join("/tmp", f"ray_test_{uuid.uuid4().hex[:8]}") + os.makedirs(ray_temp_dir, exist_ok=True) - # Setup ray task - @ray.remote - def example_task(): - 1 / 0 + try: + ray.init( + runtime_env={ + "worker_process_setup_hook": setup_sentry_with_logging_transport, + "working_dir": "./", + }, + _temp_dir=ray_temp_dir, + ) - with sentry_sdk.start_transaction(op="task", name="ray test transaction"): - with pytest.raises(ZeroDivisionError): - future = example_task.remote() - ray.get(future) + # Setup ray task + @ray.remote + def example_task(): + 1 / 0 - job_id = future.job_id().hex() - error = read_error_from_log(job_id) + with sentry_sdk.start_transaction(op="task", name="ray test transaction"): + with pytest.raises(ZeroDivisionError): + future = example_task.remote() + ray.get(future) - assert error["level"] == "error" - assert ( - error["transaction"] - == "tests.integrations.ray.test_ray.test_errors_in_ray_tasks..example_task" - ) - assert error["exception"]["values"][0]["mechanism"]["type"] == "ray" - assert not error["exception"]["values"][0]["mechanism"]["handled"] + job_id = future.job_id().hex() + error = read_error_from_log(job_id, ray_temp_dir) + + assert error["level"] == "error" + assert ( + error["transaction"] + == "tests.integrations.ray.test_ray.test_errors_in_ray_tasks..example_task" + ) + assert error["exception"]["values"][0]["mechanism"]["type"] == "ray" + assert not error["exception"]["values"][0]["mechanism"]["handled"] + + finally: + if os.path.exists(ray_temp_dir): + shutil.rmtree(ray_temp_dir, ignore_errors=True) -@pytest.mark.forked def test_tracing_in_ray_actors(): setup_sentry() @@ -194,37 +216,45 @@ def increment(self): assert worker_envelopes == [] -@pytest.mark.forked def test_errors_in_ray_actors(): setup_sentry_with_logging_transport() - ray.init( - runtime_env={ - "worker_process_setup_hook": setup_sentry_with_logging_transport, - "working_dir": "./", - } - ) - - # Setup ray actor - @ray.remote - class Counter: - def __init__(self): - self.n = 0 - - def increment(self): - with sentry_sdk.start_span(op="task", name="example actor execution"): - 1 / 0 - - return sentry_sdk.get_client().transport.envelopes - - with sentry_sdk.start_transaction(op="task", name="ray test transaction"): - with pytest.raises(ZeroDivisionError): - counter = Counter.remote() - future = counter.increment.remote() - ray.get(future) - - job_id = future.job_id().hex() - error = read_error_from_log(job_id) - - # We do not capture errors in ray actors yet - assert error is None + ray_temp_dir = os.path.join("/tmp", f"ray_test_{uuid.uuid4().hex[:8]}") + os.makedirs(ray_temp_dir, exist_ok=True) + + try: + ray.init( + runtime_env={ + "worker_process_setup_hook": setup_sentry_with_logging_transport, + "working_dir": "./", + }, + _temp_dir=ray_temp_dir, + ) + + # Setup ray actor + @ray.remote + class Counter: + def __init__(self): + self.n = 0 + + def increment(self): + with sentry_sdk.start_span(op="task", name="example actor execution"): + 1 / 0 + + return sentry_sdk.get_client().transport.envelopes + + with sentry_sdk.start_transaction(op="task", name="ray test transaction"): + with pytest.raises(ZeroDivisionError): + counter = Counter.remote() + future = counter.increment.remote() + ray.get(future) + + job_id = future.job_id().hex() + error = read_error_from_log(job_id, ray_temp_dir) + + # We do not capture errors in ray actors yet + assert error is None + + finally: + if os.path.exists(ray_temp_dir): + shutil.rmtree(ray_temp_dir, ignore_errors=True) diff --git a/tox.ini b/tox.ini index e2898ad8f3..15192bc0ad 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-18T08:05:49.500134+00:00 +# Last generated: 2025-09-18T10:26:22.484602+00:00 [tox] requires = @@ -61,10 +61,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Ray - {py3.10,py3.11}-ray-v{2.34} - {py3.10,py3.11}-ray-latest - # Redis {py3.6,py3.8}-redis-v{3} {py3.7,py3.8,py3.11}-redis-v{4} @@ -233,6 +229,9 @@ envlist = {py3.6,py3.7}-huey-v2.3.2 {py3.6,py3.11,py3.12}-huey-v2.5.3 + {py3.9,py3.10}-ray-v2.7.2 + {py3.9,py3.12,py3.13}-ray-v2.49.1 + {py3.8,py3.9}-spark-v3.0.3 {py3.8,py3.10,py3.11}-spark-v3.5.6 {py3.9,py3.12,py3.13}-spark-v4.0.1 @@ -396,10 +395,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Ray - ray-v2.34: ray~=2.34.0 - ray-latest: ray - # Redis redis: fakeredis!=1.7.4 redis: pytest<8.0.0 @@ -631,6 +626,9 @@ deps = huey-v2.3.2: huey==2.3.2 huey-v2.5.3: huey==2.5.3 + ray-v2.7.2: ray==2.7.2 + ray-v2.49.1: ray==2.49.1 + spark-v3.0.3: pyspark==3.0.3 spark-v3.5.6: pyspark==3.5.6 spark-v4.0.1: pyspark==4.0.1 From 704803355a599bf1973758ef006a5bbaa1bb3967 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 18 Sep 2025 12:53:18 +0200 Subject: [PATCH 486/591] tests: Move httpx under toxgen (#4780) * Move httpx under toxgen * This is not straightforward since we depend on `pytest-httpx` which has a very weird strict relationship to which httpx versions it supports * In addition to that, some `pytest-httpx` versions will not install on Python 3.8 despite the corresponding `httpx` versions being compatible -- added a new way to filter out these scenarios * This also removes the -latest Networking group as the httpx test suite was the last Networking test suite that wasn't completely pinned * Also move gcp to the permanent part of the ignore list as it doesn't fit the toxgen mold Ref https://github.com/getsentry/sentry-python/issues/4506 --- .../workflows/test-integrations-network.yml | 67 ------------------- scripts/populate_tox/README.md | 34 ++++++++-- scripts/populate_tox/config.py | 19 ++++++ scripts/populate_tox/populate_tox.py | 56 +++++++++++++--- scripts/populate_tox/tox.jinja | 28 -------- sentry_sdk/integrations/__init__.py | 1 + tox.ini | 45 +++++-------- 7 files changed, 109 insertions(+), 141 deletions(-) diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index e34706ff09..b2f88d261c 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -22,73 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-network-latest: - name: Network (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.9","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v6 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test grpc latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-grpc-latest" - - name: Test httpx latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-httpx-latest" - - name: Test requests latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-requests-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-network-pinned: name: Network (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index 4acebd3259..8014acff3f 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -107,9 +107,14 @@ This key is optional. ### `python` Sometimes, the whole test suite should only run on specific Python versions. -This can be achieved via the `python` key, which expects a version specifier. +This can be achieved via the `python` key. -For example, if you want AIOHTTP tests to only run on Python 3.7+, you can say: +There are two variants how to define the Python versions to run the test suite +on. + +If you want the test suite to only be run on specific Python versions, you can +set `python` to a version specifier. For example, if you want AIOHTTP tests to +only run on Python 3.7+, you can say: ```python "aiohttp": { @@ -118,12 +123,27 @@ For example, if you want AIOHTTP tests to only run on Python 3.7+, you can say: } ``` +If the Python version to use is dependent on the version of the package under +test, you can use the more expressive dictionary variant. For instance, while +HTTPX v0.28 supports Python 3.8, a test dependency of ours, `pytest-httpx`, +doesn't. If you want to specify that HTTPX test suite should not be run on +a Python version older than 3.9 if the HTTPX version is 0.28 or higher, you can +say: + +```python +"httpx": { + "python": { + # run the test suite for httpx v0.28+ on Python 3.9+ only + ">=0.28": ">=3.9", + }, +} +``` + The `python` key is optional, and when possible, it should be omitted. The script -should automatically detect which Python versions the package supports. -However, if a package has broken -metadata or the SDK is explicitly not supporting some packages on specific -Python versions (because of, for example, broken context vars), the `python` -key can be used. +should automatically detect which Python versions the package supports. However, +if a package has broken metadata or the SDK is explicitly not supporting some +packages on specific Python versions (because of, for example, broken context +vars), the `python` key can be used. ### `include` diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 346cabc2db..cf8b9ae9dc 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -150,6 +150,25 @@ }, "python": ">=3.7", }, + "httpx": { + "package": "httpx", + "deps": { + "*": ["anyio<4.0.0"], + ">=0.16,<0.17": ["pytest-httpx==0.10.0"], + ">=0.17,<0.19": ["pytest-httpx==0.12.0"], + ">=0.19,<0.21": ["pytest-httpx==0.14.0"], + ">=0.21,<0.23": ["pytest-httpx==0.19.0"], + ">=0.23,<0.24": ["pytest-httpx==0.21.0"], + ">=0.24,<0.25": ["pytest-httpx==0.22.0"], + ">=0.25,<0.26": ["pytest-httpx==0.25.0"], + ">=0.26,<0.27": ["pytest-httpx==0.28.0"], + ">=0.27,<0.28": ["pytest-httpx==0.30.0"], + ">=0.28,<0.29": ["pytest-httpx==0.35.0"], + }, + "python": { + ">=0.28": ">=3.9", + }, + }, "huey": { "package": "huey", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index a3f1ec8fea..afcef05438 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -63,13 +63,12 @@ "aws_lambda", "cloud_resource_context", "common", + "gcp", "gevent", "opentelemetry", "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "gcp", - "httpx", "redis", "requests", "rq", @@ -240,9 +239,9 @@ def _supports_lowest(release: Version) -> bool: sys.exit(1) py_versions = determine_python_versions(pypi_data) - target_python_versions = TEST_SUITE_CONFIG[integration].get("python") - if target_python_versions: - target_python_versions = SpecifierSet(target_python_versions) + target_python_versions = _transform_target_python_versions( + TEST_SUITE_CONFIG[integration].get("python") + ) return bool(supported_python_versions(py_versions, target_python_versions)) if not _supports_lowest(releases[0]): @@ -327,7 +326,10 @@ def _pick_releases( def supported_python_versions( package_python_versions: Union[SpecifierSet, list[Version]], - custom_supported_versions: Optional[SpecifierSet] = None, + custom_supported_versions: Optional[ + Union[SpecifierSet, dict[SpecifierSet, SpecifierSet]] + ] = None, + version: Optional[Version] = None, ) -> list[Version]: """ Get the intersection of Python versions supported by the package and the SDK. @@ -354,9 +356,25 @@ def supported_python_versions( curr = MIN_PYTHON_VERSION while curr <= MAX_PYTHON_VERSION: if curr in package_python_versions: - if not custom_supported_versions or curr in custom_supported_versions: + if not custom_supported_versions: supported.append(curr) + else: + if isinstance(custom_supported_versions, SpecifierSet): + if curr in custom_supported_versions: + supported.append(curr) + + elif version is not None and isinstance( + custom_supported_versions, dict + ): + for v, py in custom_supported_versions.items(): + if version in v: + if curr in py: + supported.append(curr) + break + else: + supported.append(curr) + # Construct the next Python version (i.e., bump the minor) next = [int(v) for v in str(curr).split(".")] next[1] += 1 @@ -535,20 +553,38 @@ def _add_python_versions_to_release( time.sleep(PYPI_COOLDOWN) # give PYPI some breathing room - target_python_versions = TEST_SUITE_CONFIG[integration].get("python") - if target_python_versions: - target_python_versions = SpecifierSet(target_python_versions) + target_python_versions = _transform_target_python_versions( + TEST_SUITE_CONFIG[integration].get("python") + ) release.python_versions = pick_python_versions_to_test( supported_python_versions( determine_python_versions(release_pypi_data), target_python_versions, + release, ) ) release.rendered_python_versions = _render_python_versions(release.python_versions) +def _transform_target_python_versions( + python_versions: Union[str, dict[str, str], None] +) -> Union[SpecifierSet, dict[SpecifierSet, SpecifierSet], None]: + """Wrap the contents of the `python` key in SpecifierSets.""" + if not python_versions: + return None + + if isinstance(python_versions, str): + return SpecifierSet(python_versions) + + if isinstance(python_versions, dict): + updated = {} + for key, value in python_versions.items(): + updated[SpecifierSet(key)] = SpecifierSet(value) + return updated + + def get_file_hash() -> str: """Calculate a hash of the tox.ini file.""" hasher = hashlib.md5() diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 9c511d3f1e..40980197ab 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -48,13 +48,6 @@ envlist = # GCP {py3.7}-gcp - # HTTPX - {py3.6,py3.9}-httpx-v{0.16,0.18} - {py3.6,py3.10}-httpx-v{0.20,0.22} - {py3.7,py3.11,py3.12}-httpx-v{0.23,0.24} - {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} - {py3.9,py3.12,py3.13}-httpx-latest - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -137,27 +130,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # HTTPX - httpx-v0.16: pytest-httpx==0.10.0 - httpx-v0.18: pytest-httpx==0.12.0 - httpx-v0.20: pytest-httpx==0.14.0 - httpx-v0.22: pytest-httpx==0.19.0 - httpx-v0.23: pytest-httpx==0.21.0 - httpx-v0.24: pytest-httpx==0.22.0 - httpx-v0.25: pytest-httpx==0.25.0 - httpx: pytest-httpx - # anyio is a dep of httpx - httpx: anyio<4.0.0 - httpx-v0.16: httpx~=0.16.0 - httpx-v0.18: httpx~=0.18.0 - httpx-v0.20: httpx~=0.20.0 - httpx-v0.22: httpx~=0.22.0 - httpx-v0.23: httpx~=0.23.0 - httpx-v0.24: httpx~=0.24.0 - httpx-v0.25: httpx~=0.25.0 - httpx-v0.27: httpx~=0.27.0 - httpx-latest: httpx - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 2f5a1f397e..e397c9986a 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -141,6 +141,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "gql": (3, 4, 1), "graphene": (3, 3), "grpc": (1, 32, 0), # grpcio + "httpx": (0, 16, 0), "huggingface_hub": (0, 24, 7), "langchain": (0, 1, 0), "langgraph": (0, 6, 6), diff --git a/tox.ini b/tox.ini index 15192bc0ad..b646228b7f 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-18T10:26:22.484602+00:00 +# Last generated: 2025-09-18T10:42:56.677852+00:00 [tox] requires = @@ -48,13 +48,6 @@ envlist = # GCP {py3.7}-gcp - # HTTPX - {py3.6,py3.9}-httpx-v{0.16,0.18} - {py3.6,py3.10}-httpx-v{0.20,0.22} - {py3.7,py3.11,py3.12}-httpx-v{0.23,0.24} - {py3.9,py3.11,py3.12}-httpx-v{0.25,0.27} - {py3.9,py3.12,py3.13}-httpx-latest - # OpenTelemetry (OTel) {py3.7,py3.9,py3.12,py3.13}-opentelemetry @@ -202,6 +195,11 @@ envlist = {py3.7,py3.11,py3.12}-grpc-v1.62.3 {py3.9,py3.12,py3.13}-grpc-v1.75.0 + {py3.6,py3.8,py3.9}-httpx-v0.16.1 + {py3.6,py3.9,py3.10}-httpx-v0.20.0 + {py3.7,py3.10,py3.11}-httpx-v0.24.1 + {py3.9,py3.11,py3.12}-httpx-v0.28.1 + # ~~~ Tasks ~~~ {py3.7,py3.9,py3.10}-arq-v0.23 @@ -368,27 +366,6 @@ deps = aws_lambda: requests aws_lambda: uvicorn - # HTTPX - httpx-v0.16: pytest-httpx==0.10.0 - httpx-v0.18: pytest-httpx==0.12.0 - httpx-v0.20: pytest-httpx==0.14.0 - httpx-v0.22: pytest-httpx==0.19.0 - httpx-v0.23: pytest-httpx==0.21.0 - httpx-v0.24: pytest-httpx==0.22.0 - httpx-v0.25: pytest-httpx==0.25.0 - httpx: pytest-httpx - # anyio is a dep of httpx - httpx: anyio<4.0.0 - httpx-v0.16: httpx~=0.16.0 - httpx-v0.18: httpx~=0.18.0 - httpx-v0.20: httpx~=0.20.0 - httpx-v0.22: httpx~=0.22.0 - httpx-v0.23: httpx~=0.23.0 - httpx-v0.24: httpx~=0.24.0 - httpx-v0.25: httpx~=0.25.0 - httpx-v0.27: httpx~=0.27.0 - httpx-latest: httpx - # OpenTelemetry (OTel) opentelemetry: opentelemetry-distro @@ -592,6 +569,16 @@ deps = grpc: types-protobuf grpc: pytest-asyncio + httpx-v0.16.1: httpx==0.16.1 + httpx-v0.20.0: httpx==0.20.0 + httpx-v0.24.1: httpx==0.24.1 + httpx-v0.28.1: httpx==0.28.1 + httpx: anyio<4.0.0 + httpx-v0.16.1: pytest-httpx==0.10.0 + httpx-v0.20.0: pytest-httpx==0.14.0 + httpx-v0.24.1: pytest-httpx==0.22.0 + httpx-v0.28.1: pytest-httpx==0.35.0 + # ~~~ Tasks ~~~ arq-v0.23: arq==0.23 From d5fb1bbf07d2b9b954c96f0036af7597571c61b2 Mon Sep 17 00:00:00 2001 From: Marcelo Galigniana Date: Thu, 18 Sep 2025 11:48:09 -0400 Subject: [PATCH 487/591] test(spark): Improve `test_spark` speed (#4822) Fixes GH-4623 ### Description I'm trying to improve the `test_spark.py` speed using the same Spark connection in the entire test session. In my local it has passed from image to image --- tests/integrations/spark/test_spark.py | 23 +++++++++++++++-------- 1 file changed, 15 insertions(+), 8 deletions(-) diff --git a/tests/integrations/spark/test_spark.py b/tests/integrations/spark/test_spark.py index 91882a0b8f..0ea2770b89 100644 --- a/tests/integrations/spark/test_spark.py +++ b/tests/integrations/spark/test_spark.py @@ -28,11 +28,12 @@ def sentry_init_with_reset(sentry_init): _processed_integrations.discard("spark") -@pytest.fixture(scope="function") +@pytest.fixture(scope="session") def create_spark_context(): conf = SparkConf().set("spark.driver.bindAddress", "127.0.0.1") - yield lambda: SparkContext(conf=conf, appName="Testing123") - SparkContext._active_spark_context.stop() + sc = SparkContext(conf=conf, appName="Testing123") + yield lambda: sc + sc.stop() def test_set_app_properties(create_spark_context): @@ -61,12 +62,18 @@ def test_start_sentry_listener(create_spark_context): def test_initialize_spark_integration_before_spark_context_init( mock_patch_spark_context_init, sentry_init_with_reset, - create_spark_context, ): - sentry_init_with_reset() - create_spark_context() - - mock_patch_spark_context_init.assert_called_once() + # As we are using the same SparkContext connection for the whole session, + # we clean it during this test. + original_context = SparkContext._active_spark_context + SparkContext._active_spark_context = None + + try: + sentry_init_with_reset() + mock_patch_spark_context_init.assert_called_once() + finally: + # Restore the original one. + SparkContext._active_spark_context = original_context @patch("sentry_sdk.integrations.spark.spark_driver._activate_integration") From f3b3c70dcd31b285c4248fa0c1413a52a03c3326 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 19 Sep 2025 13:30:52 +0200 Subject: [PATCH 488/591] tests: Move redis under toxgen (#4824) ### Description Moved redis under toxgen, effectively moving the whole DBs group away from the old system. #### Issues Ref https://github.com/getsentry/sentry-python/issues/4506 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-dbs.yml | 99 --------------------- scripts/populate_tox/config.py | 9 ++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 17 ---- tox.ini | 46 +++++----- 5 files changed, 33 insertions(+), 139 deletions(-) diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index efa9f8db39..d5edba3105 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -22,105 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-dbs-latest: - name: DBs (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.7","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - services: - postgres: - image: postgres - env: - POSTGRES_PASSWORD: sentry - # Set health checks to wait until postgres has started - options: >- - --health-cmd pg_isready - --health-interval 10s - --health-timeout 5s - --health-retries 5 - # Maps tcp port 5432 on service container to the host - ports: - - 5432:5432 - env: - SENTRY_PYTHON_TEST_POSTGRES_HOST: ${{ matrix.python-version == '3.6' && 'postgres' || 'localhost' }} - SENTRY_PYTHON_TEST_POSTGRES_USER: postgres - SENTRY_PYTHON_TEST_POSTGRES_PASSWORD: sentry - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v6 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: "Setup ClickHouse Server" - uses: getsentry/action-clickhouse-in-ci@v1.6 - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test asyncpg latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-asyncpg-latest" - - name: Test clickhouse_driver latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-clickhouse_driver-latest" - - name: Test pymongo latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-pymongo-latest" - - name: Test redis latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-redis-latest" - - name: Test redis_py_cluster_legacy latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-redis_py_cluster_legacy-latest" - - name: Test sqlalchemy latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-sqlalchemy-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-dbs-pinned: name: DBs (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index cf8b9ae9dc..87f181c6e3 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -277,6 +277,15 @@ "python": ">=3.9", "num_versions": 2, }, + "redis": { + "package": "redis", + "deps": { + "*": ["fakeredis!=1.7.4", "pytest<8.0.0"], + ">=4.0,<5.0": ["fakeredis<2.31.0"], + "py3.6,py3.7,py3.8": ["fakeredis<2.26.0"], + "py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13": ["pytest-asyncio"], + }, + }, "redis_py_cluster_legacy": { "package": "redis-py-cluster", }, diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index afcef05438..5a44baec24 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -69,7 +69,6 @@ "potel", # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list - "redis", "requests", "rq", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 40980197ab..f8112b7ff5 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -54,12 +54,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Redis - {py3.6,py3.8}-redis-v{3} - {py3.7,py3.8,py3.11}-redis-v{4} - {py3.7,py3.11,py3.12}-redis-v{5} - {py3.7,py3.12,py3.13}-redis-latest - # Requests {py3.6,py3.8,py3.12,py3.13}-requests @@ -136,17 +130,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Redis - redis: fakeredis!=1.7.4 - redis: pytest<8.0.0 - {py3.6,py3.7,py3.8}-redis: fakeredis<2.26.0 - {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio - redis-v3: redis~=3.0 - redis-v4: redis~=4.0 - redis-v4: fakeredis<2.31.0 - redis-v5: redis~=5.0 - redis-latest: redis - # Requests requests: requests>=2.0 diff --git a/tox.ini b/tox.ini index b646228b7f..9e864dd76d 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-18T10:42:56.677852+00:00 +# Last generated: 2025-09-19T10:59:45.339459+00:00 [tox] requires = @@ -54,12 +54,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Redis - {py3.6,py3.8}-redis-v{3} - {py3.7,py3.8,py3.11}-redis-v{4} - {py3.7,py3.11,py3.12}-redis-v{5} - {py3.7,py3.12,py3.13}-redis-latest - # Requests {py3.6,py3.8,py3.12,py3.13}-requests @@ -109,7 +103,7 @@ envlist = {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.3.0 + {py3.10,py3.12,py3.13}-openai_agents-v0.3.1 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 @@ -121,7 +115,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.33 + {py3.9,py3.12,py3.13}-boto3-v1.40.34 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -141,6 +135,13 @@ envlist = {py3.6,py3.10,py3.11}-pymongo-v3.13.0 {py3.9,py3.12,py3.13}-pymongo-v4.15.1 + {py3.6}-redis-v2.10.6 + {py3.6,py3.7,py3.8}-redis-v3.5.3 + {py3.7,py3.10,py3.11}-redis-v4.6.0 + {py3.8,py3.11,py3.12}-redis-v5.3.1 + {py3.9,py3.12,py3.13}-redis-v6.4.0 + {py3.9,py3.12,py3.13}-redis-v7.0.0b1 + {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 @@ -372,17 +373,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Redis - redis: fakeredis!=1.7.4 - redis: pytest<8.0.0 - {py3.6,py3.7,py3.8}-redis: fakeredis<2.26.0 - {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio - redis-v3: redis~=3.0 - redis-v4: redis~=4.0 - redis-v4: fakeredis<2.31.0 - redis-v5: redis~=5.0 - redis-latest: redis - # Requests requests: requests>=2.0 @@ -460,7 +450,7 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.3.0: openai-agents==0.3.0 + openai_agents-v0.3.1: openai-agents==0.3.1 openai_agents: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 @@ -474,7 +464,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.33: boto3==1.40.33 + boto3-v1.40.34: boto3==1.40.34 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -498,6 +488,18 @@ deps = pymongo-v4.15.1: pymongo==4.15.1 pymongo: mockupdb + redis-v2.10.6: redis==2.10.6 + redis-v3.5.3: redis==3.5.3 + redis-v4.6.0: redis==4.6.0 + redis-v5.3.1: redis==5.3.1 + redis-v6.4.0: redis==6.4.0 + redis-v7.0.0b1: redis==7.0.0b1 + redis: fakeredis!=1.7.4 + redis: pytest<8.0.0 + redis-v4.6.0: fakeredis<2.31.0 + {py3.6,py3.7,py3.8}-redis: fakeredis<2.26.0 + {py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-redis: pytest-asyncio + redis_py_cluster_legacy-v1.3.6: redis-py-cluster==1.3.6 redis_py_cluster_legacy-v2.1.3: redis-py-cluster==2.1.3 From 222badbbf6549c97dfe23ec655594e502d4cabbf Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 19 Sep 2025 13:50:15 +0200 Subject: [PATCH 489/591] tests: Move rq under toxgen (#4818) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit ### Description - moved rq under toxgen 🎉 it needed quite some fiddling around with fakeredis versions but seems to work now #### Issues Ref https://github.com/getsentry/sentry-python/issues/4506 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-tasks.yml | 94 ------------------- scripts/populate_tox/config.py | 11 +++ scripts/populate_tox/populate_tox.py | 1 - scripts/populate_tox/tox.jinja | 25 ----- tox.ini | 42 ++++----- 5 files changed, 27 insertions(+), 146 deletions(-) diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index d84789b767..f4c3632261 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -22,100 +22,6 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-tasks-latest: - name: Tasks (latest) - timeout-minutes: 30 - runs-on: ${{ matrix.os }} - strategy: - fail-fast: false - matrix: - python-version: ["3.7","3.12","3.13"] - # python3.6 reached EOL and is no longer being supported on - # new versions of hosted runners on Github Actions - # ubuntu-20.04 is the last version that supported python3.6 - # see https://github.com/actions/setup-python/issues/544#issuecomment-1332535877 - os: [ubuntu-22.04] - # Use Docker container only for Python 3.6 - container: ${{ matrix.python-version == '3.6' && 'python:3.6' || null }} - steps: - - uses: actions/checkout@v5.0.0 - - uses: actions/setup-python@v6 - if: ${{ matrix.python-version != '3.6' }} - with: - python-version: ${{ matrix.python-version }} - allow-prereleases: true - - name: Start Redis - uses: supercharge/redis-github-action@1.8.0 - - name: Install Java - uses: actions/setup-java@v5 - with: - distribution: 'temurin' - java-version: '21' - - name: Setup Test Env - run: | - pip install "coverage[toml]" tox - - name: Erase coverage - run: | - coverage erase - - name: Test arq latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-arq-latest" - - name: Test beam latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-beam-latest" - - name: Test celery latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-celery-latest" - - name: Test dramatiq latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-dramatiq-latest" - - name: Test huey latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-huey-latest" - - name: Test ray latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-ray-latest" - - name: Test rq latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-rq-latest" - - name: Test spark latest - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-spark-latest" - - name: Generate coverage XML (Python 3.6) - if: ${{ !cancelled() && matrix.python-version == '3.6' }} - run: | - export COVERAGE_RCFILE=.coveragerc36 - coverage combine .coverage-sentry-* - coverage xml --ignore-errors - - name: Generate coverage XML - if: ${{ !cancelled() && matrix.python-version != '3.6' }} - run: | - coverage combine .coverage-sentry-* - coverage xml - - name: Upload coverage to Codecov - if: ${{ !cancelled() }} - uses: codecov/codecov-action@v5.5.1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: coverage.xml - # make sure no plugins alter our coverage reports - plugins: noop - verbose: true - - name: Upload test results to Codecov - if: ${{ !cancelled() }} - uses: codecov/test-results-action@v1 - with: - token: ${{ secrets.CODECOV_TOKEN }} - files: .junitxml - verbose: true test-tasks-pinned: name: Tasks (pinned) timeout-minutes: 30 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 87f181c6e3..7d6a519b6b 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -292,6 +292,17 @@ "requests": { "package": "requests", }, + "rq": { + "package": "rq", + "deps": { + # https://github.com/jamesls/fakeredis/issues/245 + # https://github.com/cunla/fakeredis-py/issues/341 + "*": ["fakeredis<2.28.0"], + "<0.9": ["fakeredis<1.0", "redis<3.2.2"], + ">=0.9,<0.14": ["fakeredis>=1.0,<1.7.4"], + "py3.6,py3.7": ["fakeredis!=2.26.0"], + }, + }, "sanic": { "package": "sanic", "deps": { diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 5a44baec24..1ca9319c2e 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -70,7 +70,6 @@ # Integrations that can be migrated -- we should eventually remove all # of these from the IGNORE list "requests", - "rq", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index f8112b7ff5..6931b64cfc 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -57,13 +57,6 @@ envlist = # Requests {py3.6,py3.8,py3.12,py3.13}-requests - # RQ (Redis Queue) - {py3.6}-rq-v{0.6} - {py3.6,py3.9}-rq-v{0.13,1.0} - {py3.6,py3.11}-rq-v{1.5,1.10} - {py3.7,py3.11,py3.12}-rq-v{1.15,1.16} - {py3.7,py3.12,py3.13}-rq-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -133,24 +126,6 @@ deps = # Requests requests: requests>=2.0 - # RQ (Redis Queue) - # https://github.com/jamesls/fakeredis/issues/245 - rq-v{0.6}: fakeredis<1.0 - rq-v{0.6}: redis<3.2.2 - rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 - rq-v{1.15,1.16}: fakeredis<2.28.0 - {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-latest: fakeredis<2.28.0 - {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-v0.6: rq~=0.6.0 - rq-v0.13: rq~=0.13.0 - rq-v1.0: rq~=1.0.0 - rq-v1.5: rq~=1.5.0 - rq-v1.10: rq~=1.10.0 - rq-v1.15: rq~=1.15.0 - rq-v1.16: rq~=1.16.0 - rq-latest: rq - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. diff --git a/tox.ini b/tox.ini index 9e864dd76d..c8ec198810 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-19T10:59:45.339459+00:00 +# Last generated: 2025-09-19T11:39:21.755227+00:00 [tox] requires = @@ -57,13 +57,6 @@ envlist = # Requests {py3.6,py3.8,py3.12,py3.13}-requests - # RQ (Redis Queue) - {py3.6}-rq-v{0.6} - {py3.6,py3.9}-rq-v{0.13,1.0} - {py3.6,py3.11}-rq-v{1.5,1.10} - {py3.7,py3.11,py3.12}-rq-v{1.15,1.16} - {py3.7,py3.12,py3.13}-rq-latest - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -231,6 +224,11 @@ envlist = {py3.9,py3.10}-ray-v2.7.2 {py3.9,py3.12,py3.13}-ray-v2.49.1 + {py3.6}-rq-v0.8.2 + {py3.6,py3.7}-rq-v0.13.0 + {py3.7,py3.11,py3.12}-rq-v1.16.2 + {py3.9,py3.12,py3.13}-rq-v2.6.0 + {py3.8,py3.9}-spark-v3.0.3 {py3.8,py3.10,py3.11}-spark-v3.5.6 {py3.9,py3.12,py3.13}-spark-v4.0.1 @@ -376,24 +374,6 @@ deps = # Requests requests: requests>=2.0 - # RQ (Redis Queue) - # https://github.com/jamesls/fakeredis/issues/245 - rq-v{0.6}: fakeredis<1.0 - rq-v{0.6}: redis<3.2.2 - rq-v{0.13,1.0,1.5,1.10}: fakeredis>=1.0,<1.7.4 - rq-v{1.15,1.16}: fakeredis<2.28.0 - {py3.6,py3.7}-rq-v{1.15,1.16}: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-latest: fakeredis<2.28.0 - {py3.6,py3.7}-rq-latest: fakeredis!=2.26.0 # https://github.com/cunla/fakeredis-py/issues/341 - rq-v0.6: rq~=0.6.0 - rq-v0.13: rq~=0.13.0 - rq-v1.0: rq~=1.0.0 - rq-v1.5: rq~=1.5.0 - rq-v1.10: rq~=1.10.0 - rq-v1.15: rq~=1.15.0 - rq-v1.16: rq~=1.16.0 - rq-latest: rq - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -618,6 +598,16 @@ deps = ray-v2.7.2: ray==2.7.2 ray-v2.49.1: ray==2.49.1 + rq-v0.8.2: rq==0.8.2 + rq-v0.13.0: rq==0.13.0 + rq-v1.16.2: rq==1.16.2 + rq-v2.6.0: rq==2.6.0 + rq: fakeredis<2.28.0 + rq-v0.8.2: fakeredis<1.0 + rq-v0.8.2: redis<3.2.2 + rq-v0.13.0: fakeredis>=1.0,<1.7.4 + {py3.6,py3.7}-rq: fakeredis!=2.26.0 + spark-v3.0.3: pyspark==3.0.3 spark-v3.5.6: pyspark==3.5.6 spark-v4.0.1: pyspark==4.0.1 From 97a5c0896d9673c75d76ab5ee3c2c1608dde91f4 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 19 Sep 2025 14:20:28 +0200 Subject: [PATCH 490/591] tests: Move requests under toxgen (#4825) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit ### Description - Move requests under toxgen - This was the last test suite to port! 🎉 #### Issues ref https://github.com/getsentry/sentry-python/issues/4506 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/config.py | 1 + scripts/populate_tox/populate_tox.py | 3 --- scripts/populate_tox/tox.jinja | 6 ------ tox.ini | 14 +++++++------- 4 files changed, 8 insertions(+), 16 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 7d6a519b6b..70cecd60d9 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -291,6 +291,7 @@ }, "requests": { "package": "requests", + "num_versions": 2, }, "rq": { "package": "rq", diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 1ca9319c2e..17fced9b4f 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -67,9 +67,6 @@ "gevent", "opentelemetry", "potel", - # Integrations that can be migrated -- we should eventually remove all - # of these from the IGNORE list - "requests", } diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 6931b64cfc..98e14adb20 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -54,9 +54,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Requests - {py3.6,py3.8,py3.12,py3.13}-requests - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -123,9 +120,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Requests - requests: requests>=2.0 - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. diff --git a/tox.ini b/tox.ini index c8ec198810..662394de59 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-19T11:39:21.755227+00:00 +# Last generated: 2025-09-19T11:53:09.269997+00:00 [tox] requires = @@ -54,9 +54,6 @@ envlist = # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel - # Requests - {py3.6,py3.8,py3.12,py3.13}-requests - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -194,6 +191,9 @@ envlist = {py3.7,py3.10,py3.11}-httpx-v0.24.1 {py3.9,py3.11,py3.12}-httpx-v0.28.1 + {py3.6}-requests-v2.12.5 + {py3.9,py3.12,py3.13}-requests-v2.32.5 + # ~~~ Tasks ~~~ {py3.7,py3.9,py3.10}-arq-v0.23 @@ -371,9 +371,6 @@ deps = # OpenTelemetry Experimental (POTel) potel: -e .[opentelemetry-experimental] - # Requests - requests: requests>=2.0 - # === Integrations - Auto-generated === # These come from the populate_tox.py script. Eventually we should move all # integration tests there. @@ -561,6 +558,9 @@ deps = httpx-v0.24.1: pytest-httpx==0.22.0 httpx-v0.28.1: pytest-httpx==0.35.0 + requests-v2.12.5: requests==2.12.5 + requests-v2.32.5: requests==2.32.5 + # ~~~ Tasks ~~~ arq-v0.23: arq==0.23 From 4c8c92551ffe7e8c7431fe928de3358317afe8cf Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 22 Sep 2025 07:58:40 +0000 Subject: [PATCH 491/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(09/22)=20(#4831)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- tox.ini | 44 +++++++++++++++++++++++--------------------- 1 file changed, 23 insertions(+), 21 deletions(-) diff --git a/tox.ini b/tox.ini index 662394de59..906d075e2b 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-19T11:53:09.269997+00:00 +# Last generated: 2025-09-22T03:40:00.010016+00:00 [tox] requires = @@ -80,12 +80,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.37.2 {py3.8,py3.11,py3.12}-openai-base-v1.73.0 - {py3.8,py3.12,py3.13}-openai-base-v1.108.0 + {py3.8,py3.12,py3.13}-openai-base-v1.108.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.1 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -105,7 +105,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.34 + {py3.9,py3.12,py3.13}-boto3-v1.40.35 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -147,7 +147,7 @@ envlist = {py3.9,py3.12,py3.13}-launchdarkly-v9.12.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 - {py3.9,py3.12,py3.13}-openfeature-v0.8.2 + {py3.9,py3.12,py3.13}-openfeature-v0.8.3 {py3.7,py3.12,py3.13}-statsig-v0.55.3 {py3.7,py3.12,py3.13}-statsig-v0.58.4 @@ -222,7 +222,7 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.3 {py3.9,py3.10}-ray-v2.7.2 - {py3.9,py3.12,py3.13}-ray-v2.49.1 + {py3.9,py3.12,py3.13}-ray-v2.49.2 {py3.6}-rq-v0.8.2 {py3.6,py3.7}-rq-v0.13.0 @@ -252,9 +252,9 @@ envlist = {py3.9,py3.12,py3.13}-starlette-v0.48.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 - {py3.7,py3.10,py3.11}-fastapi-v0.91.0 - {py3.7,py3.10,py3.11}-fastapi-v0.103.2 - {py3.8,py3.12,py3.13}-fastapi-v0.116.2 + {py3.7,py3.10,py3.11}-fastapi-v0.92.0 + {py3.8,py3.10,py3.11}-fastapi-v0.105.0 + {py3.8,py3.12,py3.13}-fastapi-v0.117.1 # ~~~ Web 2 ~~~ @@ -316,7 +316,8 @@ envlist = {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 - {py3.7,py3.12,py3.13}-typer-v0.17.4 + {py3.7,py3.12,py3.13}-typer-v0.17.5 + {py3.7,py3.12,py3.13}-typer-v0.19.1 @@ -407,7 +408,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.37.2: openai==1.37.2 openai-base-v1.73.0: openai==1.73.0 - openai-base-v1.108.0: openai==1.108.0 + openai-base-v1.108.1: openai==1.108.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -416,7 +417,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.37.2: openai==1.37.2 openai-notiktoken-v1.73.0: openai==1.73.0 - openai-notiktoken-v1.108.0: openai==1.108.0 + openai-notiktoken-v1.108.1: openai==1.108.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.37.2: httpx<0.28 @@ -441,7 +442,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.34: boto3==1.40.34 + boto3-v1.40.35: boto3==1.40.35 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -492,7 +493,7 @@ deps = launchdarkly-v9.12.0: launchdarkly-server-sdk==9.12.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 - openfeature-v0.8.2: openfeature-sdk==0.8.2 + openfeature-v0.8.3: openfeature-sdk==0.8.3 statsig-v0.55.3: statsig==0.55.3 statsig-v0.58.4: statsig==0.58.4 @@ -596,7 +597,7 @@ deps = huey-v2.5.3: huey==2.5.3 ray-v2.7.2: ray==2.7.2 - ray-v2.49.1: ray==2.49.1 + ray-v2.49.2: ray==2.49.2 rq-v0.8.2: rq==0.8.2 rq-v0.13.0: rq==0.13.0 @@ -666,17 +667,17 @@ deps = {py3.6}-starlette: aiocontextvars fastapi-v0.79.1: fastapi==0.79.1 - fastapi-v0.91.0: fastapi==0.91.0 - fastapi-v0.103.2: fastapi==0.103.2 - fastapi-v0.116.2: fastapi==0.116.2 + fastapi-v0.92.0: fastapi==0.92.0 + fastapi-v0.105.0: fastapi==0.105.0 + fastapi-v0.117.1: fastapi==0.117.1 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart fastapi: requests fastapi: anyio<4 fastapi-v0.79.1: httpx<0.28.0 - fastapi-v0.91.0: httpx<0.28.0 - fastapi-v0.103.2: httpx<0.28.0 + fastapi-v0.92.0: httpx<0.28.0 + fastapi-v0.105.0: httpx<0.28.0 {py3.6}-fastapi: aiocontextvars @@ -786,7 +787,8 @@ deps = typer-v0.15.4: typer==0.15.4 typer-v0.16.1: typer==0.16.1 - typer-v0.17.4: typer==0.17.4 + typer-v0.17.5: typer==0.17.5 + typer-v0.19.1: typer==0.19.1 From a92664022dc205434734bd8aa042baa1173a9ce6 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 22 Sep 2025 12:13:19 +0200 Subject: [PATCH 492/591] tests: Remove old system for checking for regressions (#4826) ### Description With all eligible test suites now moved to toxgen, we can remove our old way of testing for regressions in new versions of packages. For context, before toxgen we had two groups of test targets for each test suite: pinned and latest. Pinned targets used to be a small selection of hand-picked pinned versions of the package to test. Latest always pulled the newest release, to check for incompatibilities. There was a lot of code to cater to having these two different categories. With the switch to toxgen, all targets are now pinned, and we make sure we pick a representative sample of versions to test against, always including the latest available version to detect regressions. I'm removing the code dealing with the `latest` group entirely, and removing references to anything `pinned` as this is now simply the default. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-ai.yml | 46 ++++++++-------- .github/workflows/test-integrations-cloud.yml | 30 +++++------ .../workflows/test-integrations-common.yml | 14 ++--- .github/workflows/test-integrations-dbs.yml | 34 ++++++------ .github/workflows/test-integrations-flags.yml | 26 ++++----- .../workflows/test-integrations-gevent.yml | 14 ++--- .../workflows/test-integrations-graphql.yml | 26 ++++----- .github/workflows/test-integrations-misc.yml | 34 ++++++------ .../workflows/test-integrations-network.yml | 22 ++++---- .github/workflows/test-integrations-tasks.yml | 42 +++++++-------- .github/workflows/test-integrations-web-1.yml | 26 ++++----- .github/workflows/test-integrations-web-2.yml | 50 ++++++++--------- scripts/populate_tox/populate_tox.py | 12 +---- scripts/runtox.sh | 18 +------ .../split_tox_gh_actions.py | 53 ++++++------------- .../split_tox_gh_actions/templates/base.jinja | 3 -- .../templates/check_required.jinja | 8 ++- .../templates/test_group.jinja | 14 ++--- 18 files changed, 210 insertions(+), 262 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 972df704e0..cf21720ff1 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-ai-pinned: - name: AI (pinned) + test-ai: + name: AI timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,42 +50,42 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test anthropic pinned + - name: Test anthropic run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-anthropic" - - name: Test cohere pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-anthropic" + - name: Test cohere run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-cohere" - - name: Test langchain-base pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-cohere" + - name: Test langchain-base run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain-base" - - name: Test langchain-notiktoken pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-base" + - name: Test langchain-notiktoken run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langchain-notiktoken" - - name: Test openai-base pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-notiktoken" + - name: Test openai-base run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai-base" - - name: Test openai-notiktoken pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai-base" + - name: Test openai-notiktoken run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai-notiktoken" - - name: Test langgraph pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai-notiktoken" + - name: Test langgraph run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-langgraph" - - name: Test openai_agents pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-langgraph" + - name: Test openai_agents run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openai_agents" - - name: Test huggingface_hub pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_agents" + - name: Test huggingface_hub run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-huggingface_hub" + ./scripts/runtox.sh "py${{ matrix.python-version }}-huggingface_hub" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -114,13 +114,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned AI tests passed - needs: test-ai-pinned + name: All AI tests passed + needs: test-ai # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-ai-pinned.result != 'success' + if: needs.test-ai.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 6aeaea8c3a..48a2881ffc 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-cloud-pinned: - name: Cloud (pinned) + test-cloud: + name: Cloud timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -54,26 +54,26 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test aws_lambda pinned + - name: Test aws_lambda run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aws_lambda" - - name: Test boto3 pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-aws_lambda" + - name: Test boto3 run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-boto3" - - name: Test chalice pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-boto3" + - name: Test chalice run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-chalice" - - name: Test cloud_resource_context pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-chalice" + - name: Test cloud_resource_context run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-cloud_resource_context" - - name: Test gcp pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-cloud_resource_context" + - name: Test gcp run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-gcp" + ./scripts/runtox.sh "py${{ matrix.python-version }}-gcp" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -102,13 +102,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Cloud tests passed - needs: test-cloud-pinned + name: All Cloud tests passed + needs: test-cloud # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-cloud-pinned.result != 'success' + if: needs.test-cloud.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index b682428dd1..eb5dde8b17 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-common-pinned: - name: Common (pinned) + test-common: + name: Common timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,10 +50,10 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test common pinned + - name: Test common run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-common" + ./scripts/runtox.sh "py${{ matrix.python-version }}-common" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -82,13 +82,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Common tests passed - needs: test-common-pinned + name: All Common tests passed + needs: test-common # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-common-pinned.result != 'success' + if: needs.test-common.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index d5edba3105..a778ce86d5 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-dbs-pinned: - name: DBs (pinned) + test-dbs: + name: DBs timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -70,30 +70,30 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test asyncpg pinned + - name: Test asyncpg run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-asyncpg" - - name: Test clickhouse_driver pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-asyncpg" + - name: Test clickhouse_driver run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-clickhouse_driver" - - name: Test pymongo pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-clickhouse_driver" + - name: Test pymongo run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-pymongo" - - name: Test redis pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-pymongo" + - name: Test redis run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-redis" - - name: Test redis_py_cluster_legacy pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-redis" + - name: Test redis_py_cluster_legacy run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-redis_py_cluster_legacy" - - name: Test sqlalchemy pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-redis_py_cluster_legacy" + - name: Test sqlalchemy run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-sqlalchemy" + ./scripts/runtox.sh "py${{ matrix.python-version }}-sqlalchemy" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -122,13 +122,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned DBs tests passed - needs: test-dbs-pinned + name: All DBs tests passed + needs: test-dbs # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-dbs-pinned.result != 'success' + if: needs.test-dbs.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index d7baeeb870..ce622d4258 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-flags-pinned: - name: Flags (pinned) + test-flags: + name: Flags timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,22 +50,22 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test launchdarkly pinned + - name: Test launchdarkly run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-launchdarkly" - - name: Test openfeature pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-launchdarkly" + - name: Test openfeature run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-openfeature" - - name: Test statsig pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-openfeature" + - name: Test statsig run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-statsig" - - name: Test unleash pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-statsig" + - name: Test unleash run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-unleash" + ./scripts/runtox.sh "py${{ matrix.python-version }}-unleash" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -94,13 +94,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Flags tests passed - needs: test-flags-pinned + name: All Flags tests passed + needs: test-flags # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-flags-pinned.result != 'success' + if: needs.test-flags.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 9af6b4d7af..8b077f357c 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-gevent-pinned: - name: Gevent (pinned) + test-gevent: + name: Gevent timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,10 +50,10 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test gevent pinned + - name: Test gevent run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-gevent" + ./scripts/runtox.sh "py${{ matrix.python-version }}-gevent" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -82,13 +82,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Gevent tests passed - needs: test-gevent-pinned + name: All Gevent tests passed + needs: test-gevent # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-gevent-pinned.result != 'success' + if: needs.test-gevent.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 5c306dff3f..81d809c94e 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-graphql-pinned: - name: GraphQL (pinned) + test-graphql: + name: GraphQL timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,22 +50,22 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test ariadne pinned + - name: Test ariadne run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-ariadne" - - name: Test gql pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-ariadne" + - name: Test gql run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-gql" - - name: Test graphene pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-gql" + - name: Test graphene run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-graphene" - - name: Test strawberry pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-graphene" + - name: Test strawberry run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-strawberry" + ./scripts/runtox.sh "py${{ matrix.python-version }}-strawberry" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -94,13 +94,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned GraphQL tests passed - needs: test-graphql-pinned + name: All GraphQL tests passed + needs: test-graphql # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-graphql-pinned.result != 'success' + if: needs.test-graphql.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 005e8395a2..2aaf2f07fe 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-misc-pinned: - name: Misc (pinned) + test-misc: + name: Misc timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,30 +50,30 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test loguru pinned + - name: Test loguru run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-loguru" - - name: Test opentelemetry pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-loguru" + - name: Test opentelemetry run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-opentelemetry" - - name: Test potel pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-opentelemetry" + - name: Test potel run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-potel" - - name: Test pure_eval pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-potel" + - name: Test pure_eval run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-pure_eval" - - name: Test trytond pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-pure_eval" + - name: Test trytond run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-trytond" - - name: Test typer pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-trytond" + - name: Test typer run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-typer" + ./scripts/runtox.sh "py${{ matrix.python-version }}-typer" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -102,13 +102,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Misc tests passed - needs: test-misc-pinned + name: All Misc tests passed + needs: test-misc # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-misc-pinned.result != 'success' + if: needs.test-misc.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index b2f88d261c..e8b5b36c98 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-network-pinned: - name: Network (pinned) + test-network: + name: Network timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,18 +50,18 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test grpc pinned + - name: Test grpc run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-grpc" - - name: Test httpx pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-grpc" + - name: Test httpx run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-httpx" - - name: Test requests pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-httpx" + - name: Test requests run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-requests" + ./scripts/runtox.sh "py${{ matrix.python-version }}-requests" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -90,13 +90,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Network tests passed - needs: test-network-pinned + name: All Network tests passed + needs: test-network # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-network-pinned.result != 'success' + if: needs.test-network.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index f4c3632261..2dfae9c9a3 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-tasks-pinned: - name: Tasks (pinned) + test-tasks: + name: Tasks timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -57,38 +57,38 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test arq pinned + - name: Test arq run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-arq" - - name: Test beam pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-arq" + - name: Test beam run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-beam" - - name: Test celery pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-beam" + - name: Test celery run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-celery" - - name: Test dramatiq pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-celery" + - name: Test dramatiq run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-dramatiq" - - name: Test huey pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-dramatiq" + - name: Test huey run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-huey" - - name: Test ray pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-huey" + - name: Test ray run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-ray" - - name: Test rq pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-ray" + - name: Test rq run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-rq" - - name: Test spark pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-rq" + - name: Test spark run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-spark" + ./scripts/runtox.sh "py${{ matrix.python-version }}-spark" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -117,13 +117,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Tasks tests passed - needs: test-tasks-pinned + name: All Tasks tests passed + needs: test-tasks # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-tasks-pinned.result != 'success' + if: needs.test-tasks.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 4b22db6155..455f15723a 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_1-pinned: - name: Web 1 (pinned) + test-web_1: + name: Web 1 timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -68,22 +68,22 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test django pinned + - name: Test django run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-django" - - name: Test flask pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-django" + - name: Test flask run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-flask" - - name: Test starlette pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-flask" + - name: Test starlette run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-starlette" - - name: Test fastapi pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-starlette" + - name: Test fastapi run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-fastapi" + ./scripts/runtox.sh "py${{ matrix.python-version }}-fastapi" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -112,13 +112,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Web 1 tests passed - needs: test-web_1-pinned + name: All Web 1 tests passed + needs: test-web_1 # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-web_1-pinned.result != 'success' + if: needs.test-web_1.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 3dbe2e1168..98cf4456e8 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -22,8 +22,8 @@ env: CACHED_BUILD_PATHS: | ${{ github.workspace }}/dist-serverless jobs: - test-web_2-pinned: - name: Web 2 (pinned) + test-web_2: + name: Web 2 timeout-minutes: 30 runs-on: ${{ matrix.os }} strategy: @@ -50,46 +50,46 @@ jobs: - name: Erase coverage run: | coverage erase - - name: Test aiohttp pinned + - name: Test aiohttp run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-aiohttp" - - name: Test asgi pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-aiohttp" + - name: Test asgi run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-asgi" - - name: Test bottle pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-asgi" + - name: Test bottle run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-bottle" - - name: Test falcon pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-bottle" + - name: Test falcon run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-falcon" - - name: Test litestar pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-falcon" + - name: Test litestar run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-litestar" - - name: Test pyramid pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-litestar" + - name: Test pyramid run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-pyramid" - - name: Test quart pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-pyramid" + - name: Test quart run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-quart" - - name: Test sanic pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-quart" + - name: Test sanic run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-sanic" - - name: Test starlite pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-sanic" + - name: Test starlite run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-starlite" - - name: Test tornado pinned + ./scripts/runtox.sh "py${{ matrix.python-version }}-starlite" + - name: Test tornado run: | set -x # print commands that are executed - ./scripts/runtox.sh --exclude-latest "py${{ matrix.python-version }}-tornado" + ./scripts/runtox.sh "py${{ matrix.python-version }}-tornado" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | @@ -118,13 +118,13 @@ jobs: files: .junitxml verbose: true check_required_tests: - name: All pinned Web 2 tests passed - needs: test-web_2-pinned + name: All Web 2 tests passed + needs: test-web_2 # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-web_2-pinned.result != 'success' + if: needs.test-web_2.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 17fced9b4f..37ddfb9daf 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -49,16 +49,8 @@ IGNORE = { # Do not try auto-generating the tox entries for these. They will be - # hardcoded in tox.ini. - # - # This set should be getting smaller over time as we migrate more test - # suites over to this script. Some entries will probably stay forever - # as they don't fit the mold (e.g. common, asgi, which don't have a 3rd party - # pypi package to install in different versions). - # - # Test suites that will have to remain hardcoded since they don't fit the - # toxgen usecase (there is no one package that should be tested in different - # versions) + # hardcoded in tox.ini since they don't fit the toxgen usecase (there is no + # one package that should be tested in different versions). "asgi", "aws_lambda", "cloud_resource_context", diff --git a/scripts/runtox.sh b/scripts/runtox.sh index 6acf4406fb..bd6e954947 100755 --- a/scripts/runtox.sh +++ b/scripts/runtox.sh @@ -13,25 +13,9 @@ else TOXPATH=./.venv/bin/tox fi -excludelatest=false -for arg in "$@" -do - if [ "$arg" = "--exclude-latest" ]; then - excludelatest=true - shift - break - fi -done - searchstring="$1" -if $excludelatest; then - echo "Excluding latest" - ENV="$($TOXPATH -l | grep -- "$searchstring" | grep -v -- '-latest' | tr $'\n' ',')" -else - echo "Including latest" - ENV="$($TOXPATH -l | grep -- "$searchstring" | tr $'\n' ',')" -fi +ENV="$($TOXPATH -l | grep -- "$searchstring" | tr $'\n' ',')" if [ -z "${ENV}" ]; then echo "No targets found. Skipping." diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 51ee614d04..a7b7c394b1 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -29,11 +29,9 @@ TOXENV_REGEX = re.compile( r""" {?(?P(py\d+\.\d+,?)+)}? - -(?P[a-z](?:[a-z_]|-(?!v{?\d|latest))*[a-z0-9]) + -(?P[a-z](?:[a-z_]|-(?!v{?\d))*[a-z0-9]) (?:-( (v{?(?P[0-9.]+[0-9a-z,.]*}?)) - | - (?Platest) ))? """, re.VERBOSE, @@ -167,13 +165,11 @@ def main(fail_on_changes): old_hash = get_files_hash() print("Parsing tox.ini...") - py_versions_pinned, py_versions_latest = parse_tox() + py_versions = parse_tox() if fail_on_changes: print("Checking if all frameworks belong in a group...") - missing_frameworks = find_frameworks_missing_from_groups( - py_versions_pinned, py_versions_latest - ) + missing_frameworks = find_frameworks_missing_from_groups(py_versions) if missing_frameworks: raise RuntimeError( "Please add the following frameworks to the corresponding group " @@ -183,9 +179,7 @@ def main(fail_on_changes): print("Rendering templates...") for group, frameworks in GROUPS.items(): - contents = render_template( - group, frameworks, py_versions_pinned, py_versions_latest - ) + contents = render_template(group, frameworks, py_versions) filename = write_file(contents, group) print(f"Created {filename}") @@ -213,8 +207,7 @@ def parse_tox(): if line.strip() and not line.strip().startswith("#") ] - py_versions_pinned = defaultdict(set) - py_versions_latest = defaultdict(set) + py_versions = defaultdict(set) parsed_correctly = True @@ -232,14 +225,10 @@ def parse_tox(): groups = parsed.groupdict() raw_python_versions = groups["py_versions"] framework = groups["framework"] - framework_versions_latest = groups.get("framework_versions_latest") or False # collect python versions to test the framework in raw_python_versions = set(raw_python_versions.split(",")) - if framework_versions_latest: - py_versions_latest[framework] |= raw_python_versions - else: - py_versions_pinned[framework] |= raw_python_versions + py_versions[framework] |= raw_python_versions except Exception: print(f"ERROR reading line {line}") @@ -248,15 +237,14 @@ def parse_tox(): if not parsed_correctly: raise RuntimeError("Failed to parse tox.ini") - py_versions_pinned = _normalize_py_versions(py_versions_pinned) - py_versions_latest = _normalize_py_versions(py_versions_latest) + py_versions = _normalize_py_versions(py_versions) - return py_versions_pinned, py_versions_latest + return py_versions -def find_frameworks_missing_from_groups(py_versions_pinned, py_versions_latest): +def find_frameworks_missing_from_groups(py_versions): frameworks_in_a_group = _union(GROUPS.values()) - all_frameworks = set(py_versions_pinned.keys()) | set(py_versions_latest.keys()) + all_frameworks = set(py_versions) return all_frameworks - frameworks_in_a_group @@ -296,33 +284,26 @@ def _union(seq): return reduce(lambda x, y: set(x) | set(y), seq) -def render_template(group, frameworks, py_versions_pinned, py_versions_latest): +def render_template(group, frameworks, py_versions): template = ENV.get_template("base.jinja") - categories = set() - py_versions = defaultdict(set) + py_versions_final = set() for framework in frameworks: - if py_versions_pinned[framework]: - categories.add("pinned") - py_versions["pinned"] |= set(py_versions_pinned[framework]) - if py_versions_latest[framework]: - categories.add("latest") - py_versions["latest"] |= set(py_versions_latest[framework]) + py_versions_final |= set(py_versions[framework]) context = { "group": group, "frameworks": frameworks, - "categories": sorted(categories), "needs_clickhouse": bool(set(frameworks) & FRAMEWORKS_NEEDING_CLICKHOUSE), "needs_docker": bool(set(frameworks) & FRAMEWORKS_NEEDING_DOCKER), "needs_postgres": bool(set(frameworks) & FRAMEWORKS_NEEDING_POSTGRES), "needs_redis": bool(set(frameworks) & FRAMEWORKS_NEEDING_REDIS), "needs_java": bool(set(frameworks) & FRAMEWORKS_NEEDING_JAVA), - "py_versions": { - category: [f'"{version}"' for version in _normalize_py_versions(versions)] - for category, versions in py_versions.items() - }, + "py_versions": [ + f'"{version}"' for version in _normalize_py_versions(py_versions_final) + ], } + rendered = template.render(context) rendered = postprocess_template(rendered) return rendered diff --git a/scripts/split_tox_gh_actions/templates/base.jinja b/scripts/split_tox_gh_actions/templates/base.jinja index 75c988e32a..5d59e63348 100644 --- a/scripts/split_tox_gh_actions/templates/base.jinja +++ b/scripts/split_tox_gh_actions/templates/base.jinja @@ -30,9 +30,6 @@ env: {% raw %}${{ github.workspace }}/dist-serverless{% endraw %} jobs: -{% for category in categories %} {% include "test_group.jinja" %} -{% endfor %} - {% include "check_required.jinja" %} {% endwith %} diff --git a/scripts/split_tox_gh_actions/templates/check_required.jinja b/scripts/split_tox_gh_actions/templates/check_required.jinja index 9a2bbed830..37d4f8fd78 100644 --- a/scripts/split_tox_gh_actions/templates/check_required.jinja +++ b/scripts/split_tox_gh_actions/templates/check_required.jinja @@ -1,13 +1,11 @@ check_required_tests: - name: All pinned {{ group }} tests passed - {% if "pinned" in categories %} - needs: test-{{ group | replace(" ", "_") | lower }}-pinned - {% endif %} + name: All {{ group }} tests passed + needs: test-{{ group | replace(" ", "_") | lower }} # Always run this, even if a dependent job failed if: always() runs-on: ubuntu-22.04 steps: - name: Check for failures - if: needs.test-{{ lowercase_group }}-pinned.result != 'success' + if: needs.test-{{ lowercase_group }}.result != 'success' run: | echo "One of the dependent jobs has failed. You may need to re-run it." && exit 1 diff --git a/scripts/split_tox_gh_actions/templates/test_group.jinja b/scripts/split_tox_gh_actions/templates/test_group.jinja index f020a44b84..59369265b3 100644 --- a/scripts/split_tox_gh_actions/templates/test_group.jinja +++ b/scripts/split_tox_gh_actions/templates/test_group.jinja @@ -1,11 +1,11 @@ - test-{{ lowercase_group }}-{{ category }}: - name: {{ group }} ({{ category }}) + test-{{ lowercase_group }}: + name: {{ group }} timeout-minutes: 30 runs-on: {% raw %}${{ matrix.os }}{% endraw %} strategy: fail-fast: false matrix: - python-version: [{{ py_versions.get(category)|join(",") }}] + python-version: [{{ py_versions|join(",") }}] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 @@ -75,14 +75,10 @@ coverage erase {% for framework in frameworks %} - - name: Test {{ framework }} {{ category }} + - name: Test {{ framework }} run: | set -x # print commands that are executed - {% if category == "pinned" %} - ./scripts/runtox.sh --exclude-latest "{% raw %}py${{ matrix.python-version }}{% endraw %}-{{ framework }}" - {% elif category == "latest" %} - ./scripts/runtox.sh "{% raw %}py${{ matrix.python-version }}{% endraw %}-{{ framework }}-latest" - {% endif %} + ./scripts/runtox.sh "{% raw %}py${{ matrix.python-version }}{% endraw %}-{{ framework }}" {% endfor %} - name: Generate coverage XML (Python 3.6) From 34a35809492cc6cd1413ab0a5da06da261863844 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Mon, 22 Sep 2025 14:02:58 +0200 Subject: [PATCH 493/591] Use weakref in dedupe where possible (#4834) We cannot use `weakref` always since built-in exception types like `ZeroDivisionError` raise a `TypeError`. This makes the memory usage at least a little better for custom exception types. #### Issues * resolves: #3165 * resolves: PY-17 --- sentry_sdk/integrations/dedupe.py | 18 ++++++++++++++++-- 1 file changed, 16 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/dedupe.py b/sentry_sdk/integrations/dedupe.py index eab2764fcd..99ac6ce164 100644 --- a/sentry_sdk/integrations/dedupe.py +++ b/sentry_sdk/integrations/dedupe.py @@ -1,3 +1,5 @@ +import weakref + import sentry_sdk from sentry_sdk.utils import ContextVar, logger from sentry_sdk.integrations import Integration @@ -35,12 +37,24 @@ def processor(event, hint): if exc_info is None: return event + last_seen = integration._last_seen.get(None) + if last_seen is not None: + # last_seen is either a weakref or the original instance + last_seen = ( + last_seen() if isinstance(last_seen, weakref.ref) else last_seen + ) + exc = exc_info[1] - if integration._last_seen.get(None) is exc: + if last_seen is exc: logger.info("DedupeIntegration dropped duplicated error event %s", exc) return None - integration._last_seen.set(exc) + # we can only weakref non builtin types + try: + integration._last_seen.set(weakref.ref(exc)) + except TypeError: + integration._last_seen.set(exc) + return event @staticmethod From 1b9f0cfe03fb9fa3ae9a399ae99da726b16189a2 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 22 Sep 2025 16:41:45 +0200 Subject: [PATCH 494/591] fix: Pin shibuya (#4839) ### Description Apidocs build started [failing](https://github.com/getsentry/sentry-python/actions/runs/17918383503/job/50946544869?pr=4837) a couple hours ago. A [new version](https://pypi.org/project/shibuya/2025.9.22/) of the shibuya theme was released recently, which seems to be the culprit. Pinning for now to unblock, will investigate whether we can just adapt. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- requirements-docs.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements-docs.txt b/requirements-docs.txt index 81e04ba3ef..b98ad4834a 100644 --- a/requirements-docs.txt +++ b/requirements-docs.txt @@ -1,5 +1,5 @@ gevent -shibuya +shibuya<2025.9.22 sphinx<8.2 sphinx-autodoc-typehints[type_comments]>=1.8.0 typing-extensions From 3f4e2a8497ae942b380ce4d2459cac7bb364506e Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 22 Sep 2025 15:10:08 +0000 Subject: [PATCH 495/591] build(deps): bump actions/setup-python from 5 to 6 (#4830) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/setup-python](https://github.com/actions/setup-python) from 5 to 6.
Release notes

Sourced from actions/setup-python's releases.

v6.0.0

What's Changed

Breaking Changes

Make sure your runner is on version v2.327.1 or later to ensure compatibility with this release. See Release Notes

Enhancements:

Bug fixes:

Dependency updates:

New Contributors

Full Changelog: https://github.com/actions/setup-python/compare/v5...v6.0.0

v5.6.0

What's Changed

Full Changelog: https://github.com/actions/setup-python/compare/v5...v5.6.0

v5.5.0

What's Changed

Enhancements:

Bug fixes:

... (truncated)

Commits
  • e797f83 Upgrade to node 24 (#1164)
  • 3d1e2d2 Revert "Enhance cache-dependency-path handling to support files outside the w...
  • 65b0712 Clarify pythonLocation behavior for PyPy and GraalPy in environment variables...
  • 5b668cf Bump actions/checkout from 4 to 5 (#1181)
  • f62a0e2 Change missing cache directory error to warning (#1182)
  • 9322b3c Upgrade setuptools to 78.1.1 to fix path traversal vulnerability in PackageIn...
  • fbeb884 Bump form-data to fix critical vulnerabilities #182 & #183 (#1163)
  • 03bb615 Bump idna from 2.9 to 3.7 in /tests/data (#843)
  • 36da51d Add version parsing from Pipfile (#1067)
  • 3c6f142 update documentation (#1156)
  • Additional commits viewable in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/setup-python&package-manager=github_actions&previous-version=5&new-version=6)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- .github/workflows/update-tox.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/update-tox.yml b/.github/workflows/update-tox.yml index 11d143a2a9..0d2640bb47 100644 --- a/.github/workflows/update-tox.yml +++ b/.github/workflows/update-tox.yml @@ -18,7 +18,7 @@ jobs: steps: - name: Setup Python - uses: actions/setup-python@v5 + uses: actions/setup-python@v6 with: python-version: 3.13 From e52c3b117a404fb7755feced02680d29e421b0ee Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Wed, 24 Sep 2025 09:37:26 +0200 Subject: [PATCH 496/591] fix(langchain): don't record tool call output if not include_prompt / should_send_default_pii (#4836) --- sentry_sdk/integrations/langchain.py | 17 ++++----- .../integrations/langchain/test_langchain.py | 36 +++++++++++++++++++ 2 files changed, 45 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 1401be06e1..b92f6983a1 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -322,14 +322,15 @@ def on_llm_end(self, response, *, run_id, **kwargs): pass try: - tool_calls = getattr(generation.message, "tool_calls", None) - if tool_calls is not None and tool_calls != []: - set_data_normalized( - span, - SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, - tool_calls, - unpack=False, - ) + if should_send_default_pii() and self.include_prompts: + tool_calls = getattr(generation.message, "tool_calls", None) + if tool_calls is not None and tool_calls != []: + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + tool_calls, + unpack=False, + ) except AttributeError: pass diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index b6b432c523..2a40945413 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -221,6 +221,19 @@ def test_langchain_agent( in chat_spans[1]["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] ) assert "5" in chat_spans[1]["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + + # Verify tool calls are recorded when PII is enabled + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in chat_spans[0].get( + "data", {} + ), "Tool calls should be recorded when send_default_pii=True and include_prompts=True" + tool_calls_data = chat_spans[0]["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS] + assert isinstance(tool_calls_data, (list, str)) # Could be serialized + if isinstance(tool_calls_data, str): + assert "get_word_length" in tool_calls_data + elif isinstance(tool_calls_data, list) and len(tool_calls_data) > 0: + # Check if tool calls contain expected function name + tool_call_str = str(tool_calls_data) + assert "get_word_length" in tool_call_str else: assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in chat_spans[0].get("data", {}) assert SPANDATA.GEN_AI_RESPONSE_TEXT not in chat_spans[0].get("data", {}) @@ -229,6 +242,29 @@ def test_langchain_agent( assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in tool_exec_span.get("data", {}) assert SPANDATA.GEN_AI_RESPONSE_TEXT not in tool_exec_span.get("data", {}) + # Verify tool calls are NOT recorded when PII is disabled + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS not in chat_spans[0].get( + "data", {} + ), ( + f"Tool calls should NOT be recorded when send_default_pii={send_default_pii} " + f"and include_prompts={include_prompts}" + ) + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS not in chat_spans[1].get( + "data", {} + ), ( + f"Tool calls should NOT be recorded when send_default_pii={send_default_pii} " + f"and include_prompts={include_prompts}" + ) + + # Verify that available tools are always recorded regardless of PII settings + for chat_span in chat_spans: + span_data = chat_span.get("data", {}) + if SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS in span_data: + tools_data = span_data[SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS] + assert ( + tools_data is not None + ), "Available tools should always be recorded regardless of PII settings" + def test_langchain_error(sentry_init, capture_events): sentry_init( From 808c1805c0253fe18a21156fb57a219ef2f71acd Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Wed, 24 Sep 2025 09:51:49 +0200 Subject: [PATCH 497/591] fix(AI): Make agents integrations set the span status in case of error (#4820) ### Description Make all AI related integrations set the `span.status` to `"error"` and also the `status` of the transaction/root_span to `"error"` in case of an error. This is also the OTel semantic convention for failing spans. #### Issues * resolves: #4752 * resolves: PY-1825 --- sentry_sdk/consts.py | 4 +- sentry_sdk/integrations/anthropic.py | 21 ++++++-- sentry_sdk/integrations/cohere.py | 4 ++ sentry_sdk/integrations/huggingface_hub.py | 5 +- sentry_sdk/integrations/langchain.py | 6 +-- sentry_sdk/integrations/openai.py | 3 ++ .../openai_agents/spans/execute_tool.py | 2 +- .../integrations/openai_agents/utils.py | 3 ++ sentry_sdk/tracing.py | 3 +- sentry_sdk/tracing_utils.py | 15 +++++- .../integrations/anthropic/test_anthropic.py | 51 ++++++++++++++++++- tests/integrations/cohere/test_cohere.py | 18 +++++++ .../huggingface_hub/test_huggingface_hub.py | 19 +++++++ .../integrations/langchain/test_langchain.py | 47 ++++++++++++++++- tests/integrations/openai/test_openai.py | 20 ++++++++ .../openai_agents/test_openai_agents.py | 26 ++++++++++ 16 files changed, 231 insertions(+), 16 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 91a1740526..606cf804b6 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -765,11 +765,12 @@ class SPANSTATUS: CANCELLED = "cancelled" DATA_LOSS = "data_loss" DEADLINE_EXCEEDED = "deadline_exceeded" + ERROR = "error" # OTel status code: https://opentelemetry.io/docs/concepts/signals/traces/#span-status FAILED_PRECONDITION = "failed_precondition" INTERNAL_ERROR = "internal_error" INVALID_ARGUMENT = "invalid_argument" NOT_FOUND = "not_found" - OK = "ok" + OK = "ok" # HTTP 200 and OTel status code: https://opentelemetry.io/docs/concepts/signals/traces/#span-status OUT_OF_RANGE = "out_of_range" PERMISSION_DENIED = "permission_denied" RESOURCE_EXHAUSTED = "resource_exhausted" @@ -777,6 +778,7 @@ class SPANSTATUS: UNAVAILABLE = "unavailable" UNIMPLEMENTED = "unimplemented" UNKNOWN_ERROR = "unknown_error" + UNSET = "unset" # OTel status code: https://opentelemetry.io/docs/concepts/signals/traces/#span-status class OP: diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 4f4c0b1a2a..d9898fa1d1 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -4,9 +4,10 @@ import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage from sentry_sdk.ai.utils import set_data_normalized, get_start_span_function -from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.consts import OP, SPANDATA, SPANSTATUS from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.tracing_utils import set_span_errored from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, @@ -52,6 +53,8 @@ def setup_once(): def _capture_exception(exc): # type: (Any) -> None + set_span_errored() + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, @@ -357,7 +360,13 @@ def _sentry_patched_create_sync(*args, **kwargs): integration = sentry_sdk.get_client().get_integration(AnthropicIntegration) kwargs["integration"] = integration - return _execute_sync(f, *args, **kwargs) + try: + return _execute_sync(f, *args, **kwargs) + finally: + span = sentry_sdk.get_current_span() + if span is not None and span.status == SPANSTATUS.ERROR: + with capture_internal_exceptions(): + span.__exit__(None, None, None) return _sentry_patched_create_sync @@ -390,6 +399,12 @@ async def _sentry_patched_create_async(*args, **kwargs): integration = sentry_sdk.get_client().get_integration(AnthropicIntegration) kwargs["integration"] = integration - return await _execute_async(f, *args, **kwargs) + try: + return await _execute_async(f, *args, **kwargs) + finally: + span = sentry_sdk.get_current_span() + if span is not None and span.status == SPANSTATUS.ERROR: + with capture_internal_exceptions(): + span.__exit__(None, None, None) return _sentry_patched_create_async diff --git a/sentry_sdk/integrations/cohere.py b/sentry_sdk/integrations/cohere.py index 57ffdb908a..3445900c80 100644 --- a/sentry_sdk/integrations/cohere.py +++ b/sentry_sdk/integrations/cohere.py @@ -7,6 +7,8 @@ from typing import TYPE_CHECKING +from sentry_sdk.tracing_utils import set_span_errored + if TYPE_CHECKING: from typing import Any, Callable, Iterator from sentry_sdk.tracing import Span @@ -84,6 +86,8 @@ def setup_once(): def _capture_exception(exc): # type: (Any) -> None + set_span_errored() + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, diff --git a/sentry_sdk/integrations/huggingface_hub.py b/sentry_sdk/integrations/huggingface_hub.py index cb76ccf507..2e2b382abd 100644 --- a/sentry_sdk/integrations/huggingface_hub.py +++ b/sentry_sdk/integrations/huggingface_hub.py @@ -7,6 +7,7 @@ from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.tracing_utils import set_span_errored from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, @@ -52,6 +53,8 @@ def setup_once(): def _capture_exception(exc): # type: (Any) -> None + set_span_errored() + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, @@ -127,8 +130,6 @@ def new_huggingface_task(*args, **kwargs): try: res = f(*args, **kwargs) except Exception as e: - # Error Handling - span.set_status("error") _capture_exception(e) span.__exit__(None, None, None) raise e from None diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index b92f6983a1..6bc3ceb93e 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -8,8 +8,7 @@ from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii -from sentry_sdk.tracing import Span -from sentry_sdk.tracing_utils import _get_value +from sentry_sdk.tracing_utils import _get_value, set_span_errored from sentry_sdk.utils import logger, capture_internal_exceptions from typing import TYPE_CHECKING @@ -26,6 +25,7 @@ Union, ) from uuid import UUID + from sentry_sdk.tracing import Span try: @@ -116,7 +116,7 @@ def _handle_error(self, run_id, error): span_data = self.span_map[run_id] span = span_data.span - span.set_status("unknown") + set_span_errored(span) sentry_sdk.capture_exception(error, span.scope) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 467116c8f4..4d72ec366c 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -7,6 +7,7 @@ from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.tracing_utils import set_span_errored from sentry_sdk.utils import ( capture_internal_exceptions, event_from_exception, @@ -83,6 +84,8 @@ def _capture_exception(exc, manual_span_cleanup=True): # Close an eventually open span # We need to do this by hand because we are not using the start_span context manager current_span = sentry_sdk.get_current_span() + set_span_errored(current_span) + if manual_span_cleanup and current_span is not None: current_span.__exit__(None, None, None) diff --git a/sentry_sdk/integrations/openai_agents/spans/execute_tool.py b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py index 5f9e4cb340..ad70762cd0 100644 --- a/sentry_sdk/integrations/openai_agents/spans/execute_tool.py +++ b/sentry_sdk/integrations/openai_agents/spans/execute_tool.py @@ -42,7 +42,7 @@ def update_execute_tool_span(span, agent, tool, result): if isinstance(result, str) and result.startswith( "An error occurred while running the tool" ): - span.set_status(SPANSTATUS.INTERNAL_ERROR) + span.set_status(SPANSTATUS.ERROR) if should_send_default_pii(): span.set_data(SPANDATA.GEN_AI_TOOL_OUTPUT, result) diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index a0487e0e3a..73d2858e7f 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -3,6 +3,7 @@ from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.tracing_utils import set_span_errored from sentry_sdk.utils import event_from_exception, safe_serialize from typing import TYPE_CHECKING @@ -20,6 +21,8 @@ def _capture_exception(exc): # type: (Any) -> None + set_span_errored() + event, hint = event_from_exception( exc, client_options=sentry_sdk.get_client().options, diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index fc43a33dc7..4edda21075 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -416,7 +416,8 @@ def __enter__(self): def __exit__(self, ty, value, tb): # type: (Optional[Any], Optional[Any], Optional[Any]) -> None if value is not None and should_be_treated_as_error(ty, value): - self.set_status(SPANSTATUS.INTERNAL_ERROR) + if self.status != SPANSTATUS.ERROR: + self.set_status(SPANSTATUS.INTERNAL_ERROR) with capture_internal_exceptions(): scope, old_span = self._context_manager_state diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index c1cfde293b..2f3e334e3f 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -11,7 +11,7 @@ import uuid import sentry_sdk -from sentry_sdk.consts import OP, SPANDATA, SPANTEMPLATE +from sentry_sdk.consts import OP, SPANDATA, SPANSTATUS, SPANTEMPLATE from sentry_sdk.utils import ( capture_internal_exceptions, filename_for_module, @@ -892,6 +892,19 @@ def get_current_span(scope=None): return current_span +def set_span_errored(span=None): + # type: (Optional[Span]) -> None + """ + Set the status of the current or given span to ERROR. + Also sets the status of the transaction (root span) to ERROR. + """ + span = span or get_current_span() + if span is not None: + span.set_status(SPANSTATUS.ERROR) + if span.containing_transaction is not None: + span.containing_transaction.set_status(SPANSTATUS.ERROR) + + def _generate_sample_rand( trace_id, # type: Optional[str] *, diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index 3893626026..04ff12eb8b 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -698,8 +698,54 @@ def test_exception_message_create(sentry_init, capture_events): max_tokens=1024, ) - (event,) = events + (event, transaction) = events assert event["level"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + +def test_span_status_error(sentry_init, capture_events): + sentry_init(integrations=[AnthropicIntegration()], traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="anthropic"): + client = Anthropic(api_key="z") + client.messages._post = mock.Mock( + side_effect=AnthropicError("API rate limit reached") + ) + with pytest.raises(AnthropicError): + client.messages.create( + model="some-model", + messages=[{"role": "system", "content": "I'm throwing an exception"}], + max_tokens=1024, + ) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + +@pytest.mark.asyncio +async def test_span_status_error_async(sentry_init, capture_events): + sentry_init(integrations=[AnthropicIntegration()], traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="anthropic"): + client = AsyncAnthropic(api_key="z") + client.messages._post = AsyncMock( + side_effect=AnthropicError("API rate limit reached") + ) + with pytest.raises(AnthropicError): + await client.messages.create( + model="some-model", + messages=[{"role": "system", "content": "I'm throwing an exception"}], + max_tokens=1024, + ) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" @pytest.mark.asyncio @@ -718,8 +764,9 @@ async def test_exception_message_create_async(sentry_init, capture_events): max_tokens=1024, ) - (event,) = events + (event, transaction) = events assert event["level"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" def test_span_origin(sentry_init, capture_events): diff --git a/tests/integrations/cohere/test_cohere.py b/tests/integrations/cohere/test_cohere.py index ee876172d1..a97d2befae 100644 --- a/tests/integrations/cohere/test_cohere.py +++ b/tests/integrations/cohere/test_cohere.py @@ -167,6 +167,24 @@ def test_bad_chat(sentry_init, capture_events): assert event["level"] == "error" +def test_span_status_error(sentry_init, capture_events): + sentry_init(integrations=[CohereIntegration()], traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="test"): + client = Client(api_key="z") + HTTPXClient.request = mock.Mock( + side_effect=httpx.HTTPError("API rate limit reached") + ) + with pytest.raises(httpx.HTTPError): + client.chat(model="some-model", message="hello") + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + @pytest.mark.parametrize( "send_default_pii, include_prompts", [(True, True), (True, False), (False, True), (False, False)], diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index 86f9c10109..5aa3928a67 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -654,6 +654,25 @@ def test_chat_completion_api_error( assert span["data"] == expected_data +def test_span_status_error(sentry_init, capture_events, mock_hf_api_with_errors): + # type: (Any, Any, Any) -> None + sentry_init(traces_sample_rate=1.0) + events = capture_events() + + client = InferenceClient(model="test-model") + + with sentry_sdk.start_transaction(name="test"): + with pytest.raises(HfHubHTTPError): + client.chat_completion( + messages=[{"role": "user", "content": "Hello!"}], + ) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_chat_completion_with_tools( diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 2a40945413..af4b6b8c56 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -285,7 +285,7 @@ def test_langchain_error(sentry_init, capture_events): ] ) global stream_result_mock - stream_result_mock = Mock(side_effect=Exception("API rate limit error")) + stream_result_mock = Mock(side_effect=ValueError("API rate limit error")) llm = MockOpenAI( model_name="gpt-3.5-turbo", temperature=0, @@ -295,13 +295,56 @@ def test_langchain_error(sentry_init, capture_events): agent_executor = AgentExecutor(agent=agent, tools=[get_word_length], verbose=True) - with start_transaction(), pytest.raises(Exception): + with start_transaction(), pytest.raises(ValueError): list(agent_executor.stream({"input": "How many letters in the word eudca"})) error = events[0] assert error["level"] == "error" +def test_span_status_error(sentry_init, capture_events): + global llm_type + llm_type = "acme-llm" + + sentry_init( + integrations=[LangchainIntegration(include_prompts=True)], + traces_sample_rate=1.0, + ) + events = capture_events() + + with start_transaction(name="test"): + prompt = ChatPromptTemplate.from_messages( + [ + ( + "system", + "You are very powerful assistant, but don't know current events", + ), + ("user", "{input}"), + MessagesPlaceholder(variable_name="agent_scratchpad"), + ] + ) + global stream_result_mock + stream_result_mock = Mock(side_effect=ValueError("API rate limit error")) + llm = MockOpenAI( + model_name="gpt-3.5-turbo", + temperature=0, + openai_api_key="badkey", + ) + agent = create_openai_tools_agent(llm, [get_word_length], prompt) + + agent_executor = AgentExecutor( + agent=agent, tools=[get_word_length], verbose=True + ) + + with pytest.raises(ValueError): + list(agent_executor.stream({"input": "How many letters in the word eudca"})) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + def test_span_origin(sentry_init, capture_events): sentry_init( integrations=[LangchainIntegration()], diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 18968fb36a..e7fbf8a7d8 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -416,6 +416,26 @@ def test_bad_chat_completion(sentry_init, capture_events): assert event["level"] == "error" +def test_span_status_error(sentry_init, capture_events): + sentry_init(integrations=[OpenAIIntegration()], traces_sample_rate=1.0) + events = capture_events() + + with start_transaction(name="test"): + client = OpenAI(api_key="z") + client.chat.completions._post = mock.Mock( + side_effect=OpenAIError("API rate limit reached") + ) + with pytest.raises(OpenAIError): + client.chat.completions.create( + model="some-model", messages=[{"role": "system", "content": "hello"}] + ) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + @pytest.mark.asyncio async def test_bad_chat_completion_async(sentry_init, capture_events): sentry_init(integrations=[OpenAIIntegration()], traces_sample_rate=1.0) diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index 047b919213..bd7f15faff 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -657,6 +657,32 @@ async def test_error_handling(sentry_init, capture_events, test_agent): assert ai_client_span["tags"]["status"] == "internal_error" +@pytest.mark.asyncio +async def test_span_status_error(sentry_init, capture_events, test_agent): + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + mock_get_response.side_effect = ValueError("Model Error") + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + with pytest.raises(ValueError, match="Model Error"): + await agents.Runner.run( + test_agent, "Test input", run_config=test_run_config + ) + + (error, transaction) = events + assert error["level"] == "error" + assert transaction["spans"][0]["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" + + @pytest.mark.asyncio async def test_multiple_agents_asyncio( sentry_init, capture_events, test_agent, mock_model_response From 13a8ae1f6891503663a1cfc1d07e498f48bbc9f0 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 24 Sep 2025 10:20:09 +0200 Subject: [PATCH 498/591] feat(toxgen): Cache release data (#4835) ### Description We're fetching the same release data from PyPI everytime toxgen is run. Cache it instead to avoid unnecessary network requests. This makes toxgen about twice as fast. I'm sorting the cache file to hopefully avoid merge conflicts (at least it should make them less likely compared to if we'd only append to the file). Let's see how this works, if the cache makes it harder to merge stuff because of merge conflicts, we can get rid of it. Ease of use is more important than runtime. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4817 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/populate_tox.py | 79 ++++++++- scripts/populate_tox/releases.jsonl | 236 +++++++++++++++++++++++++++ tox.ini | 24 ++- 3 files changed, 323 insertions(+), 16 deletions(-) create mode 100644 scripts/populate_tox/releases.jsonl diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 37ddfb9daf..28a9fe8334 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -4,6 +4,7 @@ import functools import hashlib +import json import os import sys import time @@ -34,18 +35,20 @@ # CUTOFF = datetime.now(tz=timezone.utc) - timedelta(days=365 * 5) TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" +RELEASES_CACHE_FILE = Path(__file__).resolve().parent / "releases.jsonl" ENV = Environment( loader=FileSystemLoader(Path(__file__).resolve().parent), trim_blocks=True, lstrip_blocks=True, ) -PYPI_COOLDOWN = 0.1 # seconds to wait between requests to PyPI +PYPI_COOLDOWN = 0.05 # seconds to wait between requests to PyPI PYPI_PROJECT_URL = "https://pypi.python.org/pypi/%7Bproject%7D/json" PYPI_VERSION_URL = "https://pypi.python.org/pypi/%7Bproject%7D/%7Bversion%7D/json" CLASSIFIER_PREFIX = "Programming Language :: Python :: " +CACHE = defaultdict(dict) IGNORE = { # Do not try auto-generating the tox entries for these. They will be @@ -94,9 +97,32 @@ def fetch_package(package: str) -> Optional[dict]: @functools.cache def fetch_release(package: str, version: Version) -> Optional[dict]: - """Fetch release metadata from PyPI.""" + """Fetch release metadata from cache or, failing that, PyPI.""" + release = _fetch_from_cache(package, version) + if release is not None: + return release + url = PYPI_VERSION_URL.format(project=package, version=version) - return fetch_https://github.com/url(url) + release = fetch_https://github.com/url(url) + if release is not None: + _save_to_cache(package, version, release) + return release + + +def _fetch_from_cache(package: str, version: Version) -> Optional[dict]: + package = _normalize_name(package) + if package in CACHE and str(version) in CACHE[package]: + CACHE[package][str(version)]["_accessed"] = True + return CACHE[package][str(version)] + + return None + + +def _save_to_cache(package: str, version: Version, release: Optional[dict]) -> None: + with open(RELEASES_CACHE_FILE, "a") as releases_cache: + releases_cache.write(json.dumps(_normalize_release(release)) + "\n") + + CACHE[_normalize_name(package)][str(version)] = release def _prefilter_releases( @@ -600,6 +626,24 @@ def get_last_updated() -> Optional[datetime]: return timestamp +def _normalize_name(package: str) -> str: + return package.lower().replace("-", "_") + + +def _normalize_release(release: dict) -> dict: + """Filter out unneeded parts of the release JSON.""" + normalized = { + "info": { + "classifiers": release["info"]["classifiers"], + "name": release["info"]["name"], + "requires_python": release["info"]["requires_python"], + "version": release["info"]["version"], + "yanked": release["info"]["yanked"], + }, + } + return normalized + + def main(fail_on_changes: bool = False) -> None: """ Generate tox.ini from the tox.jinja template. @@ -636,6 +680,20 @@ def main(fail_on_changes: bool = False) -> None: f"The SDK supports Python versions {MIN_PYTHON_VERSION} - {MAX_PYTHON_VERSION}." ) + # Load file cache + global CACHE + + with open(RELEASES_CACHE_FILE) as releases_cache: + for line in releases_cache: + release = json.loads(line) + name = _normalize_name(release["info"]["name"]) + version = release["info"]["version"] + CACHE[name][version] = release + CACHE[name][version][ + "_accessed" + ] = False # for cleaning up unused cache entries + + # Process packages packages = defaultdict(list) for group, integrations in GROUPS.items(): @@ -701,6 +759,21 @@ def main(fail_on_changes: bool = False) -> None: packages, update_timestamp=not fail_on_changes, last_updated=last_updated ) + # Sort the release cache file + releases = [] + with open(RELEASES_CACHE_FILE) as releases_cache: + releases = [json.loads(line) for line in releases_cache] + releases.sort(key=lambda r: (r["info"]["name"], r["info"]["version"])) + with open(RELEASES_CACHE_FILE, "w") as releases_cache: + for release in releases: + if ( + CACHE[_normalize_name(release["info"]["name"])][ + release["info"]["version"] + ]["_accessed"] + is True + ): + releases_cache.write(json.dumps(release) + "\n") + if fail_on_changes: new_file_hash = get_file_hash() if old_file_hash != new_file_hash: diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl new file mode 100644 index 0000000000..814d0c15b9 --- /dev/null +++ b/scripts/populate_tox/releases.jsonl @@ -0,0 +1,236 @@ +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.10.8", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.11.29", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.8.19", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.4", "version": "2.0.13", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.5", "version": "2.2.28", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.1.14", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.2.25", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.24", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0a1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Flask", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "1.1.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Flask", "requires_python": ">=3.9", "version": "3.1.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.16.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.17.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.18.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Quart", "requires_python": ">=3.9", "version": "0.20.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "", "version": "1.2.19", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "1.3.24", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7", "version": "1.4.54", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.43", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.1.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.2.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.12.15", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.33.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.50.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.68.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.15.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.19.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.32.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.40.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.8", "version": "2.50.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.68.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.22", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": null, "version": "0.24.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.7", "version": "0.24.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.7", "version": "0.25.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.8", "version": "0.26.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.5.0", "version": "0.23.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.6.0", "version": "0.25.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.7.0", "version": "0.27.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.8.0", "version": "0.30.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.36", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.6.0b1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.21.9", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": "", "version": "1.26.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.13.12", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.18.0", "yanked": false}} +{"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.9.4", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.6", "version": "1.12.3", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.7", "version": "1.15.0", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": "", "version": "1.4.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.117.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": null, "version": "3.5.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": "", "version": "3.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": null, "version": "3.4.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "grpcio", "requires_python": "", "version": "1.32.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.75.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.7", "version": "0.24.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.26.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.28.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": null, "version": "0.1.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.10.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.7.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "2.0.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.1.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.2.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.3.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.0", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} +{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.7", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.10.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.9.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.17.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.108.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.37.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.73.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.1", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": "", "version": "0.1.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "3.2.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.4.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.5.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.6.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.1", "yanked": false}} +{"info": {"classifiers": ["Framework :: Pylons", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": null, "version": "1.0.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.10.8", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.6.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.7.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.8.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.9.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=3.6", "version": "2.0.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.1.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.3.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.4.8", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "3.0.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.49.2", "yanked": false}} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3"], "name": "redis", "requires_python": null, "version": "2.9.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.0.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.2.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "3.5.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}} +{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}} +{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis-py-cluster", "requires_python": "", "version": "1.3.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4", "version": "2.1.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3"], "name": "requests", "requires_python": null, "version": "2.0.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.10.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.11.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.12.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.16.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development :: Libraries"], "name": "requests", "requires_python": ">=3.9", "version": "2.32.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5"], "name": "requests", "requires_python": null, "version": "2.8.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.10.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.13.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.6.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.7.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.8.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=2.7", "version": "1.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.7", "version": "1.16.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.5", "version": "1.8.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.9", "version": "2.6.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "sanic", "requires_python": "", "version": "0.8.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.6", "version": "20.12.7", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "23.12.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "25.3.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.6", "version": "0.16.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.7", "version": "0.27.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.8", "version": "0.38.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.9", "version": "0.48.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.49.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.50.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.58.4", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.61.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.64.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.8", "version": "0.233.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.257.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.282.0", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.7", "version": "6.2", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.8", "version": "6.4.2", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.4.27", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.6.22", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.8.18", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.4", "version": "5.0.63", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.5", "version": "5.4.20", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.7", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.16.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.17.5", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.19.2", "yanked": false}} diff --git a/tox.ini b/tox.ini index 906d075e2b..f80c86b51e 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-22T03:40:00.010016+00:00 +# Last generated: 2025-09-23T13:41:50.551689+00:00 [tox] requires = @@ -80,12 +80,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.37.2 {py3.8,py3.11,py3.12}-openai-base-v1.73.0 - {py3.8,py3.12,py3.13}-openai-base-v1.108.1 + {py3.8,py3.12,py3.13}-openai-base-v1.108.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.1 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.2 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -105,7 +105,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.35 + {py3.9,py3.12,py3.13}-boto3-v1.40.36 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -204,8 +204,7 @@ envlist = {py3.7}-beam-v2.14.0 {py3.7,py3.8}-beam-v2.32.0 {py3.8,py3.10,py3.11}-beam-v2.50.0 - {py3.9,py3.12,py3.13}-beam-v2.67.0 - {py3.9,py3.12,py3.13}-beam-v2.68.0rc2 + {py3.9,py3.12,py3.13}-beam-v2.68.0 {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.8,py3.12,py3.13}-celery-v5.5.3 @@ -317,7 +316,7 @@ envlist = {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.7,py3.12,py3.13}-typer-v0.16.1 {py3.7,py3.12,py3.13}-typer-v0.17.5 - {py3.7,py3.12,py3.13}-typer-v0.19.1 + {py3.8,py3.12,py3.13}-typer-v0.19.2 @@ -408,7 +407,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.37.2: openai==1.37.2 openai-base-v1.73.0: openai==1.73.0 - openai-base-v1.108.1: openai==1.108.1 + openai-base-v1.108.2: openai==1.108.2 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -417,7 +416,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.37.2: openai==1.37.2 openai-notiktoken-v1.73.0: openai==1.73.0 - openai-notiktoken-v1.108.1: openai==1.108.1 + openai-notiktoken-v1.108.2: openai==1.108.2 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.37.2: httpx<0.28 @@ -442,7 +441,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.35: boto3==1.40.35 + boto3-v1.40.36: boto3==1.40.36 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -576,8 +575,7 @@ deps = beam-v2.14.0: apache-beam==2.14.0 beam-v2.32.0: apache-beam==2.32.0 beam-v2.50.0: apache-beam==2.50.0 - beam-v2.67.0: apache-beam==2.67.0 - beam-v2.68.0rc2: apache-beam==2.68.0rc2 + beam-v2.68.0: apache-beam==2.68.0 celery-v4.4.7: celery==4.4.7 celery-v5.5.3: celery==5.5.3 @@ -788,7 +786,7 @@ deps = typer-v0.15.4: typer==0.15.4 typer-v0.16.1: typer==0.16.1 typer-v0.17.5: typer==0.17.5 - typer-v0.19.1: typer==0.19.1 + typer-v0.19.2: typer==0.19.2 From c3d367202427233757f5275271873b4f4e60a825 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 24 Sep 2025 10:37:40 +0200 Subject: [PATCH 499/591] chore: Clean up toxgen (#4855) ### Description Polish toxgen a bit, improving wording, examples, formatting, removing obsolete parts. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/README.md | 64 +++++++++++----------------- scripts/populate_tox/populate_tox.py | 49 ++++++++++++--------- scripts/populate_tox/releases.jsonl | 10 ++--- scripts/populate_tox/tox.jinja | 17 +------- tox.ini | 43 +++++++------------ 5 files changed, 76 insertions(+), 107 deletions(-) diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index 8014acff3f..68e72a8ca7 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -16,8 +16,8 @@ The `populate_tox.py` script fills out the auto-generated part of that template. It does this by querying PyPI for each framework's package and its metadata and then determining which versions make sense to test to get good coverage. -The lowest supported and latest version of a framework are always tested, with -a number of releases in between: +By default, the lowest supported and latest version of a framework are always +tested, with a number of releases in between: - If the package has majors, we pick the highest version of each major. - If the package doesn't have multiple majors, we pick two versions in between lowest and highest. @@ -34,7 +34,8 @@ the main package (framework, library) to test with; any additional test dependencies, optionally gated behind specific conditions; and optionally the Python versions to test on. -Constraints are defined using the format specified below. The following sections describe each key. +Constraints are defined using the format specified below. The following sections +describe each key. ``` integration_name: { @@ -43,7 +44,7 @@ integration_name: { rule1: [package1, package2, ...], rule2: [package3, package4, ...], }, - "python": python_version_specifier, + "python": python_version_specifier | dict[package_version_specifier, python_version_specifier], "include": package_version_specifier, "integration_name": integration_name, "num_versions": int, @@ -102,15 +103,14 @@ Python versions, you can say: ... } ``` + This key is optional. ### `python` Sometimes, the whole test suite should only run on specific Python versions. -This can be achieved via the `python` key. - -There are two variants how to define the Python versions to run the test suite -on. +This can be achieved via the `python` key. There are two variants how to define +the Python versions to run the test suite on. If you want the test suite to only be run on specific Python versions, you can set `python` to a version specifier. For example, if you want AIOHTTP tests to @@ -142,18 +142,17 @@ say: The `python` key is optional, and when possible, it should be omitted. The script should automatically detect which Python versions the package supports. However, if a package has broken metadata or the SDK is explicitly not supporting some -packages on specific Python versions (because of, for example, broken context -vars), the `python` key can be used. +packages on specific Python versions, the `python` key can be used. ### `include` -Sometimes we only want to consider testing some specific versions of packages. -For example, the Starlite package has two alpha prereleases of version 2.0.0, but -we do not want to test these, since Starlite 2.0 was renamed to Litestar. +Sometimes we only want to test specific versions of packages. For example, the +Starlite package has two alpha prereleases of version 2.0.0, but we do not want +to test these, since Starlite 2.0 was renamed to Litestar. The value of the `include` key expects a version specifier defining which versions should be considered for testing. For example, since we only want to test -versions below 2.x in Starlite, we can use +versions below 2.x in Starlite, we can use: ```python "starlite": { @@ -178,13 +177,21 @@ be expressed like so: Sometimes, the name of the test suite doesn't match the name of the integration. For example, we have the `openai_base` and `openai_notiktoken` test suites, both -of which are actually testing the `openai` integration. If this is the case, you can use the `integration_name` key to define the name of the integration. If not provided, it will default to the name of the test suite. +of which are actually testing the `openai` integration. If this is the case, you +can use the `integration_name` key to define the name of the integration. If not +provided, it will default to the name of the test suite. -Linking an integration to a test suite allows the script to access integration configuration like for example the minimum version defined in `sentry_sdk/integrations/__init__.py`. +Linking an integration to a test suite allows the script to access integration +configuration like, for example, the minimum supported version defined in +`sentry_sdk/integrations/__init__.py`. ### `num_versions` -With this option you can tweak the default version picking behavior by specifying how many package versions should be tested. It accepts an integer equal to or greater than 2, as the oldest and latest supported versions will always be picked. Additionally, if there is a recent prerelease, it'll also always be picked (this doesn't count towards `num_versions`). +With this option you can tweak the default version picking behavior by specifying +how many package versions should be tested. It accepts an integer equal to or +greater than 2, as the oldest and latest supported versions will always be +picked. Additionally, if there is a recent prerelease, it'll also always be +picked (this doesn't count towards `num_versions`). ## How-Tos @@ -202,26 +209,3 @@ With this option you can tweak the default version picking behavior by specifyin `scripts/split_tox_gh_actions/split_tox_gh_actions.py`. 4. Add the `TESTPATH` for the test suite in `tox.jinja`'s `setenv` section. 5. Run `scripts/generate-test-files.sh` and commit the changes. - -### Migrate a test suite to populate_tox.py - -A handful of integration test suites are still hardcoded. The goal is to migrate -them all to `populate_tox.py` over time. - -1. Remove the integration from the `IGNORE` list in `populate_tox.py`. -2. Remove the hardcoded entries for the integration from the `envlist` and `deps` sections of `tox.jinja`. -3. Run `scripts/generate-test-files.sh`. -4. Run the test suite, either locally or by creating a PR. -5. Address any test failures that happen. - -You might have to introduce additional version bounds on the dependencies of the -package. Try to determine the source of the failure and address it. - -Common scenarios: -- An old version of the tested package installs a dependency without defining - an upper version bound on it. A new version of the dependency is installed that - is incompatible with the package. In this case you need to determine which - versions of the dependency don't contain the breaking change and restrict this - in `TEST_SUITE_CONFIG`. -- Tests are failing on an old Python version. In this case first double-check - whether we were even testing them on that version in the original `tox.ini`. diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 28a9fe8334..070ce82492 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -1,5 +1,7 @@ """ This script populates tox.ini automatically using release data from PyPI. + +See scripts/populate_tox/README.md for more info. """ import functools @@ -123,6 +125,7 @@ def _save_to_cache(package: str, version: Version, release: Optional[dict]) -> N releases_cache.write(json.dumps(_normalize_release(release)) + "\n") CACHE[_normalize_name(package)][str(version)] = release + CACHE[_normalize_name(package)][str(version)]["_accessed"] = True def _prefilter_releases( @@ -150,7 +153,8 @@ def _prefilter_releases( min_supported = Version(".".join(map(str, min_supported))) else: print( - f" {integration} doesn't have a minimum version defined in sentry_sdk/integrations/__init__.py. Consider defining one" + f" {integration} doesn't have a minimum version defined in " + f"sentry_sdk/integrations/__init__.py. Consider defining one" ) include_versions = None @@ -225,7 +229,8 @@ def get_supported_releases( Get a list of releases that are currently supported by the SDK. This takes into account a handful of parameters (Python support, the lowest - version we've defined for the framework, the date of the release). + supported version we've defined for the framework, optionally the date + of the release). We return the list of supported releases and optionally also the newest prerelease, if it should be tested (meaning it's for a version higher than @@ -272,9 +277,9 @@ def pick_releases_to_test( ) -> list[Version]: """Pick a handful of releases to test from a sorted list of supported releases.""" # If the package has majors (or major-like releases, even if they don't do - # semver), we want to make sure we're testing them all (unless there's too - # many). If not, we just pick the oldest, the newest, and a couple - # in between. + # semver), we want to make sure we're testing them all. If it doesn't have + # multiple majors, we just pick the oldest, the newest, and a couple of + # releases in between. # # If there is a relevant prerelease, also test that in addition to the above. num_versions = TEST_SUITE_CONFIG[integration].get("num_versions") @@ -342,7 +347,7 @@ def supported_python_versions( custom_supported_versions: Optional[ Union[SpecifierSet, dict[SpecifierSet, SpecifierSet]] ] = None, - version: Optional[Version] = None, + release_version: Optional[Version] = None, ) -> list[Version]: """ Get the intersection of Python versions supported by the package and the SDK. @@ -362,6 +367,12 @@ def supported_python_versions( on Python 3.7, so we can provide this as `custom_supported_versions`. The result of this function will then by the intersection of all three, i.e., [3.7]. + - The Python SDK supports Python 3.6-3.13. The package supports 3.5-3.8. + Additionally, we have a limitation in place to only test this framework on + Python 3.5 if the framework version is <2.0. `custom_supported_versions` + will contain this restriction, and `release_version` will contain the + version of the package we're currently looking at, to determine whether the + <2.0 restriction applies in this case. """ supported = [] @@ -377,11 +388,11 @@ def supported_python_versions( if curr in custom_supported_versions: supported.append(curr) - elif version is not None and isinstance( + elif release_version is not None and isinstance( custom_supported_versions, dict ): for v, py in custom_supported_versions.items(): - if version in v: + if release_version in v: if curr in py: supported.append(curr) break @@ -435,8 +446,10 @@ def _parse_python_versions_from_classifiers(classifiers: list[str]) -> list[Vers def determine_python_versions(pypi_data: dict) -> Union[SpecifierSet, list[Version]]: """ - Given data from PyPI's release endpoint, determine the Python versions supported by the package - from the Python version classifiers, when present, or from `requires_python` if there are no classifiers. + Determine the Python versions supported by the package from PyPI data. + + We're looking at Python version classifiers, if present, and + `requires_python` if there are no classifiers. """ try: classifiers = pypi_data["info"]["classifiers"] @@ -453,7 +466,7 @@ def determine_python_versions(pypi_data: dict) -> Union[SpecifierSet, list[Versi # We only use `requires_python` if there are no classifiers. This is because # `requires_python` doesn't tell us anything about the upper bound, which - # depends on when the release first came out + # implicitly depends on when the release first came out. try: requires_python = pypi_data["info"]["requires_python"] except (AttributeError, KeyError): @@ -666,7 +679,8 @@ def main(fail_on_changes: bool = False) -> None: # timestamp so that we don't fail CI on a PR just because a new package # version was released, leading to unrelated changes in tox.ini. print( - f"Since we're in fail_on_changes mode, we're only considering releases before the last tox.ini update at {last_updated.isoformat()}." + f"Since we're in fail_on_changes mode, we're only considering " + f"releases before the last tox.ini update at {last_updated.isoformat()}." ) global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION @@ -789,19 +803,16 @@ def main(fail_on_changes: bool = False) -> None: Please don't make manual changes to `tox.ini`. Instead, make the changes to the `tox.jinja` template and/or the `populate_tox.py` - script (as applicable) and regenerate the `tox.ini` file with: - - python -m venv toxgen.env - . toxgen.env/bin/activate - pip install -r scripts/populate_tox/requirements.txt - python scripts/populate_tox/populate_tox.py + script (as applicable) and regenerate the `tox.ini` file by + running scripts/generate-test-files.sh """ ) ) print("Done checking tox.ini. Looking good!") else: print( - "Done generating tox.ini. Make sure to also update the CI YAML files to reflect the new test targets." + "Done generating tox.ini. Make sure to also update the CI YAML " + "files to reflect the new test targets." ) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 814d0c15b9..5940eb28c1 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -56,7 +56,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.36", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.37", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -110,7 +110,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} @@ -126,13 +126,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.108.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.37.2", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.73.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.2", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} @@ -159,7 +159,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.4.8", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "3.0.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.49.2", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 98e14adb20..9aa942ebe4 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -24,17 +24,6 @@ envlist = {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent # === Integrations === - # General format is {pythonversion}-{integrationname}-v{frameworkversion} - # 1 blank line between different integrations - # Each framework version should only be mentioned once. I.e: - # {py3.7,py3.10}-django-v{3.2} - # {py3.10}-django-v{4.0} - # instead of: - # {py3.7}-django-v{3.2} - # {py3.7,py3.10}-django-v{3.2,4.0} - # - # At a minimum, we should test against at least the lowest - # and the latest supported version of a framework. # Asgi {py3.7,py3.12,py3.13}-asgi @@ -55,8 +44,7 @@ envlist = {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel # === Integrations - Auto-generated === - # These come from the populate_tox.py script. Eventually we should move all - # integration tests there. + # These come from the populate_tox.py script. {% for group, integrations in groups.items() %} # ~~~ {{ group }} ~~~ @@ -121,8 +109,7 @@ deps = potel: -e .[opentelemetry-experimental] # === Integrations - Auto-generated === - # These come from the populate_tox.py script. Eventually we should move all - # integration tests there. + # These come from the populate_tox.py script. {% for group, integrations in groups.items() %} # ~~~ {{ group }} ~~~ diff --git a/tox.ini b/tox.ini index f80c86b51e..7520a4cb1f 100644 --- a/tox.ini +++ b/tox.ini @@ -10,7 +10,7 @@ # The file (and all resulting CI YAMLs) then need to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-23T13:41:50.551689+00:00 +# Last generated: 2025-09-24T08:25:49.003370+00:00 [tox] requires = @@ -24,17 +24,6 @@ envlist = {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent # === Integrations === - # General format is {pythonversion}-{integrationname}-v{frameworkversion} - # 1 blank line between different integrations - # Each framework version should only be mentioned once. I.e: - # {py3.7,py3.10}-django-v{3.2} - # {py3.10}-django-v{4.0} - # instead of: - # {py3.7}-django-v{3.2} - # {py3.7,py3.10}-django-v{3.2,4.0} - # - # At a minimum, we should test against at least the lowest - # and the latest supported version of a framework. # Asgi {py3.7,py3.12,py3.13}-asgi @@ -55,8 +44,7 @@ envlist = {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel # === Integrations - Auto-generated === - # These come from the populate_tox.py script. Eventually we should move all - # integration tests there. + # These come from the populate_tox.py script. # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 @@ -80,12 +68,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.37.2 {py3.8,py3.11,py3.12}-openai-base-v1.73.0 - {py3.8,py3.12,py3.13}-openai-base-v1.108.2 + {py3.8,py3.12,py3.13}-openai-base-v1.109.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.108.2 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -93,19 +81,19 @@ envlist = {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.3.1 + {py3.10,py3.12,py3.13}-openai_agents-v0.3.2 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.0 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.1 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.36 + {py3.9,py3.12,py3.13}-boto3-v1.40.37 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.6,py3.7,py3.8}-chalice-v1.21.9 @@ -229,7 +217,7 @@ envlist = {py3.9,py3.12,py3.13}-rq-v2.6.0 {py3.8,py3.9}-spark-v3.0.3 - {py3.8,py3.10,py3.11}-spark-v3.5.6 + {py3.8,py3.10,py3.11}-spark-v3.5.7 {py3.9,py3.12,py3.13}-spark-v4.0.1 @@ -372,8 +360,7 @@ deps = potel: -e .[opentelemetry-experimental] # === Integrations - Auto-generated === - # These come from the populate_tox.py script. Eventually we should move all - # integration tests there. + # These come from the populate_tox.py script. # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 @@ -407,7 +394,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.37.2: openai==1.37.2 openai-base-v1.73.0: openai==1.73.0 - openai-base-v1.108.2: openai==1.108.2 + openai-base-v1.109.0: openai==1.109.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -416,7 +403,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.37.2: openai==1.37.2 openai-notiktoken-v1.73.0: openai==1.73.0 - openai-notiktoken-v1.108.2: openai==1.108.2 + openai-notiktoken-v1.109.0: openai==1.109.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.37.2: httpx<0.28 @@ -427,13 +414,13 @@ deps = openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.3.1: openai-agents==0.3.1 + openai_agents-v0.3.2: openai-agents==0.3.2 openai_agents: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 - huggingface_hub-v0.35.0: huggingface_hub==0.35.0 + huggingface_hub-v0.35.1: huggingface_hub==0.35.1 huggingface_hub: responses @@ -441,7 +428,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.36: boto3==1.40.36 + boto3-v1.40.37: boto3==1.40.37 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -608,7 +595,7 @@ deps = {py3.6,py3.7}-rq: fakeredis!=2.26.0 spark-v3.0.3: pyspark==3.0.3 - spark-v3.5.6: pyspark==3.5.6 + spark-v3.5.7: pyspark==3.5.7 spark-v4.0.1: pyspark==4.0.1 From d9811feda0867beca09a4b92fc142e2ea12702bb Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 24 Sep 2025 12:36:56 +0200 Subject: [PATCH 500/591] docs: Update contributing guidelines with instructions to run tests with tox (#4857) Updating instructions to run unit tests with a little background about `tox` and some commands you can use to run the tests. Closes https://github.com/getsentry/sentry-python/issues/4548 --- CONTRIBUTING.md | 28 +++++++++++++++++++++++----- 1 file changed, 23 insertions(+), 5 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 313910fe56..89330087d9 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -74,19 +74,37 @@ That's it. You should be ready to make changes, run tests, and make commits! If ## Running Tests -You can run all tests with the following command: +We test against a number of Python language and library versions, which are automatically generated and stored in the [tox.ini](tox.ini) file. The `envlist` defines the environments you can choose from when running tests, and correspond to package versions and environment variables. The `TESTPATH` environment variable, in turn, determines which tests are run. + +The tox CLI tool is required to run the tests locally. Follow [the installation instructions](https://tox.wiki/en/latest/installation.html) for tox. Dependencies are installed for you when you run the command below, but _you_ need to bring an appropriate Python interpreter. + +[Pyenv](https://github.com/pyenv/pyenv) is a cross-platform utility for managing Python versions. You can also use a conventional package manager, but not all versions may be distributed in the package manager of your choice. For macOS, Versions 3.8 and up can be installed with Homebrew. + +An environment consists of the Python major and minor version and the library name and version. The exception to the rule is that you can provide `common` instead of the library information. The environments tied to a specific library usually run the corresponding test suite, while `common` targets all tests but skips those that require uninstalled dependencies. + +To run Celery tests for version v5.5.3 of its Python library using a 3.12 interpreter, use ```bash -pytest tests/ +tox -p auto -o -e py3.12-celery-v5.5.3 ``` -If you would like to run the tests for a specific integration, use a command similar to the one below: +or to use the `common` environment, run ```bash -pytest -rs tests/integrations/flask/ # Replace "flask" with the specific integration you wish to test +tox -p auto -o -e py3.12-common ``` -**Hint:** Tests of integrations need additional dependencies. The switch `-rs` will show you why tests were skipped and what dependencies you need to install for the tests to run. (You can also consult the [tox.ini](tox.ini) file to see what dependencies are installed for each integration) +To select specific tests, you can forward arguments to `pytest` like so + +```bash +tox -p auto -o -e py3.12-celery-v5.5.3 -- -k test_transaction_events +``` + +In general, you use + +```bash +tox -p auto -o -e -- +``` ## Adding a New Integration From 4e1b96caa049ce3a9ec7ea0d18462337c771761a Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 24 Sep 2025 15:10:01 +0200 Subject: [PATCH 501/591] chore: Slim down test matrix (#4856) ### Description Only test the lowest and highest supported version of some of our smaller integrations. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .../workflows/test-integrations-graphql.yml | 2 +- scripts/populate_tox/config.py | 20 +++++ scripts/populate_tox/releases.jsonl | 30 -------- scripts/populate_tox/tox.jinja | 2 +- tox.ini | 75 +------------------ 5 files changed, 24 insertions(+), 105 deletions(-) diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index 81d809c94e..f1a43dc77d 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.8","3.9","3.10","3.11","3.12","3.13"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 70cecd60d9..34ae680fad 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -28,6 +28,7 @@ "*": ["fastapi", "flask", "httpx"], }, "python": ">=3.8", + "num_versions": 2, }, "arq": { "package": "arq", @@ -35,6 +36,7 @@ "*": ["async-timeout", "pytest-asyncio", "fakeredis>=2.2.0,<2.8"], "<=0.23": ["pydantic<2"], }, + "num_versions": 2, }, "asyncpg": { "package": "asyncpg", @@ -46,6 +48,7 @@ "beam": { "package": "apache-beam", "python": ">=3.7", + "num_versions": 2, }, "boto3": { "package": "boto3", @@ -71,9 +74,11 @@ "deps": { "*": ["pytest-chalice"], }, + "num_versions": 2, }, "clickhouse_driver": { "package": "clickhouse-driver", + "num_versions": 2, }, "cohere": { "package": "cohere", @@ -100,6 +105,7 @@ }, "dramatiq": { "package": "dramatiq", + "num_versions": 2, }, "falcon": { "package": "falcon", @@ -135,6 +141,7 @@ }, "gql": { "package": "gql[all]", + "num_versions": 2, }, "graphene": { "package": "graphene", @@ -171,6 +178,7 @@ }, "huey": { "package": "huey", + "num_versions": 2, }, "huggingface_hub": { "package": "huggingface_hub", @@ -204,6 +212,7 @@ }, "launchdarkly": { "package": "launchdarkly-server-sdk", + "num_versions": 2, }, "litestar": { "package": "litestar", @@ -214,6 +223,7 @@ }, "loguru": { "package": "loguru", + "num_versions": 2, }, "openai-base": { "package": "openai", @@ -242,9 +252,11 @@ }, "openfeature": { "package": "openfeature-sdk", + "num_versions": 2, }, "pure_eval": { "package": "pure_eval", + "num_versions": 2, }, "pymongo": { "package": "pymongo", @@ -271,6 +283,7 @@ ], "py3.8": ["taskgroup==0.0.0a4"], }, + "num_versions": 2, }, "ray": { "package": "ray", @@ -288,6 +301,7 @@ }, "redis_py_cluster_legacy": { "package": "redis-py-cluster", + "num_versions": 2, }, "requests": { "package": "requests", @@ -352,12 +366,14 @@ }, "python": "<=3.11", "include": "!=2.0.0a1,!=2.0.0a2", # these are not relevant as there will never be a stable 2.0 release (starlite continues as litestar) + "num_versions": 2, }, "statsig": { "package": "statsig", "deps": { "*": ["typing_extensions"], }, + "num_versions": 2, }, "strawberry": { "package": "strawberry-graphql[fastapi,flask]", @@ -365,6 +381,7 @@ "*": ["httpx"], "<=0.262.5": ["pydantic<2.11"], }, + "num_versions": 2, }, "tornado": { "package": "tornado", @@ -375,6 +392,7 @@ ], # https://github.com/tornadoweb/tornado/pull/3382 "py3.6": ["aiocontextvars"], }, + "num_versions": 2, }, "trytond": { "package": "trytond", @@ -385,8 +403,10 @@ }, "typer": { "package": "typer", + "num_versions": 2, }, "unleash": { "package": "UnleashClient", + "num_versions": 2, }, } diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 5940eb28c1..fa82e6b1cd 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -12,16 +12,12 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Flask", "requires_python": ">=3.9", "version": "3.1.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.16.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.17.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.18.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Quart", "requires_python": ">=3.9", "version": "0.20.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "", "version": "1.2.19", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "1.3.24", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7", "version": "1.4.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.43", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.1.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.2.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.12.15", "yanked": false}} @@ -37,17 +33,11 @@ {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.15.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.19.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.32.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.40.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.8", "version": "2.50.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.68.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.22", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": null, "version": "0.24.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.7", "version": "0.24.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.7", "version": "0.25.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.8", "version": "0.26.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.5.0", "version": "0.23.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.6.0", "version": "0.25.0", "yanked": false}} @@ -63,16 +53,12 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.6.0b1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.21.9", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": "", "version": "1.26.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.13.12", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.18.0", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.9.4", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.6", "version": "1.12.3", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.7", "version": "1.15.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": "", "version": "1.4.1", "yanked": false}} @@ -84,7 +70,6 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": null, "version": "3.5.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": "", "version": "3.3", "yanked": false}} @@ -104,8 +89,6 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.7.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.2.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.3.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} @@ -116,10 +99,8 @@ {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.7", "yanked": false}} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.10.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.9.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.17.0", "yanked": false}} @@ -136,7 +117,6 @@ {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": "", "version": "0.1.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} @@ -205,20 +185,12 @@ {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.8", "version": "0.38.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.9", "version": "0.48.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.49.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.50.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.58.4", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.61.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.64.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.8", "version": "0.233.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.257.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.282.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.7", "version": "6.2", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.8", "version": "6.4.2", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.4.27", "yanked": false}} @@ -231,6 +203,4 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.16.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.17.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.19.2", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 9aa942ebe4..f47d5db692 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -7,7 +7,7 @@ # by "scripts/populate_tox/populate_tox.py". Any changes to the file should # be made in the template (if you want to change a hardcoded part of the file) # or in the script (if you want to change the auto-generated part). -# The file (and all resulting CI YAMLs) then need to be regenerated via +# The file (and all resulting CI YAMLs) then needs to be regenerated via # "scripts/generate-test-files.sh". # # Last generated: {{ updated }} diff --git a/tox.ini b/tox.ini index 7520a4cb1f..e0295707ab 100644 --- a/tox.ini +++ b/tox.ini @@ -7,10 +7,10 @@ # by "scripts/populate_tox/populate_tox.py". Any changes to the file should # be made in the template (if you want to change a hardcoded part of the file) # or in the script (if you want to change the auto-generated part). -# The file (and all resulting CI YAMLs) then need to be regenerated via +# The file (and all resulting CI YAMLs) then needs to be regenerated via # "scripts/generate-test-files.sh". # -# Last generated: 2025-09-24T08:25:49.003370+00:00 +# Last generated: 2025-09-24T10:52:00.519014+00:00 [tox] requires = @@ -96,8 +96,6 @@ envlist = {py3.9,py3.12,py3.13}-boto3-v1.40.37 {py3.6,py3.7,py3.8}-chalice-v1.16.0 - {py3.6,py3.7,py3.8}-chalice-v1.21.9 - {py3.6,py3.8,py3.9}-chalice-v1.26.6 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -130,32 +128,23 @@ envlist = # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 - {py3.8,py3.12,py3.13}-launchdarkly-v9.9.0 - {py3.8,py3.12,py3.13}-launchdarkly-v9.10.0 {py3.9,py3.12,py3.13}-launchdarkly-v9.12.0 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.3 {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.58.4 - {py3.7,py3.12,py3.13}-statsig-v0.61.0 {py3.7,py3.12,py3.13}-statsig-v0.64.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 - {py3.8,py3.12,py3.13}-unleash-v6.1.0 - {py3.8,py3.12,py3.13}-unleash-v6.2.2 {py3.8,py3.12,py3.13}-unleash-v6.3.0 # ~~~ GraphQL ~~~ {py3.8,py3.10,py3.11}-ariadne-v0.20.1 - {py3.8,py3.11,py3.12}-ariadne-v0.22 - {py3.8,py3.11,py3.12}-ariadne-v0.24.0 {py3.9,py3.12,py3.13}-ariadne-v0.26.2 {py3.6,py3.9,py3.10}-gql-v3.4.1 - {py3.7,py3.11,py3.12}-gql-v3.5.3 {py3.9,py3.12,py3.13}-gql-v4.0.0 {py3.9,py3.12,py3.13}-gql-v4.2.0b0 @@ -163,8 +152,6 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.8,py3.11,py3.12}-strawberry-v0.233.3 - {py3.9,py3.12,py3.13}-strawberry-v0.257.0 {py3.9,py3.12,py3.13}-strawberry-v0.282.0 @@ -185,13 +172,9 @@ envlist = # ~~~ Tasks ~~~ {py3.7,py3.9,py3.10}-arq-v0.23 - {py3.7,py3.10,py3.11}-arq-v0.24.0 - {py3.7,py3.10,py3.11}-arq-v0.25.0 {py3.8,py3.11,py3.12}-arq-v0.26.3 {py3.7}-beam-v2.14.0 - {py3.7,py3.8}-beam-v2.32.0 - {py3.8,py3.10,py3.11}-beam-v2.50.0 {py3.9,py3.12,py3.13}-beam-v2.68.0 {py3.6,py3.7,py3.8}-celery-v4.4.7 @@ -199,13 +182,9 @@ envlist = {py3.8,py3.12,py3.13}-celery-v5.6.0b1 {py3.6,py3.7}-dramatiq-v1.9.0 - {py3.6,py3.8,py3.9}-dramatiq-v1.12.3 - {py3.7,py3.10,py3.11}-dramatiq-v1.15.0 {py3.9,py3.12,py3.13}-dramatiq-v1.18.0 {py3.6,py3.7}-huey-v2.1.3 - {py3.6,py3.7}-huey-v2.2.0 - {py3.6,py3.7}-huey-v2.3.2 {py3.6,py3.11,py3.12}-huey-v2.5.3 {py3.9,py3.10}-ray-v2.7.2 @@ -268,8 +247,6 @@ envlist = {py3.6,py3.10,py3.11}-pyramid-v2.0.2 {py3.7,py3.9,py3.10}-quart-v0.16.3 - {py3.7,py3.9,py3.10}-quart-v0.17.0 - {py3.7,py3.10,py3.11}-quart-v0.18.4 {py3.9,py3.12,py3.13}-quart-v0.20.0 {py3.6}-sanic-v0.8.3 @@ -278,13 +255,9 @@ envlist = {py3.9,py3.12,py3.13}-sanic-v25.3.0 {py3.8,py3.10,py3.11}-starlite-v1.48.1 - {py3.8,py3.10,py3.11}-starlite-v1.49.0 - {py3.8,py3.10,py3.11}-starlite-v1.50.2 {py3.8,py3.10,py3.11}-starlite-v1.51.16 {py3.6,py3.7,py3.8}-tornado-v6.0.4 - {py3.7,py3.9,py3.10}-tornado-v6.2 - {py3.8,py3.10,py3.11}-tornado-v6.4.2 {py3.9,py3.12,py3.13}-tornado-v6.5.2 @@ -292,7 +265,6 @@ envlist = {py3.6,py3.12,py3.13}-loguru-v0.7.3 {py3.6,py3.7,py3.8}-pure_eval-v0.0.3 - {py3.6,py3.8,py3.9}-pure_eval-v0.1.1 {py3.7,py3.12,py3.13}-pure_eval-v0.2.3 {py3.6}-trytond-v4.6.22 @@ -302,8 +274,6 @@ envlist = {py3.9,py3.12,py3.13}-trytond-v7.6.7 {py3.7,py3.12,py3.13}-typer-v0.15.4 - {py3.7,py3.12,py3.13}-typer-v0.16.1 - {py3.7,py3.12,py3.13}-typer-v0.17.5 {py3.8,py3.12,py3.13}-typer-v0.19.2 @@ -432,8 +402,6 @@ deps = {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 - chalice-v1.21.9: chalice==1.21.9 - chalice-v1.26.6: chalice==1.26.6 chalice-v1.32.0: chalice==1.32.0 chalice: pytest-chalice @@ -474,36 +442,27 @@ deps = # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 - launchdarkly-v9.9.0: launchdarkly-server-sdk==9.9.0 - launchdarkly-v9.10.0: launchdarkly-server-sdk==9.10.0 launchdarkly-v9.12.0: launchdarkly-server-sdk==9.12.0 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.3: openfeature-sdk==0.8.3 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.58.4: statsig==0.58.4 - statsig-v0.61.0: statsig==0.61.0 statsig-v0.64.0: statsig==0.64.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 - unleash-v6.1.0: UnleashClient==6.1.0 - unleash-v6.2.2: UnleashClient==6.2.2 unleash-v6.3.0: UnleashClient==6.3.0 # ~~~ GraphQL ~~~ ariadne-v0.20.1: ariadne==0.20.1 - ariadne-v0.22: ariadne==0.22 - ariadne-v0.24.0: ariadne==0.24.0 ariadne-v0.26.2: ariadne==0.26.2 ariadne: fastapi ariadne: flask ariadne: httpx gql-v3.4.1: gql[all]==3.4.1 - gql-v3.5.3: gql[all]==3.5.3 gql-v4.0.0: gql[all]==4.0.0 gql-v4.2.0b0: gql[all]==4.2.0b0 @@ -516,13 +475,9 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.233.3: strawberry-graphql[fastapi,flask]==0.233.3 - strawberry-v0.257.0: strawberry-graphql[fastapi,flask]==0.257.0 strawberry-v0.282.0: strawberry-graphql[fastapi,flask]==0.282.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 - strawberry-v0.233.3: pydantic<2.11 - strawberry-v0.257.0: pydantic<2.11 # ~~~ Network ~~~ @@ -551,8 +506,6 @@ deps = # ~~~ Tasks ~~~ arq-v0.23: arq==0.23 - arq-v0.24.0: arq==0.24.0 - arq-v0.25.0: arq==0.25.0 arq-v0.26.3: arq==0.26.3 arq: async-timeout arq: pytest-asyncio @@ -560,8 +513,6 @@ deps = arq-v0.23: pydantic<2 beam-v2.14.0: apache-beam==2.14.0 - beam-v2.32.0: apache-beam==2.32.0 - beam-v2.50.0: apache-beam==2.50.0 beam-v2.68.0: apache-beam==2.68.0 celery-v4.4.7: celery==4.4.7 @@ -572,13 +523,9 @@ deps = {py3.7}-celery: importlib-metadata<5.0 dramatiq-v1.9.0: dramatiq==1.9.0 - dramatiq-v1.12.3: dramatiq==1.12.3 - dramatiq-v1.15.0: dramatiq==1.15.0 dramatiq-v1.18.0: dramatiq==1.18.0 huey-v2.1.3: huey==2.1.3 - huey-v2.2.0: huey==2.2.0 - huey-v2.3.2: huey==2.3.2 huey-v2.5.3: huey==2.5.3 ray-v2.7.2: ray==2.7.2 @@ -701,8 +648,6 @@ deps = pyramid: werkzeug<2.1.0 quart-v0.16.3: quart==0.16.3 - quart-v0.17.0: quart==0.17.0 - quart-v0.18.4: quart==0.18.4 quart-v0.20.0: quart==0.20.0 quart: quart-auth quart: pytest-asyncio @@ -712,14 +657,6 @@ deps = quart-v0.16.3: jinja2<3.1.0 quart-v0.16.3: Werkzeug<2.3.0 quart-v0.16.3: hypercorn<0.15.0 - quart-v0.17.0: blinker<1.6 - quart-v0.17.0: jinja2<3.1.0 - quart-v0.17.0: Werkzeug<2.3.0 - quart-v0.17.0: hypercorn<0.15.0 - quart-v0.18.4: blinker<1.6 - quart-v0.18.4: jinja2<3.1.0 - quart-v0.18.4: Werkzeug<2.3.0 - quart-v0.18.4: hypercorn<0.15.0 {py3.8}-quart: taskgroup==0.0.0a4 sanic-v0.8.3: sanic==0.8.3 @@ -734,8 +671,6 @@ deps = {py3.8}-sanic: tracerite<1.1.2 starlite-v1.48.1: starlite==1.48.1 - starlite-v1.49.0: starlite==1.49.0 - starlite-v1.50.2: starlite==1.50.2 starlite-v1.51.16: starlite==1.51.16 starlite: pytest-asyncio starlite: python-multipart @@ -745,12 +680,9 @@ deps = starlite: httpx<0.28 tornado-v6.0.4: tornado==6.0.4 - tornado-v6.2: tornado==6.2 - tornado-v6.4.2: tornado==6.4.2 tornado-v6.5.2: tornado==6.5.2 tornado: pytest tornado-v6.0.4: pytest<8.2 - tornado-v6.2: pytest<8.2 {py3.6}-tornado: aiocontextvars @@ -758,7 +690,6 @@ deps = loguru-v0.7.3: loguru==0.7.3 pure_eval-v0.0.3: pure_eval==0.0.3 - pure_eval-v0.1.1: pure_eval==0.1.1 pure_eval-v0.2.3: pure_eval==0.2.3 trytond-v4.6.22: trytond==4.6.22 @@ -771,8 +702,6 @@ deps = trytond-v4.8.18: werkzeug<1.0 typer-v0.15.4: typer==0.15.4 - typer-v0.16.1: typer==0.16.1 - typer-v0.17.5: typer==0.17.5 typer-v0.19.2: typer==0.19.2 From b7780bab8e6b4e288f8782d9c8fcc9d8b9e40bad Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 24 Sep 2025 16:44:13 +0200 Subject: [PATCH 502/591] feat(toxgen): Remove timestamp prone to merge conflicts (#4860) ### Description Most of the merge conflicts concerning `tox.ini` come from the timestamp of the last update. If two PRs regenerate `tox.ini`, they'll be guaranteed to have a merge conflict that needs to be resolved by hand. This is time consuming and unnecessary. We still need some way of figuring out when the file was last updated. This is important for checking unwanted edits to `tox.ini` (instead of to the template it's generated from). In order to check for them, we need to regenerate the file on the current branch and compare it to the committed version. However, if there have been new package versions in the meantime, these would then pop up in the diff. One solution is to regenerate the file ignoring any new releases after the last update, which is what we were using the last updated timestamp for. In this PR, we simply figure out the timestamp from `git` history, removing the need for storing it explicitly anywhere. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/ci.yml | 3 ++ scripts/populate_tox/populate_tox.py | 44 ++++++++++++---------------- scripts/populate_tox/releases.jsonl | 2 +- scripts/populate_tox/tox.jinja | 2 -- tox.ini | 10 +++---- 5 files changed, 27 insertions(+), 34 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 67b4fd3546..50ab1d39ce 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -40,6 +40,9 @@ jobs: steps: - uses: actions/checkout@v5.0.0 + with: + ref: ${{ github.event.pull_request.head.sha }} + fetch-depth: 0 - uses: actions/setup-python@v6 with: python-version: 3.12 diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 070ce82492..f13fa6d002 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -8,6 +8,7 @@ import hashlib import json import os +import subprocess import sys import time from bisect import bisect_left @@ -509,9 +510,7 @@ def _render_dependencies(integration: str, releases: list[Version]) -> list[str] return rendered -def write_tox_file( - packages: dict, update_timestamp: bool, last_updated: datetime -) -> None: +def write_tox_file(packages: dict) -> None: template = ENV.get_template("tox.jinja") context = {"groups": {}} @@ -530,11 +529,6 @@ def write_tox_file( } ) - if update_timestamp: - context["updated"] = datetime.now(tz=timezone.utc).isoformat() - else: - context["updated"] = last_updated.isoformat() - rendered = template.render(context) with open(TOX_FILE, "w") as file: @@ -623,19 +617,21 @@ def get_file_hash() -> str: def get_last_updated() -> Optional[datetime]: - timestamp = None - - with open(TOX_FILE, "r") as f: - for line in f: - if line.startswith("# Last generated:"): - timestamp = datetime.fromisoformat(line.strip().split()[-1]) - break - - if timestamp is None: - print( - "Failed to find out when tox.ini was last generated; the timestamp seems to be missing from the file." - ) - + repo_root = subprocess.run( + ["git", "rev-parse", "--show-toplevel"], + capture_output=True, + text=True, + ).stdout.strip() + tox_ini_path = Path(repo_root) / "tox.ini" + + timestamp = subprocess.run( + ["git", "log", "-1", "--pretty=%ct", str(tox_ini_path)], + capture_output=True, + text=True, + ).stdout.strip() + + timestamp = datetime.fromtimestamp(int(timestamp), timezone.utc) + print(f"Last committed tox.ini update: {timestamp}") return timestamp @@ -675,7 +671,7 @@ def main(fail_on_changes: bool = False) -> None: print(f"Running in {'fail_on_changes' if fail_on_changes else 'normal'} mode.") last_updated = get_last_updated() if fail_on_changes: - # We need to make the script ignore any new releases after the `last_updated` + # We need to make the script ignore any new releases after the last updated # timestamp so that we don't fail CI on a PR just because a new package # version was released, leading to unrelated changes in tox.ini. print( @@ -769,9 +765,7 @@ def main(fail_on_changes: bool = False) -> None: if fail_on_changes: old_file_hash = get_file_hash() - write_tox_file( - packages, update_timestamp=not fail_on_changes, last_updated=last_updated - ) + write_tox_file(packages) # Sort the release cache file releases = [] diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index fa82e6b1cd..96eb2da8e0 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -107,7 +107,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.37.2", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.73.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index f47d5db692..039b93c42e 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -9,8 +9,6 @@ # or in the script (if you want to change the auto-generated part). # The file (and all resulting CI YAMLs) then needs to be regenerated via # "scripts/generate-test-files.sh". -# -# Last generated: {{ updated }} [tox] requires = diff --git a/tox.ini b/tox.ini index e0295707ab..8e4c6e5a36 100644 --- a/tox.ini +++ b/tox.ini @@ -9,8 +9,6 @@ # or in the script (if you want to change the auto-generated part). # The file (and all resulting CI YAMLs) then needs to be regenerated via # "scripts/generate-test-files.sh". -# -# Last generated: 2025-09-24T10:52:00.519014+00:00 [tox] requires = @@ -68,12 +66,12 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.11,py3.12}-openai-base-v1.37.2 {py3.8,py3.11,py3.12}-openai-base-v1.73.0 - {py3.8,py3.12,py3.13}-openai-base-v1.109.0 + {py3.8,py3.12,py3.13}-openai-base-v1.109.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 - {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 {py3.9,py3.12,py3.13}-langgraph-v0.6.7 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 @@ -364,7 +362,7 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.37.2: openai==1.37.2 openai-base-v1.73.0: openai==1.73.0 - openai-base-v1.109.0: openai==1.109.0 + openai-base-v1.109.1: openai==1.109.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 @@ -373,7 +371,7 @@ deps = openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.37.2: openai==1.37.2 openai-notiktoken-v1.73.0: openai==1.73.0 - openai-notiktoken-v1.109.0: openai==1.109.0 + openai-notiktoken-v1.109.1: openai==1.109.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai-notiktoken-v1.37.2: httpx<0.28 From 6325924c2b8b0e4d7facdc49dd9aa7bdade9a08c Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Wed, 24 Sep 2025 17:35:11 +0200 Subject: [PATCH 503/591] Don't swallow userland exceptions in openai (#4861) #### Issues * resolves: #4853 * resolves: PY-1865 --- sentry_sdk/integrations/openai.py | 18 ++++++++++-------- 1 file changed, 10 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 4d72ec366c..e8b3b30ab2 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -282,9 +282,9 @@ def _set_output_data(span, response, kwargs, integration, finish_span=True): def new_iterator(): # type: () -> Iterator[ChatCompletionChunk] - with capture_internal_exceptions(): - count_tokens_manually = True - for x in old_iterator: + count_tokens_manually = True + for x in old_iterator: + with capture_internal_exceptions(): # OpenAI chat completion API if hasattr(x, "choices"): choice_index = 0 @@ -315,8 +315,9 @@ def new_iterator(): ) count_tokens_manually = False - yield x + yield x + with capture_internal_exceptions(): if len(data_buf) > 0: all_responses = ["".join(chunk) for chunk in data_buf] if should_send_default_pii() and integration.include_prompts: @@ -337,9 +338,9 @@ def new_iterator(): async def new_iterator_async(): # type: () -> AsyncIterator[ChatCompletionChunk] - with capture_internal_exceptions(): - count_tokens_manually = True - async for x in old_iterator: + count_tokens_manually = True + async for x in old_iterator: + with capture_internal_exceptions(): # OpenAI chat completion API if hasattr(x, "choices"): choice_index = 0 @@ -370,8 +371,9 @@ async def new_iterator_async(): ) count_tokens_manually = False - yield x + yield x + with capture_internal_exceptions(): if len(data_buf) > 0: all_responses = ["".join(chunk) for chunk in data_buf] if should_send_default_pii() and integration.include_prompts: From 01fc851b7a805816fddfd252d2cd93dfd0484b69 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 25 Sep 2025 08:51:53 +0000 Subject: [PATCH 504/591] release: 2.39.0 --- CHANGELOG.md | 31 +++++++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 34 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 7abbed7218..2c159f8afe 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,36 @@ # Changelog +## 2.39.0 + +### Various fixes & improvements + +- Don't swallow userland exceptions in openai (#4861) by @sl0thentr0py +- feat(toxgen): Remove timestamp prone to merge conflicts (#4860) by @sentrivana +- chore: Slim down test matrix (#4856) by @sentrivana +- docs: Update contributing guidelines with instructions to run tests with tox (#4857) by @alexander-alderman-webb +- chore: Clean up toxgen (#4855) by @sentrivana +- feat(toxgen): Cache release data (#4835) by @sentrivana +- fix(AI): Make agents integrations set the span status in case of error (#4820) by @antonpirker +- fix(langchain): don't record tool call output if not include_prompt / should_send_default_pii (#4836) by @shellmayr +- build(deps): bump actions/setup-python from 5 to 6 (#4830) by @dependabot +- fix: Pin shibuya (#4839) by @sentrivana +- Use weakref in dedupe where possible (#4834) by @sl0thentr0py +- tests: Remove old system for checking for regressions (#4826) by @sentrivana +- ci: 🤖 Update test matrix with new releases (09/22) (#4831) by @github-actions +- tests: Move requests under toxgen (#4825) by @sentrivana +- tests: Move rq under toxgen (#4818) by @sentrivana +- tests: Move redis under toxgen (#4824) by @sentrivana +- test(spark): Improve `test_spark` speed (#4822) by @mgaligniana +- tests: Move httpx under toxgen (#4780) by @sentrivana +- tests: Move ray under toxgen (#4810) by @sentrivana +- tests: Move sanic under toxgen & add option to control how many releases to pick (#4767) by @sentrivana +- ci: 🤖 Update test matrix with new releases (09/18) (#4816) by @github-actions +- tests: Move pure_eval under toxgen (#4815) by @sentrivana +- fix(Django): Avoid evaluating complex Django object in span.data/span.attributes (#4804) by @antonpirker +- fix: Fix links to files in toxgen PRs (#4807) by @sentrivana + +_Plus 5 more_ + ## 2.38.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 061b2bdfc8..292e0971e2 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.38.0" +release = "2.39.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 606cf804b6..5b8e840f13 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1333,4 +1333,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.38.0" +VERSION = "2.39.0" diff --git a/setup.py b/setup.py index 58101aa65f..7119e20e90 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.38.0", + version="2.39.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From ff527644c5446674c0870cb830a122f7b8fe3ac2 Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Thu, 25 Sep 2025 10:59:46 +0200 Subject: [PATCH 505/591] Updated changelog --- CHANGELOG.md | 52 ++++++++++++++++++++++++++-------------------------- 1 file changed, 26 insertions(+), 26 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2c159f8afe..b9095279f7 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,32 +4,32 @@ ### Various fixes & improvements -- Don't swallow userland exceptions in openai (#4861) by @sl0thentr0py -- feat(toxgen): Remove timestamp prone to merge conflicts (#4860) by @sentrivana -- chore: Slim down test matrix (#4856) by @sentrivana -- docs: Update contributing guidelines with instructions to run tests with tox (#4857) by @alexander-alderman-webb -- chore: Clean up toxgen (#4855) by @sentrivana -- feat(toxgen): Cache release data (#4835) by @sentrivana -- fix(AI): Make agents integrations set the span status in case of error (#4820) by @antonpirker -- fix(langchain): don't record tool call output if not include_prompt / should_send_default_pii (#4836) by @shellmayr -- build(deps): bump actions/setup-python from 5 to 6 (#4830) by @dependabot -- fix: Pin shibuya (#4839) by @sentrivana -- Use weakref in dedupe where possible (#4834) by @sl0thentr0py -- tests: Remove old system for checking for regressions (#4826) by @sentrivana -- ci: 🤖 Update test matrix with new releases (09/22) (#4831) by @github-actions -- tests: Move requests under toxgen (#4825) by @sentrivana -- tests: Move rq under toxgen (#4818) by @sentrivana -- tests: Move redis under toxgen (#4824) by @sentrivana -- test(spark): Improve `test_spark` speed (#4822) by @mgaligniana -- tests: Move httpx under toxgen (#4780) by @sentrivana -- tests: Move ray under toxgen (#4810) by @sentrivana -- tests: Move sanic under toxgen & add option to control how many releases to pick (#4767) by @sentrivana -- ci: 🤖 Update test matrix with new releases (09/18) (#4816) by @github-actions -- tests: Move pure_eval under toxgen (#4815) by @sentrivana -- fix(Django): Avoid evaluating complex Django object in span.data/span.attributes (#4804) by @antonpirker -- fix: Fix links to files in toxgen PRs (#4807) by @sentrivana - -_Plus 5 more_ +- Feat(toxgen): Cache release data (#4835) by @sentrivana +- Feat(toxgen): Remove timestamp prone to merge conflicts (#4860) by @sentrivana +- Fix: Pin shibuya (#4839) by @sentrivana +- Fix: Links to files in toxgen PRs (#4807) by @sentrivana +- Fix(AI): Make agents integrations set the span status in case of error (#4820) by @antonpirker +- Fix(dedupe): Use weakref in dedupe where possible (#4834) by @sl0thentr0py +- Fix(Django): Avoid evaluating complex Django object in span.data/span.attributes (#4804) by @antonpirker +- Fix(Dangchain): don't record tool call output if not include_prompt / should_send_default_pii (#4836) by @shellmayr +- Fix(OpenAI): Don't swallow userland exceptions in openai (#4861) by @sl0thentr0py +- Docs: Update contributing guidelines with instructions to run tests with tox (#4857) by @alexander-alderman-webb +- Test(Spark): Improve `test_spark` speed (#4822) by @mgaligniana +- Tests: Remove old system for checking for regressions (#4826) by @sentrivana +- Tests: Move requests under toxgen (#4825) by @sentrivana +- Tests: Move rq under toxgen (#4818) by @sentrivana +- Tests: Move redis under toxgen (#4824) by @sentrivana +- Tests: Move httpx under toxgen (#4780) by @sentrivana +- Tests: Move ray under toxgen (#4810) by @sentrivana +- Tests: Move sanic under toxgen & add option to control how many releases to pick (#4767) by @sentrivana +- Tests: Move pure_eval under toxgen (#4815) by @sentrivana +- Chore: Clean up toxgen (#4855) by @sentrivana +- Chore: Slim down test matrix (#4856) by @sentrivana +- Note: This is my last release. So long, and thanks for all the fish! by @antonpirker +- Build(deps): bump actions/setup-python from 5 to 6 (#4830) by @dependabot +- CI: 🤖 Update test matrix with new releases (09/18) (#4816) by @github-actions +- CI: 🤖 Update test matrix with new releases (09/22) (#4831) by @github-actions + ## 2.38.0 From 0f99a91e7b5103382c8198a7fcae70ab2933d85d Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 25 Sep 2025 11:05:18 +0200 Subject: [PATCH 506/591] Update CHANGELOG.md --- CHANGELOG.md | 22 +++------------------- 1 file changed, 3 insertions(+), 19 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index b9095279f7..941aec99e6 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,31 +4,15 @@ ### Various fixes & improvements -- Feat(toxgen): Cache release data (#4835) by @sentrivana -- Feat(toxgen): Remove timestamp prone to merge conflicts (#4860) by @sentrivana -- Fix: Pin shibuya (#4839) by @sentrivana -- Fix: Links to files in toxgen PRs (#4807) by @sentrivana - Fix(AI): Make agents integrations set the span status in case of error (#4820) by @antonpirker - Fix(dedupe): Use weakref in dedupe where possible (#4834) by @sl0thentr0py - Fix(Django): Avoid evaluating complex Django object in span.data/span.attributes (#4804) by @antonpirker -- Fix(Dangchain): don't record tool call output if not include_prompt / should_send_default_pii (#4836) by @shellmayr +- Fix(Langchain): Don't record tool call output if not include_prompt / should_send_default_pii (#4836) by @shellmayr - Fix(OpenAI): Don't swallow userland exceptions in openai (#4861) by @sl0thentr0py - Docs: Update contributing guidelines with instructions to run tests with tox (#4857) by @alexander-alderman-webb - Test(Spark): Improve `test_spark` speed (#4822) by @mgaligniana -- Tests: Remove old system for checking for regressions (#4826) by @sentrivana -- Tests: Move requests under toxgen (#4825) by @sentrivana -- Tests: Move rq under toxgen (#4818) by @sentrivana -- Tests: Move redis under toxgen (#4824) by @sentrivana -- Tests: Move httpx under toxgen (#4780) by @sentrivana -- Tests: Move ray under toxgen (#4810) by @sentrivana -- Tests: Move sanic under toxgen & add option to control how many releases to pick (#4767) by @sentrivana -- Tests: Move pure_eval under toxgen (#4815) by @sentrivana -- Chore: Clean up toxgen (#4855) by @sentrivana -- Chore: Slim down test matrix (#4856) by @sentrivana -- Note: This is my last release. So long, and thanks for all the fish! by @antonpirker -- Build(deps): bump actions/setup-python from 5 to 6 (#4830) by @dependabot -- CI: 🤖 Update test matrix with new releases (09/18) (#4816) by @github-actions -- CI: 🤖 Update test matrix with new releases (09/22) (#4831) by @github-actions + +Note: This is my last release. So long, and thanks for all the fish! by @antonpirker ## 2.38.0 From 3a43dc9351b38079290851d13c625e39b96a65cc Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 25 Sep 2025 12:29:00 +0200 Subject: [PATCH 507/591] feat(toxgen): Generate `TESTPATH` for integrated test suites (#4863) ### Description Make toxgen generate the `TESTPATH` env var for integrated test suites, removing a manual step when adding a new test suite. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4536 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/README.md | 3 +- scripts/populate_tox/populate_tox.py | 18 ++++++++- scripts/populate_tox/releases.jsonl | 2 +- scripts/populate_tox/tox.jinja | 59 +++------------------------- tox.ini | 25 ++++++------ 5 files changed, 39 insertions(+), 68 deletions(-) diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index 68e72a8ca7..9bdb3567b8 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -207,5 +207,4 @@ picked (this doesn't count towards `num_versions`). "Defining constraints" section for the format. 3. Add the integration to one of the groups in the `GROUPS` dictionary in `scripts/split_tox_gh_actions/split_tox_gh_actions.py`. -4. Add the `TESTPATH` for the test suite in `tox.jinja`'s `setenv` section. -5. Run `scripts/generate-test-files.sh` and commit the changes. +4. Run `scripts/generate-test-files.sh` and commit the changes. diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index f13fa6d002..4bb4d78231 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -513,7 +513,11 @@ def _render_dependencies(integration: str, releases: list[Version]) -> list[str] def write_tox_file(packages: dict) -> None: template = ENV.get_template("tox.jinja") - context = {"groups": {}} + context = { + "groups": {}, + "testpaths": [], + } + for group, integrations in packages.items(): context["groups"][group] = [] for integration in integrations: @@ -528,6 +532,14 @@ def write_tox_file(packages: dict) -> None: ), } ) + context["testpaths"].append( + ( + integration["name"], + f"tests/integrations/{integration['integration_name']}", + ) + ) + + context["testpaths"].sort() rendered = template.render(context) @@ -759,6 +771,10 @@ def main(fail_on_changes: bool = False) -> None: "package": package, "extra": extra, "releases": test_releases, + "integration_name": TEST_SUITE_CONFIG[integration].get( + "integration_name" + ) + or integration, } ) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 96eb2da8e0..c68fe87734 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.37", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.38", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 039b93c42e..39d2406b84 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -136,69 +136,22 @@ setenv = django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 + # TESTPATH definitions for test suites not managed by toxgen common: TESTPATH=tests gevent: TESTPATH=tests - aiohttp: TESTPATH=tests/integrations/aiohttp - anthropic: TESTPATH=tests/integrations/anthropic - ariadne: TESTPATH=tests/integrations/ariadne - arq: TESTPATH=tests/integrations/arq asgi: TESTPATH=tests/integrations/asgi - asyncpg: TESTPATH=tests/integrations/asyncpg aws_lambda: TESTPATH=tests/integrations/aws_lambda - beam: TESTPATH=tests/integrations/beam - boto3: TESTPATH=tests/integrations/boto3 - bottle: TESTPATH=tests/integrations/bottle - celery: TESTPATH=tests/integrations/celery - chalice: TESTPATH=tests/integrations/chalice - clickhouse_driver: TESTPATH=tests/integrations/clickhouse_driver - cohere: TESTPATH=tests/integrations/cohere cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context - django: TESTPATH=tests/integrations/django - dramatiq: TESTPATH=tests/integrations/dramatiq - falcon: TESTPATH=tests/integrations/falcon - fastapi: TESTPATH=tests/integrations/fastapi - flask: TESTPATH=tests/integrations/flask gcp: TESTPATH=tests/integrations/gcp - gql: TESTPATH=tests/integrations/gql - graphene: TESTPATH=tests/integrations/graphene - grpc: TESTPATH=tests/integrations/grpc - httpx: TESTPATH=tests/integrations/httpx - huey: TESTPATH=tests/integrations/huey - huggingface_hub: TESTPATH=tests/integrations/huggingface_hub - langchain-base: TESTPATH=tests/integrations/langchain - langchain-notiktoken: TESTPATH=tests/integrations/langchain - langgraph: TESTPATH=tests/integrations/langgraph - launchdarkly: TESTPATH=tests/integrations/launchdarkly - litestar: TESTPATH=tests/integrations/litestar - loguru: TESTPATH=tests/integrations/loguru - openai-base: TESTPATH=tests/integrations/openai - openai-notiktoken: TESTPATH=tests/integrations/openai - openai_agents: TESTPATH=tests/integrations/openai_agents - openfeature: TESTPATH=tests/integrations/openfeature opentelemetry: TESTPATH=tests/integrations/opentelemetry potel: TESTPATH=tests/integrations/opentelemetry - pure_eval: TESTPATH=tests/integrations/pure_eval - pymongo: TESTPATH=tests/integrations/pymongo - pyramid: TESTPATH=tests/integrations/pyramid - quart: TESTPATH=tests/integrations/quart - ray: TESTPATH=tests/integrations/ray - redis: TESTPATH=tests/integrations/redis - redis_py_cluster_legacy: TESTPATH=tests/integrations/redis_py_cluster_legacy - requests: TESTPATH=tests/integrations/requests - rq: TESTPATH=tests/integrations/rq - sanic: TESTPATH=tests/integrations/sanic - spark: TESTPATH=tests/integrations/spark - sqlalchemy: TESTPATH=tests/integrations/sqlalchemy - starlette: TESTPATH=tests/integrations/starlette - starlite: TESTPATH=tests/integrations/starlite - statsig: TESTPATH=tests/integrations/statsig - strawberry: TESTPATH=tests/integrations/strawberry - tornado: TESTPATH=tests/integrations/tornado - trytond: TESTPATH=tests/integrations/trytond - typer: TESTPATH=tests/integrations/typer - unleash: TESTPATH=tests/integrations/unleash socket: TESTPATH=tests/integrations/socket + # These TESTPATH definitions are auto-generated by toxgen + {% for integration, testpath in testpaths %} + {{ integration }}: TESTPATH={{ testpath }} + {% endfor %} + passenv = SENTRY_PYTHON_TEST_POSTGRES_HOST SENTRY_PYTHON_TEST_POSTGRES_USER diff --git a/tox.ini b/tox.ini index 8e4c6e5a36..b993397389 100644 --- a/tox.ini +++ b/tox.ini @@ -91,7 +91,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.37 + {py3.9,py3.12,py3.13}-boto3-v1.40.38 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -396,7 +396,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.37: boto3==1.40.37 + boto3-v1.40.38: boto3==1.40.38 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -713,15 +713,23 @@ setenv = django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 + # TESTPATH definitions for test suites not managed by toxgen common: TESTPATH=tests gevent: TESTPATH=tests + asgi: TESTPATH=tests/integrations/asgi + aws_lambda: TESTPATH=tests/integrations/aws_lambda + cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context + gcp: TESTPATH=tests/integrations/gcp + opentelemetry: TESTPATH=tests/integrations/opentelemetry + potel: TESTPATH=tests/integrations/opentelemetry + socket: TESTPATH=tests/integrations/socket + + # These TESTPATH definitions are auto-generated by toxgen aiohttp: TESTPATH=tests/integrations/aiohttp anthropic: TESTPATH=tests/integrations/anthropic ariadne: TESTPATH=tests/integrations/ariadne arq: TESTPATH=tests/integrations/arq - asgi: TESTPATH=tests/integrations/asgi asyncpg: TESTPATH=tests/integrations/asyncpg - aws_lambda: TESTPATH=tests/integrations/aws_lambda beam: TESTPATH=tests/integrations/beam boto3: TESTPATH=tests/integrations/boto3 bottle: TESTPATH=tests/integrations/bottle @@ -729,13 +737,11 @@ setenv = chalice: TESTPATH=tests/integrations/chalice clickhouse_driver: TESTPATH=tests/integrations/clickhouse_driver cohere: TESTPATH=tests/integrations/cohere - cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context django: TESTPATH=tests/integrations/django dramatiq: TESTPATH=tests/integrations/dramatiq falcon: TESTPATH=tests/integrations/falcon - fastapi: TESTPATH=tests/integrations/fastapi + fastapi: TESTPATH=tests/integrations/fastapi flask: TESTPATH=tests/integrations/flask - gcp: TESTPATH=tests/integrations/gcp gql: TESTPATH=tests/integrations/gql graphene: TESTPATH=tests/integrations/graphene grpc: TESTPATH=tests/integrations/grpc @@ -744,7 +750,7 @@ setenv = huggingface_hub: TESTPATH=tests/integrations/huggingface_hub langchain-base: TESTPATH=tests/integrations/langchain langchain-notiktoken: TESTPATH=tests/integrations/langchain - langgraph: TESTPATH=tests/integrations/langgraph + langgraph: TESTPATH=tests/integrations/langgraph launchdarkly: TESTPATH=tests/integrations/launchdarkly litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru @@ -752,8 +758,6 @@ setenv = openai-notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature - opentelemetry: TESTPATH=tests/integrations/opentelemetry - potel: TESTPATH=tests/integrations/opentelemetry pure_eval: TESTPATH=tests/integrations/pure_eval pymongo: TESTPATH=tests/integrations/pymongo pyramid: TESTPATH=tests/integrations/pyramid @@ -774,7 +778,6 @@ setenv = trytond: TESTPATH=tests/integrations/trytond typer: TESTPATH=tests/integrations/typer unleash: TESTPATH=tests/integrations/unleash - socket: TESTPATH=tests/integrations/socket passenv = SENTRY_PYTHON_TEST_POSTGRES_HOST From ceefa6902703dcaedb3454bb0894ae7b3dd1da3e Mon Sep 17 00:00:00 2001 From: Anton Pirker Date: Fri, 26 Sep 2025 12:17:09 +0200 Subject: [PATCH 508/591] ci: Replace `black` and `flake8` with `ruff`. (#4866) ### Description Replace `black` and `flake8` with `ruff`. With this change also reformat the code base (only minor changes, because the configuration is basically the same.) The commits for the reformatting are not included in `git blame` calls (see `.git-blame-ignore-revs`). Also updates the `pre-commit` config so we have exactly the same linting/formatting in CI and local dev. We now have: - Less dependencies - Less configuration - Faster #### Issues * resolves: #2527 --- .coveragerc36 | 4 +- .cursor/rules/core-architecture.mdc | 4 +- .cursor/rules/integrations-guide.mdc | 8 +- .cursor/rules/project-overview.mdc | 4 +- .cursor/rules/quick-reference.mdc | 6 +- .cursor/rules/testing-guide.mdc | 8 +- .git-blame-ignore-revs | 3 + .pre-commit-config.yaml | 32 ++------ codecov.yml | 4 +- pyproject.toml | 79 ++++++++++--------- requirements-linting.txt | 6 +- scripts/build_aws_lambda_layer.py | 3 +- scripts/populate_tox/populate_tox.py | 8 +- scripts/populate_tox/releases.jsonl | 4 +- scripts/populate_tox/tox.jinja | 12 ++- scripts/ready_yet/main.py | 2 +- scripts/ready_yet/run.sh | 4 +- scripts/test-lambda-locally/README.md | 16 ++-- .../deploy-lambda-locally.sh | 4 +- .../test-lambda-locally/lambda_function.py | 7 +- sentry_sdk/client.py | 12 +-- sentry_sdk/consts.py | 1 - sentry_sdk/envelope.py | 42 ++++++---- sentry_sdk/feature_flags.py | 1 - sentry_sdk/hub.py | 26 +++--- sentry_sdk/integrations/asgi.py | 5 +- sentry_sdk/integrations/grpc/aio/client.py | 3 +- sentry_sdk/integrations/grpc/client.py | 7 +- sentry_sdk/integrations/langchain.py | 1 - sentry_sdk/integrations/launchdarkly.py | 1 - sentry_sdk/integrations/litestar.py | 4 +- sentry_sdk/integrations/pure_eval.py | 4 +- sentry_sdk/integrations/spark/spark_driver.py | 3 +- sentry_sdk/integrations/sqlalchemy.py | 8 +- sentry_sdk/integrations/starlette.py | 4 +- sentry_sdk/integrations/starlite.py | 4 +- sentry_sdk/integrations/wsgi.py | 5 +- sentry_sdk/metrics.py | 28 ++++--- sentry_sdk/profiler/utils.py | 8 +- sentry_sdk/scope.py | 9 ++- sentry_sdk/serializer.py | 4 +- sentry_sdk/session.py | 6 +- sentry_sdk/sessions.py | 6 +- sentry_sdk/tracing.py | 6 +- sentry_sdk/tracing_utils.py | 4 +- sentry_sdk/transport.py | 17 ++-- sentry_sdk/utils.py | 8 +- tests/conftest.py | 1 - tests/integrations/asyncio/test_asyncio.py | 3 +- .../RaiseErrorPerformanceDisabled/.gitignore | 4 +- .../RaiseErrorPerformanceEnabled/.gitignore | 4 +- .../TracesSampler/.gitignore | 4 +- tests/integrations/beam/test_beam.py | 8 +- tests/integrations/celery/test_celery.py | 10 ++- tests/integrations/django/asgi/test_asgi.py | 4 +- .../integrations/django/test_cache_module.py | 18 ++--- .../excepthook/test_excepthook.py | 8 +- tests/integrations/fastapi/test_fastapi.py | 18 ++--- tests/integrations/flask/test_flask.py | 12 +-- tests/integrations/gcp/test_gcp.py | 7 +- tests/integrations/gql/test_gql.py | 24 +++--- tests/integrations/httpx/test_httpx.py | 3 +- .../integrations/langchain/test_langchain.py | 18 ++--- .../integrations/langgraph/test_langgraph.py | 5 -- .../opentelemetry/test_span_processor.py | 12 ++- tests/integrations/rq/test_rq.py | 12 ++- tests/integrations/spark/test_spark.py | 1 - tests/integrations/sys_exit/test_sys_exit.py | 6 +- .../integrations/threading/test_threading.py | 6 +- tests/integrations/unleash/testutils.py | 1 - .../unraisablehook/test_unraisablehook.py | 4 +- tests/integrations/wsgi/test_wsgi.py | 9 ++- tests/test_basics.py | 6 +- tests/test_client.py | 10 +-- tests/test_conftest.py | 8 +- tests/test_gevent.py | 6 +- tests/test_propagationcontext.py | 3 +- tests/test_transport.py | 6 +- tests/test_utils.py | 6 +- tests/tracing/test_sampling.py | 7 +- tox.ini | 20 +++-- 81 files changed, 358 insertions(+), 361 deletions(-) create mode 100644 .git-blame-ignore-revs diff --git a/.coveragerc36 b/.coveragerc36 index 8642882ab1..e0c19bb634 100644 --- a/.coveragerc36 +++ b/.coveragerc36 @@ -3,12 +3,12 @@ [run] branch = true -omit = +omit = /tmp/* */tests/* */.venv/* [report] -exclude_lines = +exclude_lines = if TYPE_CHECKING: diff --git a/.cursor/rules/core-architecture.mdc b/.cursor/rules/core-architecture.mdc index 885773f16d..0af65f4815 100644 --- a/.cursor/rules/core-architecture.mdc +++ b/.cursor/rules/core-architecture.mdc @@ -1,6 +1,6 @@ --- -description: -globs: +description: +globs: alwaysApply: false --- # Core Architecture diff --git a/.cursor/rules/integrations-guide.mdc b/.cursor/rules/integrations-guide.mdc index 869a7f742a..6785b1522d 100644 --- a/.cursor/rules/integrations-guide.mdc +++ b/.cursor/rules/integrations-guide.mdc @@ -1,6 +1,6 @@ --- -description: -globs: +description: +globs: alwaysApply: false --- # Integrations Guide @@ -133,10 +133,10 @@ from sentry_sdk.integrations import Integration class MyIntegration(Integration): identifier = "my_integration" - + def __init__(self, param=None): self.param = param - + @staticmethod def setup_once(): # Install hooks, monkey patches, etc. diff --git a/.cursor/rules/project-overview.mdc b/.cursor/rules/project-overview.mdc index 13fad83ae7..c8c0cf77b9 100644 --- a/.cursor/rules/project-overview.mdc +++ b/.cursor/rules/project-overview.mdc @@ -1,6 +1,6 @@ --- -description: -globs: +description: +globs: alwaysApply: false --- # Sentry Python SDK - Project Overview diff --git a/.cursor/rules/quick-reference.mdc b/.cursor/rules/quick-reference.mdc index 453869fa83..ef90c22f78 100644 --- a/.cursor/rules/quick-reference.mdc +++ b/.cursor/rules/quick-reference.mdc @@ -1,6 +1,6 @@ --- -description: -globs: +description: +globs: alwaysApply: false --- # Quick Reference @@ -44,7 +44,7 @@ tox -e py3.12-django-v5.2.3 ### Code Quality -Our `linters` tox environment runs `black` for formatting, `flake8` for linting and `mypy` for type checking. +Our `linters` tox environment runs `ruff-format` for formatting, `ruff-check` for linting and `mypy` for type checking. ```bash tox -e linters diff --git a/.cursor/rules/testing-guide.mdc b/.cursor/rules/testing-guide.mdc index e336bb337a..939f1a095b 100644 --- a/.cursor/rules/testing-guide.mdc +++ b/.cursor/rules/testing-guide.mdc @@ -1,6 +1,6 @@ --- -description: -globs: +description: +globs: alwaysApply: false --- # Testing Guide @@ -65,10 +65,10 @@ def test_flask_integration(sentry_init, capture_events): # Test setup sentry_init(integrations=[FlaskIntegration()]) events = capture_events() - + # Test execution # ... test code ... - + # Assertions assert len(events) == 1 assert events[0]["exception"]["values"][0]["type"] == "ValueError" diff --git a/.git-blame-ignore-revs b/.git-blame-ignore-revs new file mode 100644 index 0000000000..147eaebfe8 --- /dev/null +++ b/.git-blame-ignore-revs @@ -0,0 +1,3 @@ +# Formatting commits to ignore in git blame +afea4a017bf13f78e82f725ea9d6a56a8e02cb34 +23a340a9dca60eea36de456def70c00952a33556 diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 9787e136bb..7ec62a8f37 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -1,31 +1,9 @@ # See https://pre-commit.com for more information # See https://pre-commit.com/hooks.html for more hooks repos: -- repo: https://github.com/pre-commit/pre-commit-hooks - rev: v4.3.0 +- repo: https://github.com/astral-sh/ruff-pre-commit + rev: v0.13.2 hooks: - - id: trailing-whitespace - - id: end-of-file-fixer - -- repo: https://github.com/psf/black - rev: 24.1.0 - hooks: - - id: black - exclude: ^(.*_pb2.py|.*_pb2_grpc.py) - -- repo: https://github.com/pycqa/flake8 - rev: 5.0.4 - hooks: - - id: flake8 - additional_dependencies: - [ - flake8-pyproject, - flake8-bugbear, - pep8-naming, - ] - -# Disabled for now, because it lists a lot of problems. -#- repo: https://github.com/pre-commit/mirrors-mypy -# rev: 'v0.931' -# hooks: -# - id: mypy + - id: ruff-check + args: [--fix] + - id: ruff-format diff --git a/codecov.yml b/codecov.yml index 086157690e..b7abcf8c86 100644 --- a/codecov.yml +++ b/codecov.yml @@ -19,9 +19,9 @@ comment: # Comments will only post when coverage changes. Furthermore, if a comment # already exists, and a newer commit results in no coverage change for the # entire pull, the comment will be deleted. - require_changes: true + require_changes: true require_base: true # must have a base report to post require_head: true # must have a head report to post github_checks: - annotations: false \ No newline at end of file + annotations: false diff --git a/pyproject.toml b/pyproject.toml index 44eded7641..8e6fe345f4 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,19 +1,3 @@ -# -# Tool: Black -# - -[tool.black] -# 'extend-exclude' excludes files or directories in addition to the defaults -extend-exclude = ''' -# A regex preceded with ^/ will apply only to files and directories -# in the root of the project. -( - .*_pb2.py # exclude autogenerated Protocol Buffer files anywhere in the project - | .*_pb2_grpc.py # exclude autogenerated Protocol Buffer files anywhere in the project -) -''' - - # # Tool: Coverage # @@ -196,29 +180,48 @@ module = "agents.*" ignore_missing_imports = true # -# Tool: Flake8 +# Tool: Ruff (linting and formatting) # -[tool.flake8] -extend-ignore = [ - # Handled by black (Whitespace before ':' -- handled by black) - "E203", - # Handled by black (Line too long) - "E501", - # Sometimes not possible due to execution order (Module level import is not at top of file) - "E402", - # I don't care (Do not assign a lambda expression, use a def) - "E731", - # does not apply to Python 2 (redundant exception types by flake8-bugbear) - "B014", - # I don't care (Lowercase imported as non-lowercase by pep8-naming) - "N812", - # is a worse version of and conflicts with B902 (first argument of a classmethod should be named cls) - "N804", +[tool.ruff] +# Target Python 3.7+ (minimum version supported by ruff) +target-version = "py37" + +# Exclude files and directories +extend-exclude = [ + "*_pb2.py", # Protocol Buffer files (covers all pb2 files including grpc_test_service_pb2.py) + "*_pb2_grpc.py", # Protocol Buffer files (covers all pb2_grpc files including grpc_test_service_pb2_grpc.py) + "checkouts", # From flake8 + "lol*", # From flake8 ] -extend-exclude = ["checkouts", "lol*"] -exclude = [ - # gRCP generated files - "grpc_test_service_pb2.py", - "grpc_test_service_pb2_grpc.py", + +[tool.ruff.lint] +# Match flake8's default rule selection exactly +# Flake8 by default only enables E and W (pycodestyle) + F (pyflakes) +select = [ + "E", # pycodestyle errors (same as flake8 default) + "W", # pycodestyle warnings (same as flake8 default) + "F", # Pyflakes (same as flake8 default) + # Note: B and N rules are NOT enabled by default in flake8 + # They were only active through the plugins, which may not have been fully enabled ] + +# Use ONLY the same ignores as the original flake8 config + compatibility for this codebase +ignore = [ + "E203", # Whitespace before ':' + "E501", # Line too long + "E402", # Module level import not at top of file + "E731", # Do not assign a lambda expression, use a def + "B014", # Redundant exception types + "N812", # Lowercase imported as non-lowercase + "N804", # First argument of classmethod should be named cls + + # Additional ignores for codebase compatibility + "F401", # Unused imports - many in TYPE_CHECKING blocks used for type comments + "E721", # Use isinstance instead of type() == - existing pattern in this codebase +] + +[tool.ruff.format] +# ruff format already excludes the same files as specified in extend-exclude +# Ensure Python 3.7 compatibility - avoid using Python 3.9+ syntax features +skip-magic-trailing-comma = false diff --git a/requirements-linting.txt b/requirements-linting.txt index 20db2151d0..1cc8274795 100644 --- a/requirements-linting.txt +++ b/requirements-linting.txt @@ -1,9 +1,5 @@ mypy -black -flake8==5.0.4 -flake8-pyproject # Flake8 plugin to support configuration in pyproject.toml -flake8-bugbear # Flake8 plugin -pep8-naming # Flake8 plugin +ruff types-certifi types-protobuf types-gevent diff --git a/scripts/build_aws_lambda_layer.py b/scripts/build_aws_lambda_layer.py index a7e2397546..a1078f4e19 100644 --- a/scripts/build_aws_lambda_layer.py +++ b/scripts/build_aws_lambda_layer.py @@ -75,8 +75,7 @@ def create_init_serverless_sdk_package(self): sentry-python-serverless zip """ serverless_sdk_path = ( - f"{self.python_site_packages}/sentry_sdk/" - f"integrations/init_serverless_sdk" + f"{self.python_site_packages}/sentry_sdk/integrations/init_serverless_sdk" ) if not os.path.exists(serverless_sdk_path): os.makedirs(serverless_sdk_path) diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 4bb4d78231..d625c1da72 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -601,7 +601,7 @@ def _add_python_versions_to_release( def _transform_target_python_versions( - python_versions: Union[str, dict[str, str], None] + python_versions: Union[str, dict[str, str], None], ) -> Union[SpecifierSet, dict[SpecifierSet, SpecifierSet], None]: """Wrap the contents of the `python` key in SpecifierSets.""" if not python_versions: @@ -711,9 +711,9 @@ def main(fail_on_changes: bool = False) -> None: name = _normalize_name(release["info"]["name"]) version = release["info"]["version"] CACHE[name][version] = release - CACHE[name][version][ - "_accessed" - ] = False # for cleaning up unused cache entries + CACHE[name][version]["_accessed"] = ( + False # for cleaning up unused cache entries + ) # Process packages packages = defaultdict(list) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index c68fe87734..fa24b089cd 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.38", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.39", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -77,7 +77,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "grpcio", "requires_python": "", "version": "1.32.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.75.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.75.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 39d2406b84..2a33e7790d 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -176,11 +176,9 @@ basepython = py3.12: python3.12 py3.13: python3.13 - # Python version is pinned here because flake8 actually behaves differently - # depending on which version is used. You can patch this out to point to - # some random Python 3 binary, but then you get guaranteed mismatches with - # CI. Other tools such as mypy and black have options that pin the Python - # version. + # Python version is pinned here for consistency across environments. + # Tools like ruff and mypy have options that pin the target Python + # version (configured in pyproject.toml), ensuring consistent behavior. linters: python3.12 commands = @@ -196,6 +194,6 @@ commands = [testenv:linters] commands = - flake8 tests sentry_sdk - black --check tests sentry_sdk + ruff check tests sentry_sdk + ruff format --check tests sentry_sdk mypy sentry_sdk diff --git a/scripts/ready_yet/main.py b/scripts/ready_yet/main.py index bba97d0c98..1f94adeb70 100644 --- a/scripts/ready_yet/main.py +++ b/scripts/ready_yet/main.py @@ -56,7 +56,7 @@ def main(): Check if libraries in our tox.ini are ready for Python version defined in `PYTHON_VERSION`. """ print(f"Checking libs from tox.ini for Python {PYTHON_VERSION} compatibility:") - + ready = set() not_ready = set() not_found = set() diff --git a/scripts/ready_yet/run.sh b/scripts/ready_yet/run.sh index f32bd7bdda..f872079c87 100755 --- a/scripts/ready_yet/run.sh +++ b/scripts/ready_yet/run.sh @@ -3,7 +3,7 @@ # exit on first error set -xe -reset +reset # create and activate virtual environment python -m venv .venv @@ -13,4 +13,4 @@ source .venv/bin/activate python -m pip install -r requirements.txt # Run the script -python main.py \ No newline at end of file +python main.py diff --git a/scripts/test-lambda-locally/README.md b/scripts/test-lambda-locally/README.md index 115927cc2b..2c02d6301f 100644 --- a/scripts/test-lambda-locally/README.md +++ b/scripts/test-lambda-locally/README.md @@ -1,28 +1,28 @@ # Test AWS Lambda functions locally -An easy way to run an AWS Lambda function with the Sentry SDK locally. +An easy way to run an AWS Lambda function with the Sentry SDK locally. -This is a small helper to create a AWS Lambda function that includes the +This is a small helper to create a AWS Lambda function that includes the currently checked out Sentry SDK and runs it in a local AWS Lambda environment. -Currently only embedding the Sentry SDK into the Lambda function package -is supported. Adding the SDK as Lambda Layer is not possible at the moment. +Currently only embedding the Sentry SDK into the Lambda function package +is supported. Adding the SDK as Lambda Layer is not possible at the moment. ## Prerequisites - Set `SENTRY_DSN` environment variable. The Lambda function will use this DSN. -- You need to have Docker installed and running. +- You need to have Docker installed and running. ## Run Lambda function -- Update `lambda_function.py` to include your test code. +- Update `lambda_function.py` to include your test code. - Run `./deploy-lambda-locally.sh`. This will: - Install [AWS SAM](https://aws.amazon.com/serverless/sam/) in a virtual Python environment - Create a lambda function package in `package/` that includes - The currently checked out Sentry SDK - All dependencies of the Sentry SDK (certifi and urllib3) - - The actual function defined in `lamdba_function.py`. + - The actual function defined in `lamdba_function.py`. - Zip everything together into lambda_deployment_package.zip - Run a local Lambda environment that serves that Lambda function. - Point your browser to `http://127.0.0.1:3000` to access your Lambda function. - - Currently GET and POST requests are possible. This is defined in `template.yaml`. \ No newline at end of file + - Currently GET and POST requests are possible. This is defined in `template.yaml`. diff --git a/scripts/test-lambda-locally/deploy-lambda-locally.sh b/scripts/test-lambda-locally/deploy-lambda-locally.sh index 495c1259dc..5da4fee6ba 100755 --- a/scripts/test-lambda-locally/deploy-lambda-locally.sh +++ b/scripts/test-lambda-locally/deploy-lambda-locally.sh @@ -13,9 +13,9 @@ fi uv sync # Create a deployment package of the lambda function in `lambda_function.py`. -rm -rf package && mkdir -p package +rm -rf package && mkdir -p package pip install ../../../sentry-python -t package/ --upgrade -cp lambda_function.py package/ +cp lambda_function.py package/ cd package && zip -r ../lambda_deployment_package.zip . && cd .. # Start the local Lambda server with the new function (defined in template.yaml) diff --git a/scripts/test-lambda-locally/lambda_function.py b/scripts/test-lambda-locally/lambda_function.py index ceab090499..fdaa2160eb 100644 --- a/scripts/test-lambda-locally/lambda_function.py +++ b/scripts/test-lambda-locally/lambda_function.py @@ -5,21 +5,22 @@ from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration from sentry_sdk.integrations.logging import LoggingIntegration + def lambda_handler(event, context): sentry_sdk.init( dsn=os.environ.get("SENTRY_DSN"), attach_stacktrace=True, integrations=[ LoggingIntegration(level=logging.INFO, event_level=logging.ERROR), - AwsLambdaIntegration(timeout_warning=True) + AwsLambdaIntegration(timeout_warning=True), ], traces_sample_rate=1.0, debug=True, ) try: - my_dict = {"a" : "test"} - value = my_dict["b"] # This should raise exception + my_dict = {"a": "test"} + _ = my_dict["b"] # This should raise exception except: logging.exception("Key Does not Exists") raise diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index c45d5e2f4f..c06043ebe2 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -178,9 +178,7 @@ class BaseClient: def __init__(self, options=None): # type: (Optional[Dict[str, Any]]) -> None - self.options = ( - options if options is not None else DEFAULT_OPTIONS - ) # type: Dict[str, Any] + self.options = options if options is not None else DEFAULT_OPTIONS # type: Dict[str, Any] self.transport = None # type: Optional[Transport] self.monitor = None # type: Optional[Monitor] @@ -956,7 +954,7 @@ def _capture_experimental_log(self, log): debug = self.options.get("debug", False) if debug: logger.debug( - f'[Sentry Logs] [{log.get("severity_text")}] {log.get("body")}' + f"[Sentry Logs] [{log.get('severity_text')}] {log.get('body')}" ) before_send_log = get_before_send_log(self.options) @@ -970,7 +968,8 @@ def _capture_experimental_log(self, log): self.log_batcher.add(log) def capture_session( - self, session # type: Session + self, + session, # type: Session ): # type: (...) -> None if not session.release: @@ -991,7 +990,8 @@ def get_integration(self, name_or_class): ... def get_integration( - self, name_or_class # type: Union[str, Type[Integration]] + self, + name_or_class, # type: Union[str, Type[Integration]] ): # type: (...) -> Optional[Integration] """Returns the integration for this client by name or class. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 5b8e840f13..9e84dc3dd2 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -852,7 +852,6 @@ class OP: # This type exists to trick mypy and PyCharm into thinking `init` and `Client` # take these arguments (even though they take opaque **kwargs) class ClientConstructor: - def __init__( self, dsn=None, # type: Optional[str] diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index 7dbbdec5c8..d9b2c1629a 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -57,25 +57,29 @@ def description(self): ) def add_event( - self, event # type: Event + self, + event, # type: Event ): # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=event), type="event")) def add_transaction( - self, transaction # type: Event + self, + transaction, # type: Event ): # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=transaction), type="transaction")) def add_profile( - self, profile # type: Any + self, + profile, # type: Any ): # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=profile), type="profile")) def add_profile_chunk( - self, profile_chunk # type: Any + self, + profile_chunk, # type: Any ): # type: (...) -> None self.add_item( @@ -87,13 +91,15 @@ def add_profile_chunk( ) def add_checkin( - self, checkin # type: Any + self, + checkin, # type: Any ): # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=checkin), type="check_in")) def add_session( - self, session # type: Union[Session, Any] + self, + session, # type: Union[Session, Any] ): # type: (...) -> None if isinstance(session, Session): @@ -101,13 +107,15 @@ def add_session( self.add_item(Item(payload=PayloadRef(json=session), type="session")) def add_sessions( - self, sessions # type: Any + self, + sessions, # type: Any ): # type: (...) -> None self.add_item(Item(payload=PayloadRef(json=sessions), type="sessions")) def add_item( - self, item # type: Item + self, + item, # type: Item ): # type: (...) -> None self.items.append(item) @@ -133,7 +141,8 @@ def __iter__(self): return iter(self.items) def serialize_into( - self, f # type: Any + self, + f, # type: Any ): # type: (...) -> None f.write(json_dumps(self.headers)) @@ -149,7 +158,8 @@ def serialize(self): @classmethod def deserialize_from( - cls, f # type: Any + cls, + f, # type: Any ): # type: (...) -> Envelope headers = parse_json(f.readline()) @@ -163,7 +173,8 @@ def deserialize_from( @classmethod def deserialize( - cls, bytes # type: bytes + cls, + bytes, # type: bytes ): # type: (...) -> Envelope return cls.deserialize_from(io.BytesIO(bytes)) @@ -307,7 +318,8 @@ def get_transaction_event(self): return None def serialize_into( - self, f # type: Any + self, + f, # type: Any ): # type: (...) -> None headers = dict(self.headers) @@ -326,7 +338,8 @@ def serialize(self): @classmethod def deserialize_from( - cls, f # type: Any + cls, + f, # type: Any ): # type: (...) -> Optional[Item] line = f.readline().rstrip() @@ -349,7 +362,8 @@ def deserialize_from( @classmethod def deserialize( - cls, bytes # type: bytes + cls, + bytes, # type: bytes ): # type: (...) -> Optional[Item] return cls.deserialize_from(io.BytesIO(bytes)) diff --git a/sentry_sdk/feature_flags.py b/sentry_sdk/feature_flags.py index eb53acae5d..03fba9c53c 100644 --- a/sentry_sdk/feature_flags.py +++ b/sentry_sdk/feature_flags.py @@ -15,7 +15,6 @@ class FlagBuffer: - def __init__(self, capacity): # type: (int) -> None self.capacity = capacity diff --git a/sentry_sdk/hub.py b/sentry_sdk/hub.py index 7fda9202df..6f2d1bbf13 100644 --- a/sentry_sdk/hub.py +++ b/sentry_sdk/hub.py @@ -205,7 +205,8 @@ def __exit__( scope._isolation_scope.set(old_isolation_scope) def run( - self, callback # type: Callable[[], T] + self, + callback, # type: Callable[[], T] ): # type: (...) -> T """ @@ -219,7 +220,8 @@ def run( return callback() def get_integration( - self, name_or_class # type: Union[str, Type[Integration]] + self, + name_or_class, # type: Union[str, Type[Integration]] ): # type: (...) -> Any """ @@ -277,7 +279,8 @@ def last_event_id(self): return self._last_event_id def bind_client( - self, new # type: Optional[BaseClient] + self, + new, # type: Optional[BaseClient] ): # type: (...) -> None """ @@ -430,7 +433,7 @@ def start_transaction( transaction=None, instrumenter=INSTRUMENTER.SENTRY, custom_sampling_context=None, - **kwargs + **kwargs, ): # type: (Optional[Transaction], str, Optional[SamplingContext], Unpack[TransactionKwargs]) -> Union[Transaction, NoOpSpan] """ @@ -487,14 +490,16 @@ def continue_trace(self, environ_or_headers, op=None, name=None, source=None): @overload def push_scope( - self, callback=None # type: Optional[None] + self, + callback=None, # type: Optional[None] ): # type: (...) -> ContextManager[Scope] pass @overload def push_scope( # noqa: F811 - self, callback # type: Callable[[Scope], None] + self, + callback, # type: Callable[[Scope], None] ): # type: (...) -> None pass @@ -540,14 +545,16 @@ def pop_scope_unsafe(self): @overload def configure_scope( - self, callback=None # type: Optional[None] + self, + callback=None, # type: Optional[None] ): # type: (...) -> ContextManager[Scope] pass @overload def configure_scope( # noqa: F811 - self, callback # type: Callable[[Scope], None] + self, + callback, # type: Callable[[Scope], None] ): # type: (...) -> None pass @@ -587,7 +594,8 @@ def inner(): return inner() def start_session( - self, session_mode="application" # type: str + self, + session_mode="application", # type: str ): # type: (...) -> None """ diff --git a/sentry_sdk/integrations/asgi.py b/sentry_sdk/integrations/asgi.py index dde8128a33..28b44cc7ab 100644 --- a/sentry_sdk/integrations/asgi.py +++ b/sentry_sdk/integrations/asgi.py @@ -233,14 +233,15 @@ async def _run_app(self, scope, receive, send, asgi_version): if transaction: transaction.set_tag("asgi.type", ty) - with ( + transaction_context = ( sentry_sdk.start_transaction( transaction, custom_sampling_context={"asgi_scope": scope}, ) if transaction is not None else nullcontext() - ): + ) + with transaction_context: try: async def _sentry_wrapped_send(event): diff --git a/sentry_sdk/integrations/grpc/aio/client.py b/sentry_sdk/integrations/grpc/aio/client.py index ff3c213176..7462675a97 100644 --- a/sentry_sdk/integrations/grpc/aio/client.py +++ b/sentry_sdk/integrations/grpc/aio/client.py @@ -65,7 +65,8 @@ async def intercept_unary_unary( class SentryUnaryStreamClientInterceptor( - ClientInterceptor, UnaryStreamClientInterceptor # type: ignore + ClientInterceptor, + UnaryStreamClientInterceptor, # type: ignore ): async def intercept_unary_stream( self, diff --git a/sentry_sdk/integrations/grpc/client.py b/sentry_sdk/integrations/grpc/client.py index a5b4f9f52e..ef24821ed2 100644 --- a/sentry_sdk/integrations/grpc/client.py +++ b/sentry_sdk/integrations/grpc/client.py @@ -19,7 +19,8 @@ class ClientInterceptor( - grpc.UnaryUnaryClientInterceptor, grpc.UnaryStreamClientInterceptor # type: ignore + grpc.UnaryUnaryClientInterceptor, # type: ignore + grpc.UnaryStreamClientInterceptor, # type: ignore ): _is_intercepted = False @@ -60,9 +61,7 @@ def intercept_unary_stream(self, continuation, client_call_details, request): client_call_details ) - response = continuation( - client_call_details, request - ) # type: UnaryStreamCall + response = continuation(client_call_details, request) # type: UnaryStreamCall # Setting code on unary-stream leads to execution getting stuck # span.set_data("code", response.code().name) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 6bc3ceb93e..fdba26569d 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -554,7 +554,6 @@ def _simplify_langchain_tools(tools): for tool in tools: try: if isinstance(tool, dict): - if "function" in tool and isinstance(tool["function"], dict): func = tool["function"] simplified_tool = { diff --git a/sentry_sdk/integrations/launchdarkly.py b/sentry_sdk/integrations/launchdarkly.py index d3c423e7be..6dfc1958b7 100644 --- a/sentry_sdk/integrations/launchdarkly.py +++ b/sentry_sdk/integrations/launchdarkly.py @@ -44,7 +44,6 @@ def setup_once(): class LaunchDarklyHook(Hook): - @property def metadata(self): # type: () -> Metadata diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 2be4d376e0..745a00bcba 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -222,9 +222,7 @@ async def handle_wrapper(self, scope, receive, send): return await old_handle(self, scope, receive, send) sentry_scope = sentry_sdk.get_isolation_scope() - request = scope["app"].request_class( - scope=scope, receive=receive, send=send - ) # type: Request[Any, Any] + request = scope["app"].request_class(scope=scope, receive=receive, send=send) # type: Request[Any, Any] extracted_request_data = ConnectionDataExtractor( parse_body=True, parse_query=True )(request) diff --git a/sentry_sdk/integrations/pure_eval.py b/sentry_sdk/integrations/pure_eval.py index c1c3d63871..6ac10dfe1b 100644 --- a/sentry_sdk/integrations/pure_eval.py +++ b/sentry_sdk/integrations/pure_eval.py @@ -116,7 +116,9 @@ def start(n): return (n.lineno, n.col_offset) nodes_before_stmt = [ - node for node in nodes if start(node) < stmt.last_token.end # type: ignore + node + for node in nodes + if start(node) < stmt.last_token.end # type: ignore ] if nodes_before_stmt: # The position of the last node before or in the statement diff --git a/sentry_sdk/integrations/spark/spark_driver.py b/sentry_sdk/integrations/spark/spark_driver.py index fac985357f..b22dc2c807 100644 --- a/sentry_sdk/integrations/spark/spark_driver.py +++ b/sentry_sdk/integrations/spark/spark_driver.py @@ -158,7 +158,8 @@ def onExecutorBlacklisted(self, executorBlacklisted): # noqa: N802,N803 pass def onExecutorBlacklistedForStage( # noqa: N802 - self, executorBlacklistedForStage # noqa: N803 + self, + executorBlacklistedForStage, # noqa: N803 ): # type: (Any) -> None pass diff --git a/sentry_sdk/integrations/sqlalchemy.py b/sentry_sdk/integrations/sqlalchemy.py index 068d373053..0e039f93f3 100644 --- a/sentry_sdk/integrations/sqlalchemy.py +++ b/sentry_sdk/integrations/sqlalchemy.py @@ -64,9 +64,7 @@ def _before_cursor_execute( @ensure_integration_enabled(SqlalchemyIntegration) def _after_cursor_execute(conn, cursor, statement, parameters, context, *args): # type: (Any, Any, Any, Any, Any, *Any) -> None - ctx_mgr = getattr( - context, "_sentry_sql_span_manager", None - ) # type: Optional[ContextManager[Any]] + ctx_mgr = getattr(context, "_sentry_sql_span_manager", None) # type: Optional[ContextManager[Any]] if ctx_mgr is not None: context._sentry_sql_span_manager = None @@ -92,9 +90,7 @@ def _handle_error(context, *args): # _after_cursor_execute does not get called for crashing SQL stmts. Judging # from SQLAlchemy codebase it does seem like any error coming into this # handler is going to be fatal. - ctx_mgr = getattr( - execution_context, "_sentry_sql_span_manager", None - ) # type: Optional[ContextManager[Any]] + ctx_mgr = getattr(execution_context, "_sentry_sql_span_manager", None) # type: Optional[ContextManager[Any]] if ctx_mgr is not None: execution_context._sentry_sql_span_manager = None diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index c7ce40618b..f1a0e360bb 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -103,9 +103,7 @@ def __init__( self.http_methods_to_capture = tuple(map(str.upper, http_methods_to_capture)) if isinstance(failed_request_status_codes, Set): - self.failed_request_status_codes = ( - failed_request_status_codes - ) # type: Container[int] + self.failed_request_status_codes = failed_request_status_codes # type: Container[int] else: warnings.warn( "Passing a list or None for failed_request_status_codes is deprecated. " diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index b402aa2184..daab82d642 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -200,9 +200,7 @@ async def handle_wrapper(self, scope, receive, send): return await old_handle(self, scope, receive, send) sentry_scope = sentry_sdk.get_isolation_scope() - request = scope["app"].request_class( - scope=scope, receive=receive, send=send - ) # type: Request[Any, Any] + request = scope["app"].request_class(scope=scope, receive=receive, send=send) # type: Request[Any, Any] extracted_request_data = ConnectionDataExtractor( parse_body=True, parse_query=True )(request) diff --git a/sentry_sdk/integrations/wsgi.py b/sentry_sdk/integrations/wsgi.py index e628e50e69..fa79ec96da 100644 --- a/sentry_sdk/integrations/wsgi.py +++ b/sentry_sdk/integrations/wsgi.py @@ -119,14 +119,15 @@ def __call__(self, environ, start_response): origin=self.span_origin, ) - with ( + transaction_context = ( sentry_sdk.start_transaction( transaction, custom_sampling_context={"wsgi_environ": environ}, ) if transaction is not None else nullcontext() - ): + ) + with transaction_context: try: response = self.app( environ, diff --git a/sentry_sdk/metrics.py b/sentry_sdk/metrics.py index 4bdbc62253..d0041114ce 100644 --- a/sentry_sdk/metrics.py +++ b/sentry_sdk/metrics.py @@ -159,7 +159,8 @@ class CounterMetric(Metric): __slots__ = ("value",) def __init__( - self, first # type: MetricValue + self, + first, # type: MetricValue ): # type: (...) -> None self.value = float(first) @@ -170,7 +171,8 @@ def weight(self): return 1 def add( - self, value # type: MetricValue + self, + value, # type: MetricValue ): # type: (...) -> None self.value += float(value) @@ -190,7 +192,8 @@ class GaugeMetric(Metric): ) def __init__( - self, first # type: MetricValue + self, + first, # type: MetricValue ): # type: (...) -> None first = float(first) @@ -207,7 +210,8 @@ def weight(self): return 5 def add( - self, value # type: MetricValue + self, + value, # type: MetricValue ): # type: (...) -> None value = float(value) @@ -232,7 +236,8 @@ class DistributionMetric(Metric): __slots__ = ("value",) def __init__( - self, first # type: MetricValue + self, + first, # type: MetricValue ): # type(...) -> None self.value = [float(first)] @@ -243,7 +248,8 @@ def weight(self): return len(self.value) def add( - self, value # type: MetricValue + self, + value, # type: MetricValue ): # type: (...) -> None self.value.append(float(value)) @@ -257,7 +263,8 @@ class SetMetric(Metric): __slots__ = ("value",) def __init__( - self, first # type: MetricValue + self, + first, # type: MetricValue ): # type: (...) -> None self.value = {first} @@ -268,7 +275,8 @@ def weight(self): return len(self.value) def add( - self, value # type: MetricValue + self, + value, # type: MetricValue ): # type: (...) -> None self.value.add(value) @@ -373,9 +381,7 @@ class LocalAggregator: def __init__(self): # type: (...) -> None - self._measurements = ( - {} - ) # type: Dict[Tuple[str, MetricTagsInternal], Tuple[float, float, int, float]] + self._measurements = {} # type: Dict[Tuple[str, MetricTagsInternal], Tuple[float, float, int, float]] def add( self, diff --git a/sentry_sdk/profiler/utils.py b/sentry_sdk/profiler/utils.py index 3554cddb5d..7d311e91f4 100644 --- a/sentry_sdk/profiler/utils.py +++ b/sentry_sdk/profiler/utils.py @@ -85,9 +85,7 @@ def get_frame_name(frame): if ( # the co_varnames start with the frame's positional arguments # and we expect the first to be `self` if its an instance method - co_varnames - and co_varnames[0] == "self" - and "self" in frame.f_locals + co_varnames and co_varnames[0] == "self" and "self" in frame.f_locals ): for cls in type(frame.f_locals["self"]).__mro__: if name in cls.__dict__: @@ -101,9 +99,7 @@ def get_frame_name(frame): if ( # the co_varnames start with the frame's positional arguments # and we expect the first to be `cls` if its a class method - co_varnames - and co_varnames[0] == "cls" - and "cls" in frame.f_locals + co_varnames and co_varnames[0] == "cls" and "cls" in frame.f_locals ): for cls in frame.f_locals["cls"].__mro__: if name in cls.__dict__: diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 3356de57a8..c871e6a467 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -894,7 +894,8 @@ def set_context( self._contexts[key] = value def remove_context( - self, key # type: str + self, + key, # type: str ): # type: (...) -> None """Removes a context.""" @@ -910,7 +911,8 @@ def set_extra( self._extras[key] = value def remove_extra( - self, key # type: str + self, + key, # type: str ): # type: (...) -> None """Removes a specific extra key.""" @@ -1321,7 +1323,8 @@ def resume_auto_session_tracking(self): self._force_auto_session_tracking = None def add_event_processor( - self, func # type: EventProcessor + self, + func, # type: EventProcessor ): # type: (...) -> None """Register a scope local event processor on the scope. diff --git a/sentry_sdk/serializer.py b/sentry_sdk/serializer.py index 6bde5c08bd..1775b1b555 100644 --- a/sentry_sdk/serializer.py +++ b/sentry_sdk/serializer.py @@ -128,9 +128,7 @@ def serialize(event, **kwargs): path = [] # type: List[Segment] meta_stack = [] # type: List[Dict[str, Any]] - keep_request_bodies = ( - kwargs.pop("max_request_body_size", None) == "always" - ) # type: bool + keep_request_bodies = kwargs.pop("max_request_body_size", None) == "always" # type: bool max_value_length = kwargs.pop("max_value_length", None) # type: Optional[int] is_vars = kwargs.pop("is_vars", False) custom_repr = kwargs.pop("custom_repr", None) # type: Callable[..., Optional[str]] diff --git a/sentry_sdk/session.py b/sentry_sdk/session.py index c1d422c115..af9551c56e 100644 --- a/sentry_sdk/session.py +++ b/sentry_sdk/session.py @@ -130,7 +130,8 @@ def update( self.status = status def close( - self, status=None # type: Optional[SessionStatus] + self, + status=None, # type: Optional[SessionStatus] ): # type: (...) -> Any if status is None and self.status == "ok": @@ -139,7 +140,8 @@ def close( self.update(status=status) def get_json_attrs( - self, with_user_info=True # type: Optional[bool] + self, + with_user_info=True, # type: Optional[bool] ): # type: (...) -> Any attrs = {} diff --git a/sentry_sdk/sessions.py b/sentry_sdk/sessions.py index 00fda23200..2bf4ee707a 100644 --- a/sentry_sdk/sessions.py +++ b/sentry_sdk/sessions.py @@ -228,7 +228,8 @@ def _thread(): return None def add_aggregate_session( - self, session # type: Session + self, + session, # type: Session ): # type: (...) -> None # NOTE on `session.did`: @@ -259,7 +260,8 @@ def add_aggregate_session( state["exited"] = state.get("exited", 0) + 1 def add_session( - self, session # type: Session + self, + session, # type: Session ): # type: (...) -> None if session.session_mode == "request": diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 4edda21075..a82b99ff4d 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -998,9 +998,7 @@ def finish( # For backwards compatibility, we must handle the case where `scope` # or `hub` could both either be a `Scope` or a `Hub`. - scope = self._get_scope_from_finish_args( - scope, hub - ) # type: Optional[sentry_sdk.Scope] + scope = self._get_scope_from_finish_args(scope, hub) # type: Optional[sentry_sdk.Scope] scope = scope or self.scope or sentry_sdk.get_current_scope() client = sentry_sdk.get_client() @@ -1209,8 +1207,8 @@ def _set_initial_sampling_decision(self, sampling_context): sample_rate = ( client.options["traces_sampler"](sampling_context) if callable(client.options.get("traces_sampler")) + # default inheritance behavior else ( - # default inheritance behavior sampling_context["parent_sampled"] if sampling_context["parent_sampled"] is not None else client.options["traces_sample_rate"] diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 2f3e334e3f..b81d647c6d 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -527,9 +527,7 @@ def _fill_sample_rand(self): ) return - self.dynamic_sampling_context["sample_rand"] = ( - f"{sample_rand:.6f}" # noqa: E231 - ) + self.dynamic_sampling_context["sample_rand"] = f"{sample_rand:.6f}" # noqa: E231 def _sample_rand(self): # type: () -> Optional[str] diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index e904081959..75384519e9 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -203,9 +203,7 @@ def __init__(self, options): self._disabled_until = {} # type: Dict[Optional[EventDataCategory], datetime] # We only use this Retry() class for the `get_retry_after` method it exposes self._retry = urllib3.util.Retry() - self._discarded_events = defaultdict( - int - ) # type: DefaultDict[Tuple[EventDataCategory, str], int] + self._discarded_events = defaultdict(int) # type: DefaultDict[Tuple[EventDataCategory, str], int] self._last_client_report_sent = time.time() self._pool = self._make_pool() @@ -549,7 +547,8 @@ def _request( raise NotImplementedError() def capture_envelope( - self, envelope # type: Envelope + self, + envelope, # type: Envelope ): # type: (...) -> None def send_envelope_wrapper(): @@ -862,14 +861,16 @@ class _FunctionTransport(Transport): """ def __init__( - self, func # type: Callable[[Event], None] + self, + func, # type: Callable[[Event], None] ): # type: (...) -> None Transport.__init__(self) self._func = func def capture_event( - self, event # type: Event + self, + event, # type: Event ): # type: (...) -> None self._func(event) @@ -891,9 +892,7 @@ def make_transport(options): use_http2_transport = options.get("_experiments", {}).get("transport_http2", False) # By default, we use the http transport class - transport_cls = ( - Http2Transport if use_http2_transport else HttpTransport - ) # type: Type[Transport] + transport_cls = Http2Transport if use_http2_transport else HttpTransport # type: Type[Transport] if isinstance(ref_transport, Transport): return ref_transport diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 3fe3ac3eec..2083fd296c 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -389,7 +389,8 @@ def __init__( self.client = client def get_api_url( - self, type=EndpointType.ENVELOPE # type: EndpointType + self, + type=EndpointType.ENVELOPE, # type: EndpointType ): # type: (...) -> str """Returns the API url for storing events.""" @@ -850,7 +851,9 @@ def exceptions_from_error( parent_id = exception_id exception_id += 1 - should_supress_context = hasattr(exc_value, "__suppress_context__") and exc_value.__suppress_context__ # type: ignore + should_supress_context = ( + hasattr(exc_value, "__suppress_context__") and exc_value.__suppress_context__ # type: ignore + ) if should_supress_context: # Add direct cause. # The field `__cause__` is set when raised with the exception (using the `from` keyword). @@ -1845,7 +1848,6 @@ def now(): from gevent import get_hub as get_gevent_hub from gevent.monkey import is_module_patched except ImportError: - # it's not great that the signatures are different, get_hub can't return None # consider adding an if TYPE_CHECKING to change the signature to Optional[Hub] def get_gevent_hub(): # type: ignore[misc] diff --git a/tests/conftest.py b/tests/conftest.py index 01b1e9a81f..faa0251d9b 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -384,7 +384,6 @@ def render_span(span): root_span = event["contexts"]["trace"] - # Return a list instead of a multiline string because black will know better how to format that return "\n".join(render_span(root_span)) return inner diff --git a/tests/integrations/asyncio/test_asyncio.py b/tests/integrations/asyncio/test_asyncio.py index fb75bfc69b..66113746bf 100644 --- a/tests/integrations/asyncio/test_asyncio.py +++ b/tests/integrations/asyncio/test_asyncio.py @@ -291,7 +291,8 @@ def test_sentry_task_factory_with_factory(mock_get_running_loop): @patch("asyncio.get_running_loop") @patch("sentry_sdk.integrations.asyncio.Task") def test_sentry_task_factory_context_no_factory( - MockTask, mock_get_running_loop # noqa: N803 + MockTask, + mock_get_running_loop, # noqa: N803 ): mock_loop = mock_get_running_loop.return_value mock_coro = MagicMock() diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore index ee0b7b9305..1c56884372 100644 --- a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceDisabled/.gitignore @@ -1,4 +1,4 @@ -# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies # into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. # Ignore everything @@ -8,4 +8,4 @@ !index.py # And not .gitignore itself -!.gitignore \ No newline at end of file +!.gitignore diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore index ee0b7b9305..1c56884372 100644 --- a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/RaiseErrorPerformanceEnabled/.gitignore @@ -1,4 +1,4 @@ -# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies # into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. # Ignore everything @@ -8,4 +8,4 @@ !index.py # And not .gitignore itself -!.gitignore \ No newline at end of file +!.gitignore diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore index ee0b7b9305..1c56884372 100644 --- a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TracesSampler/.gitignore @@ -1,4 +1,4 @@ -# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies # into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. # Ignore everything @@ -8,4 +8,4 @@ !index.py # And not .gitignore itself -!.gitignore \ No newline at end of file +!.gitignore diff --git a/tests/integrations/beam/test_beam.py b/tests/integrations/beam/test_beam.py index 8c503b4c8c..809c4122e4 100644 --- a/tests/integrations/beam/test_beam.py +++ b/tests/integrations/beam/test_beam.py @@ -144,10 +144,10 @@ def test_monkey_patch_signature(f, args, kwargs): try: expected_signature = inspect.signature(f) test_signature = inspect.signature(f_temp) - assert ( - expected_signature == test_signature - ), "Failed on {}, signature {} does not match {}".format( - f, expected_signature, test_signature + assert expected_signature == test_signature, ( + "Failed on {}, signature {} does not match {}".format( + f, expected_signature, test_signature + ) ) except Exception: # expected to pass for py2.7 diff --git a/tests/integrations/celery/test_celery.py b/tests/integrations/celery/test_celery.py index 80b4a423cb..b8fc2bb3e8 100644 --- a/tests/integrations/celery/test_celery.py +++ b/tests/integrations/celery/test_celery.py @@ -374,9 +374,9 @@ def dummy_task(self): assert submit_transaction["type"] == "transaction" assert submit_transaction["transaction"] == "submit_celery" - assert len( - submit_transaction["spans"] - ), 4 # Because redis integration was auto enabled + assert len(submit_transaction["spans"]), ( + 4 + ) # Because redis integration was auto enabled span = submit_transaction["spans"][0] assert span["op"] == "queue.submit.celery" assert span["description"] == "dummy_task" @@ -439,7 +439,9 @@ def dummy_task(self, x, y): def test_traces_sampler_gets_task_info_in_sampling_context( - init_celery, celery_invocation, DictionaryContaining # noqa:N803 + init_celery, + celery_invocation, + DictionaryContaining, # noqa:N803 ): traces_sampler = mock.Mock() celery = init_celery(traces_sampler=traces_sampler) diff --git a/tests/integrations/django/asgi/test_asgi.py b/tests/integrations/django/asgi/test_asgi.py index 3c78ac3f38..8a30c7f5c0 100644 --- a/tests/integrations/django/asgi/test_asgi.py +++ b/tests/integrations/django/asgi/test_asgi.py @@ -432,9 +432,7 @@ async def test_trace_from_headers_if_performance_disabled(sentry_init, capture_e PICTURE = os.path.join(os.path.dirname(os.path.abspath(__file__)), "image.png") BODY_FORM = """--fd721ef49ea403a6\r\nContent-Disposition: form-data; name="username"\r\n\r\nJane\r\n--fd721ef49ea403a6\r\nContent-Disposition: form-data; name="password"\r\n\r\nhello123\r\n--fd721ef49ea403a6\r\nContent-Disposition: form-data; name="photo"; filename="image.png"\r\nContent-Type: image/png\r\nContent-Transfer-Encoding: base64\r\n\r\n{{image_data}}\r\n--fd721ef49ea403a6--\r\n""".replace( "{{image_data}}", base64.b64encode(open(PICTURE, "rb").read()).decode("utf-8") -).encode( - "utf-8" -) +).encode("utf-8") BODY_FORM_CONTENT_LENGTH = str(len(BODY_FORM)).encode("utf-8") diff --git a/tests/integrations/django/test_cache_module.py b/tests/integrations/django/test_cache_module.py index 263f9f36f8..bc58dd8471 100644 --- a/tests/integrations/django/test_cache_module.py +++ b/tests/integrations/django/test_cache_module.py @@ -529,33 +529,33 @@ def test_cache_spans_get_many(sentry_init, capture_events, use_django_caching): from django.core.cache import cache with sentry_sdk.start_transaction(): - cache.get_many([f"S{id}", f"S{id+1}"]) + cache.get_many([f"S{id}", f"S{id + 1}"]) cache.set(f"S{id}", "Sensitive1") - cache.get_many([f"S{id}", f"S{id+1}"]) + cache.get_many([f"S{id}", f"S{id + 1}"]) (transaction,) = events assert len(transaction["spans"]) == 7 assert transaction["spans"][0]["op"] == "cache.get" - assert transaction["spans"][0]["description"] == f"S{id}, S{id+1}" + assert transaction["spans"][0]["description"] == f"S{id}, S{id + 1}" assert transaction["spans"][1]["op"] == "cache.get" assert transaction["spans"][1]["description"] == f"S{id}" assert transaction["spans"][2]["op"] == "cache.get" - assert transaction["spans"][2]["description"] == f"S{id+1}" + assert transaction["spans"][2]["description"] == f"S{id + 1}" assert transaction["spans"][3]["op"] == "cache.put" assert transaction["spans"][3]["description"] == f"S{id}" assert transaction["spans"][4]["op"] == "cache.get" - assert transaction["spans"][4]["description"] == f"S{id}, S{id+1}" + assert transaction["spans"][4]["description"] == f"S{id}, S{id + 1}" assert transaction["spans"][5]["op"] == "cache.get" assert transaction["spans"][5]["description"] == f"S{id}" assert transaction["spans"][6]["op"] == "cache.get" - assert transaction["spans"][6]["description"] == f"S{id+1}" + assert transaction["spans"][6]["description"] == f"S{id + 1}" @pytest.mark.forked @@ -578,20 +578,20 @@ def test_cache_spans_set_many(sentry_init, capture_events, use_django_caching): from django.core.cache import cache with sentry_sdk.start_transaction(): - cache.set_many({f"S{id}": "Sensitive1", f"S{id+1}": "Sensitive2"}) + cache.set_many({f"S{id}": "Sensitive1", f"S{id + 1}": "Sensitive2"}) cache.get(f"S{id}") (transaction,) = events assert len(transaction["spans"]) == 4 assert transaction["spans"][0]["op"] == "cache.put" - assert transaction["spans"][0]["description"] == f"S{id}, S{id+1}" + assert transaction["spans"][0]["description"] == f"S{id}, S{id + 1}" assert transaction["spans"][1]["op"] == "cache.put" assert transaction["spans"][1]["description"] == f"S{id}" assert transaction["spans"][2]["op"] == "cache.put" - assert transaction["spans"][2]["description"] == f"S{id+1}" + assert transaction["spans"][2]["description"] == f"S{id + 1}" assert transaction["spans"][3]["op"] == "cache.get" assert transaction["spans"][3]["description"] == f"S{id}" diff --git a/tests/integrations/excepthook/test_excepthook.py b/tests/integrations/excepthook/test_excepthook.py index 745f62d818..5a19b4f985 100644 --- a/tests/integrations/excepthook/test_excepthook.py +++ b/tests/integrations/excepthook/test_excepthook.py @@ -32,9 +32,7 @@ def capture_envelope(self, envelope): frame_value = "LOL" 1/0 - """.format( - transport=transport, options=options - ) + """.format(transport=transport, options=options) ) ) @@ -75,9 +73,7 @@ def capture_envelope(self, envelope): frame_value = "LOL" 1/0 - """.format( - transport=transport, options=options - ) + """.format(transport=transport, options=options) ) ) diff --git a/tests/integrations/fastapi/test_fastapi.py b/tests/integrations/fastapi/test_fastapi.py index 3d79da92cc..a69978ded4 100644 --- a/tests/integrations/fastapi/test_fastapi.py +++ b/tests/integrations/fastapi/test_fastapi.py @@ -271,9 +271,9 @@ def test_response_status_code_ok_in_transaction_context(sentry_init, capture_env assert transaction["type"] == "transaction" assert len(transaction["contexts"]) > 0 - assert ( - "response" in transaction["contexts"].keys() - ), "Response context not found in transaction" + assert "response" in transaction["contexts"].keys(), ( + "Response context not found in transaction" + ) assert transaction["contexts"]["response"]["status_code"] == 200 @@ -307,9 +307,9 @@ def test_response_status_code_error_in_transaction_context( assert transaction["type"] == "transaction" assert len(transaction["contexts"]) > 0 - assert ( - "response" in transaction["contexts"].keys() - ), "Response context not found in transaction" + assert "response" in transaction["contexts"].keys(), ( + "Response context not found in transaction" + ) assert transaction["contexts"]["response"]["status_code"] == 500 @@ -338,9 +338,9 @@ def test_response_status_code_not_found_in_transaction_context( assert transaction["type"] == "transaction" assert len(transaction["contexts"]) > 0 - assert ( - "response" in transaction["contexts"].keys() - ), "Response context not found in transaction" + assert "response" in transaction["contexts"].keys(), ( + "Response context not found in transaction" + ) assert transaction["contexts"]["response"]["status_code"] == 404 diff --git a/tests/integrations/flask/test_flask.py b/tests/integrations/flask/test_flask.py index 49ee684797..e117b98ca9 100644 --- a/tests/integrations/flask/test_flask.py +++ b/tests/integrations/flask/test_flask.py @@ -943,9 +943,9 @@ def test_response_status_code_ok_in_transaction_context( assert transaction["type"] == "transaction" assert len(transaction["contexts"]) > 0 - assert ( - "response" in transaction["contexts"].keys() - ), "Response context not found in transaction" + assert "response" in transaction["contexts"].keys(), ( + "Response context not found in transaction" + ) assert transaction["contexts"]["response"]["status_code"] == 200 @@ -970,9 +970,9 @@ def test_response_status_code_not_found_in_transaction_context( assert transaction["type"] == "transaction" assert len(transaction["contexts"]) > 0 - assert ( - "response" in transaction["contexts"].keys() - ), "Response context not found in transaction" + assert "response" in transaction["contexts"].keys(), ( + "Response context not found in transaction" + ) assert transaction["contexts"]["response"]["status_code"] == 404 diff --git a/tests/integrations/gcp/test_gcp.py b/tests/integrations/gcp/test_gcp.py index 22d104c817..d088f134fe 100644 --- a/tests/integrations/gcp/test_gcp.py +++ b/tests/integrations/gcp/test_gcp.py @@ -101,14 +101,14 @@ def inner(code, subprocess_kwargs=()): subprocess.check_call( [sys.executable, "setup.py", "sdist", "-d", os.path.join(tmpdir, "..")], - **subprocess_kwargs + **subprocess_kwargs, ) subprocess.check_call( "pip install ../*.tar.gz -t .", cwd=tmpdir, shell=True, - **subprocess_kwargs + **subprocess_kwargs, ) stream = os.popen("python {}/main.py".format(tmpdir)) @@ -280,7 +280,8 @@ def cloud_function(functionhandler, event): def test_traces_sampler_gets_correct_values_in_sampling_context( - run_cloud_function, DictionaryContaining # noqa:N803 + run_cloud_function, + DictionaryContaining, # noqa:N803 ): # TODO: There are some decent sized hacks below. For more context, see the # long comment in the test of the same name in the AWS integration. The diff --git a/tests/integrations/gql/test_gql.py b/tests/integrations/gql/test_gql.py index f87fb974d0..147f7a06a8 100644 --- a/tests/integrations/gql/test_gql.py +++ b/tests/integrations/gql/test_gql.py @@ -44,16 +44,16 @@ def _make_erroneous_query(capture_events): with pytest.raises(TransportQueryError): _execute_mock_query(response_json) - assert ( - len(events) == 1 - ), "the sdk captured %d events, but 1 event was expected" % len(events) + assert len(events) == 1, ( + "the sdk captured %d events, but 1 event was expected" % len(events) + ) (event,) = events (exception,) = event["exception"]["values"] - assert ( - exception["type"] == "TransportQueryError" - ), "%s was captured, but we expected a TransportQueryError" % exception(type) + assert exception["type"] == "TransportQueryError", ( + "%s was captured, but we expected a TransportQueryError" % exception(type) + ) assert "request" in event @@ -79,12 +79,12 @@ def test_real_gql_request_no_error(sentry_init, capture_events): result = _execute_mock_query(response_json) - assert ( - result == response_data - ), "client.execute returned a different value from what it received from the server" - assert ( - len(events) == 0 - ), "the sdk captured an event, even though the query was successful" + assert result == response_data, ( + "client.execute returned a different value from what it received from the server" + ) + assert len(events) == 0, ( + "the sdk captured an event, even though the query was successful" + ) def test_real_gql_request_with_error_no_pii(sentry_init, capture_events): diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index ba2575ce59..4fd5275fb7 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -328,7 +328,8 @@ def test_option_trace_propagation_targets( integrations=[HttpxIntegration()], ) - with sentry_sdk.start_transaction(): # Must be in a transaction to propagate headers + # Must be in a transaction to propagate headers + with sentry_sdk.start_transaction(): if asyncio.iscoroutinefunction(httpx_client.get): asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) else: diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index af4b6b8c56..ba49b2e508 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -223,9 +223,9 @@ def test_langchain_agent( assert "5" in chat_spans[1]["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] # Verify tool calls are recorded when PII is enabled - assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in chat_spans[0].get( - "data", {} - ), "Tool calls should be recorded when send_default_pii=True and include_prompts=True" + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in chat_spans[0].get("data", {}), ( + "Tool calls should be recorded when send_default_pii=True and include_prompts=True" + ) tool_calls_data = chat_spans[0]["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS] assert isinstance(tool_calls_data, (list, str)) # Could be serialized if isinstance(tool_calls_data, str): @@ -261,9 +261,9 @@ def test_langchain_agent( span_data = chat_span.get("data", {}) if SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS in span_data: tools_data = span_data[SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS] - assert ( - tools_data is not None - ), "Available tools should always be recorded regardless of PII settings" + assert tools_data is not None, ( + "Available tools should always be recorded regardless of PII settings" + ) def test_langchain_error(sentry_init, capture_events): @@ -537,9 +537,9 @@ def test_span_map_is_instance_variable(): callback2 = SentryLangchainCallback(max_span_map_size=100, include_prompts=True) # Verify they have different span_map instances - assert ( - callback1.span_map is not callback2.span_map - ), "span_map should be an instance variable, not shared between instances" + assert callback1.span_map is not callback2.span_map, ( + "span_map should be an instance variable, not shared between instances" + ) def test_langchain_callback_manager(sentry_init): diff --git a/tests/integrations/langgraph/test_langgraph.py b/tests/integrations/langgraph/test_langgraph.py index 5e35f772f5..1510305b06 100644 --- a/tests/integrations/langgraph/test_langgraph.py +++ b/tests/integrations/langgraph/test_langgraph.py @@ -333,7 +333,6 @@ async def original_ainvoke(self, *args, **kwargs): async def run_test(): with start_transaction(): - wrapped_ainvoke = _wrap_pregel_ainvoke(original_ainvoke) result = await wrapped_ainvoke(pregel, test_state) return result @@ -394,7 +393,6 @@ def original_invoke(self, *args, **kwargs): raise Exception("Graph execution failed") with start_transaction(), pytest.raises(Exception, match="Graph execution failed"): - wrapped_invoke = _wrap_pregel_invoke(original_invoke) wrapped_invoke(pregel, test_state) @@ -426,7 +424,6 @@ async def run_error_test(): with start_transaction(), pytest.raises( Exception, match="Async graph execution failed" ): - wrapped_ainvoke = _wrap_pregel_ainvoke(original_ainvoke) await wrapped_ainvoke(pregel, test_state) @@ -482,7 +479,6 @@ def test_pregel_invoke_with_different_graph_names( pregel = MockPregelInstance(graph_name) if graph_name else MockPregelInstance() if not graph_name: - delattr(pregel, "name") delattr(pregel, "graph_name") @@ -490,7 +486,6 @@ def original_invoke(self, *args, **kwargs): return {"result": "test"} with start_transaction(): - wrapped_invoke = _wrap_pregel_invoke(original_invoke) wrapped_invoke(pregel, {"messages": []}) diff --git a/tests/integrations/opentelemetry/test_span_processor.py b/tests/integrations/opentelemetry/test_span_processor.py index ec5cf6af23..cbee14f4d6 100644 --- a/tests/integrations/opentelemetry/test_span_processor.py +++ b/tests/integrations/opentelemetry/test_span_processor.py @@ -549,8 +549,10 @@ def test_pruning_old_spans_on_start(): current_time_minutes = int(time.time() / 60) span_processor.open_spans = { current_time_minutes - 3: {"111111111abcdef"}, # should stay - current_time_minutes - - 11: {"2222222222abcdef", "3333333333abcdef"}, # should go + current_time_minutes - 11: { + "2222222222abcdef", + "3333333333abcdef", + }, # should go } span_processor.on_start(otel_span, parent_context) @@ -599,8 +601,10 @@ def test_pruning_old_spans_on_end(): span_processor.open_spans = { current_time_minutes: {"1234567890abcdef"}, # should go (because it is closed) current_time_minutes - 3: {"111111111abcdef"}, # should stay - current_time_minutes - - 11: {"2222222222abcdef", "3333333333abcdef"}, # should go + current_time_minutes - 11: { + "2222222222abcdef", + "3333333333abcdef", + }, # should go } span_processor.on_end(otel_span) diff --git a/tests/integrations/rq/test_rq.py b/tests/integrations/rq/test_rq.py index e445b588be..23603ad91d 100644 --- a/tests/integrations/rq/test_rq.py +++ b/tests/integrations/rq/test_rq.py @@ -97,7 +97,9 @@ def test_transport_shutdown(sentry_init, capture_events_forksafe): def test_transaction_with_error( - sentry_init, capture_events, DictionaryContaining # noqa:N803 + sentry_init, + capture_events, + DictionaryContaining, # noqa:N803 ): sentry_init(integrations=[RqIntegration()], traces_sample_rate=1.0) events = capture_events() @@ -195,7 +197,9 @@ def test_tracing_disabled( def test_transaction_no_error( - sentry_init, capture_events, DictionaryContaining # noqa:N803 + sentry_init, + capture_events, + DictionaryContaining, # noqa:N803 ): sentry_init(integrations=[RqIntegration()], traces_sample_rate=1.0) events = capture_events() @@ -222,7 +226,9 @@ def test_transaction_no_error( def test_traces_sampler_gets_correct_values_in_sampling_context( - sentry_init, DictionaryContaining, ObjectDescribedBy # noqa:N803 + sentry_init, + DictionaryContaining, + ObjectDescribedBy, # noqa:N803 ): traces_sampler = mock.Mock(return_value=True) sentry_init(integrations=[RqIntegration()], traces_sampler=traces_sampler) diff --git a/tests/integrations/spark/test_spark.py b/tests/integrations/spark/test_spark.py index 0ea2770b89..c5bb70f4d1 100644 --- a/tests/integrations/spark/test_spark.py +++ b/tests/integrations/spark/test_spark.py @@ -90,7 +90,6 @@ def test_initialize_spark_integration_after_spark_context_init( @pytest.fixture def sentry_listener(): - listener = SentryListener() return listener diff --git a/tests/integrations/sys_exit/test_sys_exit.py b/tests/integrations/sys_exit/test_sys_exit.py index 81a950c7c0..a9909ae3c2 100644 --- a/tests/integrations/sys_exit/test_sys_exit.py +++ b/tests/integrations/sys_exit/test_sys_exit.py @@ -66,6 +66,6 @@ def test_sys_exit_integration_not_auto_enabled(sentry_init, capture_events): "sys.exit should not be patched, but it must have been because it did not raise SystemExit" ) - assert ( - len(events) == 0 - ), "No events should have been captured because sys.exit should not have been patched" + assert len(events) == 0, ( + "No events should have been captured because sys.exit should not have been patched" + ) diff --git a/tests/integrations/threading/test_threading.py b/tests/integrations/threading/test_threading.py index 4577c846d8..799298910b 100644 --- a/tests/integrations/threading/test_threading.py +++ b/tests/integrations/threading/test_threading.py @@ -208,9 +208,9 @@ def do_some_work(): t.join() # check if the initial scope data is not modified by the started thread - assert initial_iso_scope._tags == { - "initial_tag": "initial_value" - }, "The isolation scope in the main thread should not be modified by the started thread." + assert initial_iso_scope._tags == {"initial_tag": "initial_value"}, ( + "The isolation scope in the main thread should not be modified by the started thread." + ) @pytest.mark.parametrize( diff --git a/tests/integrations/unleash/testutils.py b/tests/integrations/unleash/testutils.py index 07b065e2f0..4e91b6190b 100644 --- a/tests/integrations/unleash/testutils.py +++ b/tests/integrations/unleash/testutils.py @@ -34,7 +34,6 @@ def mock_unleash_client(): class MockUnleashClient: - def __init__(self, *a, **kw): self.features = { "hello": True, diff --git a/tests/integrations/unraisablehook/test_unraisablehook.py b/tests/integrations/unraisablehook/test_unraisablehook.py index 2f97886ce8..dbe8164cf5 100644 --- a/tests/integrations/unraisablehook/test_unraisablehook.py +++ b/tests/integrations/unraisablehook/test_unraisablehook.py @@ -42,9 +42,7 @@ def capture_envelope(self, envelope): undeletable = Undeletable() del undeletable - """.format( - transport=transport, options=options - ) + """.format(transport=transport, options=options) ) ) diff --git a/tests/integrations/wsgi/test_wsgi.py b/tests/integrations/wsgi/test_wsgi.py index 656fc1757f..a741d1c57b 100644 --- a/tests/integrations/wsgi/test_wsgi.py +++ b/tests/integrations/wsgi/test_wsgi.py @@ -136,7 +136,10 @@ def test_keyboard_interrupt_is_captured(sentry_init, capture_events): def test_transaction_with_error( - sentry_init, crashing_app, capture_events, DictionaryContaining # noqa:N803 + sentry_init, + crashing_app, + capture_events, + DictionaryContaining, # noqa:N803 ): def dogpark(environ, start_response): raise ValueError("Fetch aborted. The ball was not returned.") @@ -173,7 +176,9 @@ def dogpark(environ, start_response): def test_transaction_no_error( - sentry_init, capture_events, DictionaryContaining # noqa:N803 + sentry_init, + capture_events, + DictionaryContaining, # noqa:N803 ): def dogpark(environ, start_response): start_response("200 OK", []) diff --git a/tests/test_basics.py b/tests/test_basics.py index 45303c9a59..b0b577b796 100644 --- a/tests/test_basics.py +++ b/tests/test_basics.py @@ -1218,6 +1218,6 @@ def recurse(): # On my machine, it takes about 100-200ms to capture the exception, # so this limit should be generous enough. - assert ( - capture_end_time - capture_start_time < 10**9 * 2 - ), "stacktrace capture took too long, check that frame limit is set correctly" + assert capture_end_time - capture_start_time < 10**9 * 2, ( + "stacktrace capture took too long, check that frame limit is set correctly" + ) diff --git a/tests/test_client.py b/tests/test_client.py index a02ea6e56a..ff3f61d702 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -608,9 +608,7 @@ def capture_envelope(self, envelope): for _ in range({num_messages}): capture_message("HI") - """.format( - transport=transport, options=options, num_messages=num_messages - ) + """.format(transport=transport, options=options, num_messages=num_messages) ) ) @@ -1194,9 +1192,9 @@ def test_spotlight_option( client = sentry_sdk.get_client() url = client.spotlight.url if client.spotlight else None - assert ( - url == spotlight_url_expected - ), f"With config {client_option} and env {env_var_value}" + assert url == spotlight_url_expected, ( + f"With config {client_option} and env {env_var_value}" + ) class IssuesSamplerTestConfig: diff --git a/tests/test_conftest.py b/tests/test_conftest.py index 3b8cd098f5..a36fe17894 100644 --- a/tests/test_conftest.py +++ b/tests/test_conftest.py @@ -22,7 +22,9 @@ ], ) def test_string_containing( - test_string, expected_result, StringContaining # noqa: N803 + test_string, + expected_result, + StringContaining, # noqa: N803 ): assert (test_string == StringContaining("dogs")) is expected_result @@ -46,7 +48,9 @@ def test_string_containing( ], ) def test_dictionary_containing( - test_dict, expected_result, DictionaryContaining # noqa: N803 + test_dict, + expected_result, + DictionaryContaining, # noqa: N803 ): assert ( test_dict == DictionaryContaining({"dogs": "yes", "cats": "maybe"}) diff --git a/tests/test_gevent.py b/tests/test_gevent.py index d330760adf..05fa6ed2e8 100644 --- a/tests/test_gevent.py +++ b/tests/test_gevent.py @@ -104,13 +104,11 @@ def test_transport_works_gevent( (compression_level is None) or ( # setting compression level to 0 means don't compress - compression_level - > 0 + compression_level > 0 ) ) and ( # if we couldn't resolve to a known algo, we don't compress - compression_algo - != "" + compression_algo != "" ) assert capturing_server.captured[0].compressed == should_compress diff --git a/tests/test_propagationcontext.py b/tests/test_propagationcontext.py index 078a69c72b..e014012956 100644 --- a/tests/test_propagationcontext.py +++ b/tests/test_propagationcontext.py @@ -155,8 +155,7 @@ def mock_random_class(seed): ) assert ( - ctx.dynamic_sampling_context["sample_rand"] - == f"{expected_interval[0]:.6f}" # noqa: E231 + ctx.dynamic_sampling_context["sample_rand"] == f"{expected_interval[0]:.6f}" # noqa: E231 ) assert mock_randrange.call_count == 1 assert mock_randrange.call_args[0] == ( diff --git a/tests/test_transport.py b/tests/test_transport.py index e493515e9a..68669fa24d 100644 --- a/tests/test_transport.py +++ b/tests/test_transport.py @@ -137,13 +137,11 @@ def test_transport_works( (compression_level is None) or ( # setting compression level to 0 means don't compress - compression_level - > 0 + compression_level > 0 ) ) and ( # if we couldn't resolve to a known algo, we don't compress - compression_algo - != "" + compression_algo != "" ) assert capturing_server.captured[0].compressed == should_compress diff --git a/tests/test_utils.py b/tests/test_utils.py index b268fbd57b..e1c6786e1b 100644 --- a/tests/test_utils.py +++ b/tests/test_utils.py @@ -196,9 +196,9 @@ def test_datetime_from_isoformat_with_py_36_or_lower(input_str, expected_output) ], ) def test_env_to_bool(env_var_value, strict, expected): - assert ( - env_to_bool(env_var_value, strict=strict) == expected - ), f"Value: {env_var_value}, strict: {strict}" + assert env_to_bool(env_var_value, strict=strict) == expected, ( + f"Value: {env_var_value}, strict: {strict}" + ) @pytest.mark.parametrize( diff --git a/tests/tracing/test_sampling.py b/tests/tracing/test_sampling.py index 1761a3dbac..63da5a1399 100644 --- a/tests/tracing/test_sampling.py +++ b/tests/tracing/test_sampling.py @@ -212,7 +212,8 @@ def mock_set_initial_sampling_decision(_, sampling_context): def test_passes_custom_sampling_context_from_start_transaction_to_traces_sampler( - sentry_init, DictionaryContaining # noqa: N803 + sentry_init, + DictionaryContaining, # noqa: N803 ): traces_sampler = mock.Mock() sentry_init(traces_sampler=traces_sampler) @@ -251,7 +252,9 @@ def test_sample_rate_affects_errors(sentry_init, capture_events): ], ) def test_warns_and_sets_sampled_to_false_on_invalid_traces_sampler_return_value( - sentry_init, traces_sampler_return_value, StringContaining # noqa: N803 + sentry_init, + traces_sampler_return_value, + StringContaining, # noqa: N803 ): sentry_init(traces_sampler=mock.Mock(return_value=traces_sampler_return_value)) diff --git a/tox.ini b/tox.ini index b993397389..e310a6245b 100644 --- a/tox.ini +++ b/tox.ini @@ -91,7 +91,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.38 + {py3.9,py3.12,py3.13}-boto3-v1.40.39 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -157,7 +157,7 @@ envlist = {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.47.5 {py3.7,py3.11,py3.12}-grpc-v1.62.3 - {py3.9,py3.12,py3.13}-grpc-v1.75.0 + {py3.9,py3.12,py3.13}-grpc-v1.75.1 {py3.6,py3.8,py3.9}-httpx-v0.16.1 {py3.6,py3.9,py3.10}-httpx-v0.20.0 @@ -396,7 +396,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.38: boto3==1.40.38 + boto3-v1.40.39: boto3==1.40.39 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -482,7 +482,7 @@ deps = grpc-v1.32.0: grpcio==1.32.0 grpc-v1.47.5: grpcio==1.47.5 grpc-v1.62.3: grpcio==1.62.3 - grpc-v1.75.0: grpcio==1.75.0 + grpc-v1.75.1: grpcio==1.75.1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -803,11 +803,9 @@ basepython = py3.12: python3.12 py3.13: python3.13 - # Python version is pinned here because flake8 actually behaves differently - # depending on which version is used. You can patch this out to point to - # some random Python 3 binary, but then you get guaranteed mismatches with - # CI. Other tools such as mypy and black have options that pin the Python - # version. + # Python version is pinned here for consistency across environments. + # Tools like ruff and mypy have options that pin the target Python + # version (configured in pyproject.toml), ensuring consistent behavior. linters: python3.12 commands = @@ -823,6 +821,6 @@ commands = [testenv:linters] commands = - flake8 tests sentry_sdk - black --check tests sentry_sdk + ruff check tests sentry_sdk + ruff format --check tests sentry_sdk mypy sentry_sdk From 7e493b481e2cf9b5ded12c3c66be22abe059168f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 26 Sep 2025 15:44:24 +0200 Subject: [PATCH 509/591] docs: Update CONTRIBUTING.md (#4870) ### Description Removed hub mention, some obsolete/unclear instructions, added a note about defensiveness, toxgen. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- CONTRIBUTING.md | 29 +++++++++-------------------- 1 file changed, 9 insertions(+), 20 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 89330087d9..753b169214 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -78,7 +78,7 @@ We test against a number of Python language and library versions, which are auto The tox CLI tool is required to run the tests locally. Follow [the installation instructions](https://tox.wiki/en/latest/installation.html) for tox. Dependencies are installed for you when you run the command below, but _you_ need to bring an appropriate Python interpreter. -[Pyenv](https://github.com/pyenv/pyenv) is a cross-platform utility for managing Python versions. You can also use a conventional package manager, but not all versions may be distributed in the package manager of your choice. For macOS, Versions 3.8 and up can be installed with Homebrew. +[Pyenv](https://github.com/pyenv/pyenv) is a cross-platform utility for managing Python versions. You can also use a conventional package manager, but not all versions may be distributed in the package manager of your choice. For macOS, versions 3.8 and up can be installed with Homebrew. An environment consists of the Python major and minor version and the library name and version. The exception to the rule is that you can provide `common` instead of the library information. The environments tied to a specific library usually run the corresponding test suite, while `common` targets all tests but skips those that require uninstalled dependencies. @@ -109,29 +109,20 @@ tox -p auto -o -e -- ## Adding a New Integration 1. Write the integration. - - - Instrument all application instances by default. Prefer global signals/patches instead of configuring a specific instance. Don't make the user pass anything to your integration for anything to work. Aim for zero configuration. - - - Everybody monkeypatches. That means: - - - Make sure to think about conflicts with other monkeypatches when monkeypatching. - - - You don't need to feel bad about it. - + - Instrument all application instances by default. Prefer global signals/patches. + - Don't make the user pass anything to your integration for anything to work. Aim for zero configuration. + - Everybody monkeypatches. That means you don't need to feel bad about it. - Make sure your changes don't break end user contracts. The SDK should never alter the expected behavior of the underlying library or framework from the user's perspective and it shouldn't have any side effects. - - - Avoid modifying the hub, registering a new client or the like. The user drives the client, and the client owns integrations. - - - Allow the user to turn off the integration by changing the client. Check `Hub.current.get_integration(MyIntegration)` from within your signal handlers to see if your integration is still active before you do anything impactful (such as sending an event). + - Be defensive. Don't assume the code you're patching will stay the same forever, especially if it's an internal function. Allow for future variability whenever it makes sense. + - Avoid registering a new client or the like. The user drives the client, and the client owns integrations. + - Allow the user to turn off the integration by changing the client. Check `sentry_sdk.get_client().get_integration(MyIntegration)` from within your signal handlers to see if your integration is still active before you do anything impactful (such as sending an event). 2. Write tests. - - - Consider the minimum versions supported, and test each version in a separate env in `tox.ini`. - + - Consider the minimum versions supported, and document in `_MIN_VERSIONS` in `integrations/__init__.py`. - Create a new folder in `tests/integrations/`, with an `__init__` file that skips the entire suite if the package is not installed. + - Add the test suite to the script generating our test matrix. See [`scripts/populate_tox/README.md`](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md#add-a-new-test-suite). 3. Update package metadata. - - We use `extras_require` in `setup.py` to communicate minimum version requirements for integrations. People can use this in combination with tools like Poetry or Pipenv to detect conflicts between our supported versions and their used versions programmatically. Do not set upper bounds on version requirements as people are often faster in adopting new versions of a web framework than we are in adding them to the test matrix or our package metadata. @@ -140,8 +131,6 @@ tox -p auto -o -e -- 5. Merge docs after new version has been released. The docs are built and deployed after each merge, so your changes should go live in a few minutes. -6. (optional, if possible) Update data in [`sdk_updates.py`](https://github.com/getsentry/sentry/blob/master/src/sentry/sdk_updates.py) to give users in-app suggestions to use your integration. This step will only apply to some integrations. - ## Releasing a New Version _(only relevant for Python SDK core team)_ From e13b7a7b4b022df24af3eaaa93884d9c9ea3b6a7 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 26 Sep 2025 15:44:45 +0200 Subject: [PATCH 510/591] feat: Add script to determine lowest supported versions (#4867) ### Description With toxgen, we now have an automated way of detecting the effective minimum version of a framework. This depends on both package metadata (e.g., the min version needs to support at least some of the Python versions we support) as well as on our test matrix and explicitly defined lower bounds (in `sentry_sdk/integrations/__init__.py`'s `_MIN_VERSIONS`). When we release a new major in which we drop support for a Python version, we can use this script to automatically generate the new `_MIN_VERSIONS`. Example output: ``` Effective minimal versions: - The format is the same as _MIN_VERSIONS in sentry_sdk/integrations/__init__.py for easy replacing. - When updating these, make sure to also update: - The docs page for the integration - The lower bounds in extras_require in setup.py "aiohttp": (3, 4, 4), "anthropic": (0, 16, 0), "ariadne": (0, 20, 1), "arq": (0, 23), "asyncpg": (0, 23, 0), "beam": (2, 14, 0), "boto3": (1, 12, 49), "bottle": (0, 12, 25), "celery": (4, 4, 7), "chalice": (1, 16, 0), "clickhouse_driver": (0, 2, 9), "cohere": (5, 4, 0), "django": (1, 11, 29), "dramatiq": (1, 9, 0), "falcon": (1, 4, 1), "fastapi": (0, 79, 1), "flask": (1, 1, 4), "gql": (3, 4, 1), "graphene": (3, 3), "grpc": (1, 32, 0), "httpx": (0, 16, 1), "huey": (2, 1, 3), "huggingface_hub": (0, 24, 7), "langchain": (0, 1, 20), "langgraph": (0, 6, 7), "launchdarkly": (9, 8, 1), "litestar": (2, 0, 1), "loguru": (0, 7, 3), "openai": (1, 0, 1), "openai_agents": (0, 0, 19), "openfeature": (0, 7, 5), "pure_eval": (0, 0, 3), "pymongo": (3, 5, 1), "pyramid": (1, 8, 6), "quart": (0, 16, 3), "ray": (2, 7, 2), "redis": (2, 10, 6), "redis_py_cluster_legacy": (1, 3, 6), "requests": (2, 12, 5), "rq": (0, 8, 2), "sanic": (0, 8, 3), "spark": (3, 0, 3), "sqlalchemy": (1, 3, 24), "starlette": (0, 16, 0), "starlite": (1, 48, 1), "statsig": (0, 55, 3), "strawberry": (0, 209, 8), "tornado": (6, 0, 4), "trytond": (4, 6, 22), "typer": (0, 15, 4), "unleash": (6, 0, 1), ``` #### Issues Ref https://github.com/getsentry/sentry-python/issues/4047 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/populate_tox.py | 4 +- scripts/update_integration_support.py | 55 +++++++++++++++++++++++++++ 2 files changed, 58 insertions(+), 1 deletion(-) create mode 100644 scripts/update_integration_support.py diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index d625c1da72..c0bf7f1a9f 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -665,7 +665,7 @@ def _normalize_release(release: dict) -> dict: return normalized -def main(fail_on_changes: bool = False) -> None: +def main(fail_on_changes: bool = False) -> dict[str, list]: """ Generate tox.ini from the tox.jinja template. @@ -825,6 +825,8 @@ def main(fail_on_changes: bool = False) -> None: "files to reflect the new test targets." ) + return packages + if __name__ == "__main__": fail_on_changes = len(sys.argv) == 2 and sys.argv[1] == "--fail-on-changes" diff --git a/scripts/update_integration_support.py b/scripts/update_integration_support.py new file mode 100644 index 0000000000..7e686b2729 --- /dev/null +++ b/scripts/update_integration_support.py @@ -0,0 +1,55 @@ +""" +Small utility to determine the actual minimum supported version of each framework/library. +""" + +import os +import sys +from textwrap import dedent + +populate_tox_dir = os.path.join( + os.path.dirname(os.path.abspath(__file__)), "populate_tox" +) +sys.path.append(populate_tox_dir) + +from populate_tox import main + + +def update(): + print("Running populate_tox.py...") + packages = main() + + print("Figuring out the lowest supported version of integrations...") + min_versions = [] + + for _, integrations in packages.items(): + for integration in integrations: + min_versions.append( + (integration["integration_name"], str(integration["releases"][0])) + ) + + min_versions = sorted( + set( + [ + (integration, tuple([int(v) for v in min_version.split(".")])) + for integration, min_version in min_versions + ] + ) + ) + + print() + print("Effective minimal versions:") + print( + dedent(""" + - The format is the same as _MIN_VERSIONS in sentry_sdk/integrations/__init__.py for easy replacing. + - When updating these, make sure to also update: + - The docs page for the integration + - The lower bounds in extras_require in setup.py + """) + ) + print() + for integration, min_version in min_versions: + print(f'"{integration}": {min_version},') + + +if __name__ == "__main__": + update() From 5e24a79ffe95abcf0dfaee2fa7dd7b4cc72e0b2d Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 29 Sep 2025 06:54:47 +0000 Subject: [PATCH 511/591] build(deps): update shibuya requirement from <2025.9.22 (#4871) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Updates the requirements on [shibuya](https://github.com/lepture/shibuya) to permit the latest version.
Release notes

Sourced from shibuya's releases.

2025.9.25

   🐞 Bug Fixes

    View changes on GitHub
Changelog

Sourced from shibuya's changelog.

2025.9.25

  • Fix: Add dark theme support for sphinxcontrib-mermaid extension.
  • Fix: Fix building error when toc is empty, via :issue:91.
  • Fix: Improve shibuya.sponsors extension.

2025.9.24

  • New: add a shibuya.sponsors extension.
  • Fix: remove useless pygments css files.
  • Fix: hide username when repository name is too long.
  • Fix: update head backrop blur style.
  • Fix: update base template's block theme_scripts to themescripts.

2025.9.23

  • Fix: fix pageurl context when it is None, via :issue:90.

2025.9.22

  • Refactor: Update python code structure.
  • Fix: Support for Pygments extension, via :issue:89.
  • Fix: Build pygments styles dynamically.
  • Fix: Fix text overflow for math block and pre.
  • Fix: Fix style for buysellads.

2025.8.16

  • Update: Use tailwindcss v4.
  • Fix: Improve copybutton's position in code blocks.
  • Fix: Improve style for sphinx-contributors.
  • Fix: Fix JS errors when scrolling on landing page.

2025.7.24

  • Fix: Improve accessibility for nav links, you can now using keyboard to open sub nav links.
  • Fix: Update sphinx-design's style for iconify-icon.

2025.7.14

  • New: Add data-accent-color style for :ref:sphinx-iconify.
  • New: Add outline variant style for sphinx-design's tab-set.

... (truncated)

Commits
  • e3832e3 chore: release 2025.9.25
  • fb9446b docs: update changelog
  • b61847f chore: ruff format
  • 4cea3db docs: add a sponsors page
  • 214dde7 fix: update sponsors style
  • f623d24 fix: improve SponsorsDirective
  • cf082e1 fix: prevent building error when toc is empty
  • 8c6e8b7 fix: dark theme support for sphinxcontrib-mermaid extension
  • 0055315 chore: release 2025.9.24
  • 53d76dc fix: update block theme_scripts to themescripts
  • Additional commits viewable in compare view

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
--- > [!NOTE] > Removes the version cap on `shibuya` in `requirements-docs.txt` to allow the latest release. > > Written by [Cursor Bugbot](https://cursor.com/dashboard?tab=bugbot) for commit cbacc632a603a969a2042355a17349200d4d27e7. This will update automatically on new commits. Configure [here](https://cursor.com/dashboard?tab=bugbot). --------- Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- requirements-docs.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements-docs.txt b/requirements-docs.txt index b98ad4834a..81e04ba3ef 100644 --- a/requirements-docs.txt +++ b/requirements-docs.txt @@ -1,5 +1,5 @@ gevent -shibuya<2025.9.22 +shibuya sphinx<8.2 sphinx-autodoc-typehints[type_comments]>=1.8.0 typing-extensions From 9b58d3155db85947cb951649f8651d3afdf559f9 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 29 Sep 2025 06:57:35 +0000 Subject: [PATCH 512/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(09/29)=20(#4872)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --- > [!NOTE] > Bumps test targets and dependency pins for anthropic, boto3, fastapi, and redis to their latest versions across releases.jsonl and tox.ini. > > - **Integrations matrix updates** > - **AI**: > - `anthropic`: `0.68.0` → `0.68.1` (envlist and deps) > - **Cloud**: > - `boto3`: `1.40.39` → `1.40.40` (envlist and deps) > - **Web**: > - `fastapi`: `0.117.1` → `0.118.0` (envlist and deps) > - **DB/Cache**: > - `redis`: `7.0.0b1` → `7.0.0b2` (envlist and deps) > > Written by [Cursor Bugbot](https://cursor.com/dashboard?tab=bugbot) for commit 6c871c3bb56dac1e8540e1edb2dcc9f9bde2b451. This will update automatically on new commits. Configure [here](https://cursor.com/dashboard?tab=bugbot). --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/releases.jsonl | 8 ++++---- tox.ini | 16 ++++++++-------- 2 files changed, 12 insertions(+), 12 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index fa24b089cd..3532b61c75 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -26,7 +26,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.33.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.50.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.68.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.68.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.39", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.40", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -66,7 +66,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.117.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} @@ -153,7 +153,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}} diff --git a/tox.ini b/tox.ini index e310a6245b..2354c66c7c 100644 --- a/tox.ini +++ b/tox.ini @@ -48,7 +48,7 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.33.1 {py3.8,py3.11,py3.12}-anthropic-v0.50.0 - {py3.8,py3.12,py3.13}-anthropic-v0.68.0 + {py3.8,py3.12,py3.13}-anthropic-v0.68.1 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 @@ -91,7 +91,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.39 + {py3.9,py3.12,py3.13}-boto3-v1.40.40 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -114,7 +114,7 @@ envlist = {py3.7,py3.10,py3.11}-redis-v4.6.0 {py3.8,py3.11,py3.12}-redis-v5.3.1 {py3.9,py3.12,py3.13}-redis-v6.4.0 - {py3.9,py3.12,py3.13}-redis-v7.0.0b1 + {py3.9,py3.12,py3.13}-redis-v7.0.0b2 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 @@ -218,7 +218,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.117.1 + {py3.8,py3.12,py3.13}-fastapi-v0.118.0 # ~~~ Web 2 ~~~ @@ -334,7 +334,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.33.1: anthropic==0.33.1 anthropic-v0.50.0: anthropic==0.50.0 - anthropic-v0.68.0: anthropic==0.68.0 + anthropic-v0.68.1: anthropic==0.68.1 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.33.1: httpx<0.28.0 @@ -396,7 +396,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.39: boto3==1.40.39 + boto3-v1.40.40: boto3==1.40.40 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -423,7 +423,7 @@ deps = redis-v4.6.0: redis==4.6.0 redis-v5.3.1: redis==5.3.1 redis-v6.4.0: redis==6.4.0 - redis-v7.0.0b1: redis==7.0.0b1 + redis-v7.0.0b2: redis==7.0.0b2 redis: fakeredis!=1.7.4 redis: pytest<8.0.0 redis-v4.6.0: fakeredis<2.31.0 @@ -599,7 +599,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.92.0: fastapi==0.92.0 fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.117.1: fastapi==0.117.1 + fastapi-v0.118.0: fastapi==0.118.0 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart From 2872675b8b37b3a654fc9cac5590f0de64183c9f Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Wed, 1 Oct 2025 08:53:19 +0200 Subject: [PATCH 513/591] fix(openai-agents): Move _set_agent_data call to ai_client_span function (#4876) ### Description We can set the agent data early, so that we get input model and other data, even if there's an exception while running the agent #### Issues Resolves: TET-1205 --- > [!NOTE] > Move `_set_agent_data` to `ai_client_span` so agent/model metadata is recorded when the span is created. > > - **Tracing (`sentry_sdk/integrations/openai_agents/spans/ai_client.py`)**: > - Set agent metadata in `ai_client_span` via `_set_agent_data` at span creation. > - Remove `_set_agent_data` from `update_ai_client_span` to avoid duplication. > > Written by [Cursor Bugbot](https://cursor.com/dashboard?tab=bugbot) for commit 27e94cc89cd48716d9949fd402e04dda0f932e98. This will update automatically on new commits. Configure [here](https://cursor.com/dashboard?tab=bugbot). --- sentry_sdk/integrations/openai_agents/spans/ai_client.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/openai_agents/spans/ai_client.py b/sentry_sdk/integrations/openai_agents/spans/ai_client.py index d325ae86e3..e215edfd26 100644 --- a/sentry_sdk/integrations/openai_agents/spans/ai_client.py +++ b/sentry_sdk/integrations/openai_agents/spans/ai_client.py @@ -28,12 +28,13 @@ def ai_client_span(agent, get_response_kwargs): # TODO-anton: remove hardcoded stuff and replace something that also works for embedding and so on span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + _set_agent_data(span, agent) + return span def update_ai_client_span(span, agent, get_response_kwargs, result): # type: (sentry_sdk.tracing.Span, Agent, dict[str, Any], Any) -> None - _set_agent_data(span, agent) _set_usage_data(span, result.usage) _set_input_data(span, get_response_kwargs) _set_output_data(span, result) From b838765160305aacc9b1ea0f42d6ca2e090c4b2e Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 1 Oct 2025 15:45:14 +0200 Subject: [PATCH 514/591] feat: Option to not trace HTTP requests based on status codes (#4869) Add `trace_ignore_status_codes` option to exclude HTTP requests from tracing if they have certain status codes. A debug-level log message is printed if a transaction is dropped because its status code is ignored. By default no status codes automatically cause transactions to be dropped. Closes https://github.com/getsentry/sentry-python/issues/4812 --------- Co-authored-by: Anton Pirker --- sentry_sdk/consts.py | 10 ++ sentry_sdk/tracing.py | 37 +++++- tests/tracing/test_ignore_status_codes.py | 139 ++++++++++++++++++++++ 3 files changed, 183 insertions(+), 3 deletions(-) create mode 100644 tests/tracing/test_ignore_status_codes.py diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 9e84dc3dd2..5bcc487037 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -40,6 +40,7 @@ class CompressionAlgo(Enum): from typing import Any from typing import Sequence from typing import Tuple + from typing import AbstractSet from typing_extensions import Literal from typing_extensions import TypedDict @@ -919,6 +920,7 @@ def __init__( max_stack_frames=DEFAULT_MAX_STACK_FRAMES, # type: Optional[int] enable_logs=False, # type: bool before_send_log=None, # type: Optional[Callable[[Log, Hint], Optional[Log]]] + trace_ignore_status_codes=frozenset(), # type: AbstractSet[int] ): # type: (...) -> None """Initialize the Sentry SDK with the given parameters. All parameters described here can be used in a call to `sentry_sdk.init()`. @@ -1307,6 +1309,14 @@ def __init__( function will be retained. If the function returns None, the log will not be sent to Sentry. + :param trace_ignore_status_codes: An optional property that disables tracing for + HTTP requests with certain status codes. + + Requests are not traced if the status code is contained in the provided set. + + If `trace_ignore_status_codes` is not provided, requests with any status code + may be traced. + :param _experiments: """ pass diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index a82b99ff4d..1697df1f22 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -30,6 +30,7 @@ from typing import Tuple from typing import Union from typing import TypeVar + from typing import Set from typing_extensions import TypedDict, Unpack @@ -970,6 +971,12 @@ def _get_scope_from_finish_args( return scope_or_hub + def _get_log_representation(self): + # type: () -> str + return "{op}transaction <{name}>".format( + op=("<" + self.op + "> " if self.op else ""), name=self.name + ) + def finish( self, scope=None, # type: Optional[sentry_sdk.Scope] @@ -1039,6 +1046,32 @@ def finish( super().finish(scope, end_timestamp) + status_code = self._data.get(SPANDATA.HTTP_STATUS_CODE) + if ( + status_code is not None + and status_code in client.options["trace_ignore_status_codes"] + ): + logger.debug( + "[Tracing] Discarding {transaction_description} because the HTTP status code {status_code} is matched by trace_ignore_status_codes: {trace_ignore_status_codes}".format( + transaction_description=self._get_log_representation(), + status_code=self._data[SPANDATA.HTTP_STATUS_CODE], + trace_ignore_status_codes=client.options[ + "trace_ignore_status_codes" + ], + ) + ) + if client.transport: + client.transport.record_lost_event( + "event_processor", data_category="transaction" + ) + + num_spans = len(self._span_recorder.spans) + 1 + client.transport.record_lost_event( + "event_processor", data_category="span", quantity=num_spans + ) + + self.sampled = False + if not self.sampled: # At this point a `sampled = None` should have already been resolved # to a concrete decision. @@ -1186,9 +1219,7 @@ def _set_initial_sampling_decision(self, sampling_context): """ client = sentry_sdk.get_client() - transaction_description = "{op}transaction <{name}>".format( - op=("<" + self.op + "> " if self.op else ""), name=self.name - ) + transaction_description = self._get_log_representation() # nothing to do if tracing is disabled if not has_tracing_enabled(client.options): diff --git a/tests/tracing/test_ignore_status_codes.py b/tests/tracing/test_ignore_status_codes.py new file mode 100644 index 0000000000..b2899e0ad9 --- /dev/null +++ b/tests/tracing/test_ignore_status_codes.py @@ -0,0 +1,139 @@ +import sentry_sdk +from sentry_sdk import start_transaction, start_span + +import pytest + +from collections import Counter + + +def test_no_ignored_codes(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + ) + events = capture_events() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", 404) + + assert len(events) == 1 + + +@pytest.mark.parametrize("status_code", [200, 404]) +def test_single_code_ignored(sentry_init, capture_events, status_code): + sentry_init( + traces_sample_rate=1.0, + trace_ignore_status_codes={ + 404, + }, + ) + events = capture_events() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", status_code) + + if status_code == 404: + assert not events + else: + assert len(events) == 1 + + +@pytest.mark.parametrize("status_code", [200, 305, 307, 399, 404]) +def test_range_ignored(sentry_init, capture_events, status_code): + sentry_init( + traces_sample_rate=1.0, + trace_ignore_status_codes=set( + range( + 305, + 400, + ), + ), + ) + events = capture_events() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", status_code) + + if 305 <= status_code <= 399: + assert not events + else: + assert len(events) == 1 + + +@pytest.mark.parametrize("status_code", [200, 301, 303, 355, 404]) +def test_variety_ignored(sentry_init, capture_events, status_code): + sentry_init( + traces_sample_rate=1.0, + trace_ignore_status_codes={ + 301, + 302, + 303, + *range( + 305, + 400, + ), + *range( + 401, + 405, + ), + }, + ) + events = capture_events() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", status_code) + + if ( + 301 <= status_code <= 303 + or 305 <= status_code <= 399 + or 401 <= status_code <= 404 + ): + assert not events + else: + assert len(events) == 1 + + +def test_transaction_not_ignored_when_status_code_has_invalid_type( + sentry_init, capture_events +): + sentry_init( + traces_sample_rate=1.0, + trace_ignore_status_codes=set( + range(401, 404), + ), + ) + events = capture_events() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", "404") + + assert len(events) == 1 + + +def test_records_lost_events(sentry_init, capture_record_lost_event_calls): + sentry_init( + traces_sample_rate=1.0, + trace_ignore_status_codes={ + 404, + }, + ) + record_lost_event_calls = capture_record_lost_event_calls() + + with start_transaction(op="http", name="GET /"): + span_or_tx = sentry_sdk.get_current_span() + span_or_tx.set_data("http.response.status_code", 404) + + with start_span(op="child-span"): + with start_span(op="child-child-span"): + pass + + assert Counter(record_lost_event_calls) == Counter( + [ + ("event_processor", "transaction", None, 1), + ("event_processor", "span", None, 3), + ] + ) From 2e4457c8da5112e095362899ef8464bc304cc54a Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Thu, 2 Oct 2025 16:35:31 +0200 Subject: [PATCH 515/591] fix(openai-agents): also emit spans for MCP tool calls done by the LLM (#4875) ### Description We cannot directly intercept MCP Tool calls, as they are done remotely by the LLM and not in the Agent itself. However, we see when such a tool call took place, so we can emit a zero-length span with the tool call specifics. It will start at the same time as the parent span. Closes https://linear.app/getsentry/issue/TET-1192/openai-agents-hosted-mcp-calls-cannot-be-wrapped-in-an-execute-tool --- > [!NOTE] > Emit execute_tool spans for MCP tool calls detected in agent results, with tool metadata, input/output (PII-gated), and error status. > > - **Tracing/Spans (openai_agents)**: > - Add `utils._create_mcp_execute_tool_spans` to emit `OP.GEN_AI_EXECUTE_TOOL` spans for MCP tool calls (`McpCall`) found in `result.output`. > - Sets `GEN_AI_TOOL_TYPE=mcp`, `GEN_AI_TOOL_NAME`, propagates input/output when PII allowed, and marks `SPANSTATUS.ERROR` on error. > - Spans start at the parent span's start time (zero-length representation of remote call). > - Wire into `spans/ai_client.update_ai_client_span` to create these tool spans after setting usage/input/output data. > - Update imports to include `SPANSTATUS` and `OP`. > > Written by [Cursor Bugbot](https://cursor.com/dashboard?tab=bugbot) for commit 96df8c1c08768d7f1d1369e66a9c4e7f6ebfc04c. This will update automatically on new commits. Configure [here](https://cursor.com/dashboard?tab=bugbot). --------- Co-authored-by: Ivana Kellyer --- .../openai_agents/spans/ai_client.py | 2 + .../integrations/openai_agents/utils.py | 26 +- .../openai_agents/test_openai_agents.py | 302 ++++++++++++++++++ 3 files changed, 329 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/openai_agents/spans/ai_client.py b/sentry_sdk/integrations/openai_agents/spans/ai_client.py index e215edfd26..88b403ba85 100644 --- a/sentry_sdk/integrations/openai_agents/spans/ai_client.py +++ b/sentry_sdk/integrations/openai_agents/spans/ai_client.py @@ -7,6 +7,7 @@ _set_input_data, _set_output_data, _set_usage_data, + _create_mcp_execute_tool_spans, ) from typing import TYPE_CHECKING @@ -38,3 +39,4 @@ def update_ai_client_span(span, agent, get_response_kwargs, result): _set_usage_data(span, result.usage) _set_input_data(span, get_response_kwargs) _set_output_data(span, result) + _create_mcp_execute_tool_spans(span, result) diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index 73d2858e7f..b0ad6bf903 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -1,6 +1,6 @@ import sentry_sdk from sentry_sdk.ai.utils import set_data_normalized -from sentry_sdk.consts import SPANDATA +from sentry_sdk.consts import SPANDATA, SPANSTATUS, OP from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii from sentry_sdk.tracing_utils import set_span_errored @@ -156,3 +156,27 @@ def _set_output_data(span, result): set_data_normalized( span, SPANDATA.GEN_AI_RESPONSE_TEXT, output_messages["response"] ) + + +def _create_mcp_execute_tool_spans(span, result): + # type: (sentry_sdk.tracing.Span, agents.Result) -> None + for output in result.output: + if output.__class__.__name__ == "McpCall": + with sentry_sdk.start_span( + op=OP.GEN_AI_EXECUTE_TOOL, + description=f"execute_tool {output.name}", + start_timestamp=span.start_timestamp, + ) as execute_tool_span: + set_data_normalized(execute_tool_span, SPANDATA.GEN_AI_TOOL_TYPE, "mcp") + set_data_normalized( + execute_tool_span, SPANDATA.GEN_AI_TOOL_NAME, output.name + ) + if should_send_default_pii(): + execute_tool_span.set_data( + SPANDATA.GEN_AI_TOOL_INPUT, output.arguments + ) + execute_tool_span.set_data( + SPANDATA.GEN_AI_TOOL_OUTPUT, output.output + ) + if output.error: + execute_tool_span.set_status(SPANSTATUS.ERROR) diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index bd7f15faff..1768971c99 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -15,6 +15,7 @@ ModelSettings, ) from agents.items import ( + McpCall, ResponseOutputMessage, ResponseOutputText, ResponseFunctionToolCall, @@ -683,6 +684,307 @@ async def test_span_status_error(sentry_init, capture_events, test_agent): assert transaction["contexts"]["trace"]["status"] == "error" +@pytest.mark.asyncio +async def test_mcp_tool_execution_spans(sentry_init, capture_events, test_agent): + """ + Test that MCP (Model Context Protocol) tool calls create execute_tool spans. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Create a McpCall object + mcp_call = McpCall( + id="mcp_call_123", + name="test_mcp_tool", + arguments='{"query": "search term"}', + output="MCP tool executed successfully", + error=None, + type="mcp_call", + server_label="test_server", + ) + + # Create a ModelResponse with an McpCall in the output + mcp_response = ModelResponse( + output=[mcp_call], + usage=Usage( + requests=1, + input_tokens=10, + output_tokens=5, + total_tokens=15, + ), + response_id="resp_mcp_123", + ) + + # Final response after MCP tool execution + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="Task completed using MCP tool", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=Usage( + requests=1, + input_tokens=15, + output_tokens=10, + total_tokens=25, + ), + response_id="resp_final_123", + ) + + mock_get_response.side_effect = [mcp_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await agents.Runner.run( + test_agent, + "Please use MCP tool", + run_config=test_run_config, + ) + + (transaction,) = events + spans = transaction["spans"] + + # Find the MCP execute_tool span + mcp_tool_span = None + for span in spans: + if ( + span.get("description") == "execute_tool test_mcp_tool" + and span.get("data", {}).get("gen_ai.tool.type") == "mcp" + ): + mcp_tool_span = span + break + + # Verify the MCP tool span was created + assert mcp_tool_span is not None, "MCP execute_tool span was not created" + assert mcp_tool_span["description"] == "execute_tool test_mcp_tool" + assert mcp_tool_span["data"]["gen_ai.tool.type"] == "mcp" + assert mcp_tool_span["data"]["gen_ai.tool.name"] == "test_mcp_tool" + assert mcp_tool_span["data"]["gen_ai.tool.input"] == '{"query": "search term"}' + assert ( + mcp_tool_span["data"]["gen_ai.tool.output"] == "MCP tool executed successfully" + ) + + # Verify no error status since error was None + assert mcp_tool_span.get("tags", {}).get("status") != "error" + + +@pytest.mark.asyncio +async def test_mcp_tool_execution_with_error(sentry_init, capture_events, test_agent): + """ + Test that MCP tool calls with errors are tracked with error status. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Create a McpCall object with an error + mcp_call_with_error = McpCall( + id="mcp_call_error_123", + name="failing_mcp_tool", + arguments='{"query": "test"}', + output=None, + error="MCP tool execution failed", + type="mcp_call", + server_label="test_server", + ) + + # Create a ModelResponse with a failing McpCall + mcp_response = ModelResponse( + output=[mcp_call_with_error], + usage=Usage( + requests=1, + input_tokens=10, + output_tokens=5, + total_tokens=15, + ), + response_id="resp_mcp_error_123", + ) + + # Final response after error + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="The MCP tool encountered an error", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=Usage( + requests=1, + input_tokens=15, + output_tokens=10, + total_tokens=25, + ), + response_id="resp_final_error_123", + ) + + mock_get_response.side_effect = [mcp_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await agents.Runner.run( + test_agent, + "Please use failing MCP tool", + run_config=test_run_config, + ) + + (transaction,) = events + spans = transaction["spans"] + + # Find the MCP execute_tool span with error + mcp_tool_span = None + for span in spans: + if ( + span.get("description") == "execute_tool failing_mcp_tool" + and span.get("data", {}).get("gen_ai.tool.type") == "mcp" + ): + mcp_tool_span = span + break + + # Verify the MCP tool span was created with error status + assert mcp_tool_span is not None, "MCP execute_tool span was not created" + assert mcp_tool_span["description"] == "execute_tool failing_mcp_tool" + assert mcp_tool_span["data"]["gen_ai.tool.type"] == "mcp" + assert mcp_tool_span["data"]["gen_ai.tool.name"] == "failing_mcp_tool" + assert mcp_tool_span["data"]["gen_ai.tool.input"] == '{"query": "test"}' + assert mcp_tool_span["data"]["gen_ai.tool.output"] is None + + # Verify error status was set + assert mcp_tool_span["tags"]["status"] == "error" + + +@pytest.mark.asyncio +async def test_mcp_tool_execution_without_pii(sentry_init, capture_events, test_agent): + """ + Test that MCP tool input/output are not included when send_default_pii is False. + """ + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Create a McpCall object + mcp_call = McpCall( + id="mcp_call_pii_123", + name="test_mcp_tool", + arguments='{"query": "sensitive data"}', + output="Result with sensitive info", + error=None, + type="mcp_call", + server_label="test_server", + ) + + # Create a ModelResponse with an McpCall + mcp_response = ModelResponse( + output=[mcp_call], + usage=Usage( + requests=1, + input_tokens=10, + output_tokens=5, + total_tokens=15, + ), + response_id="resp_mcp_123", + ) + + # Final response + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="Task completed", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=Usage( + requests=1, + input_tokens=15, + output_tokens=10, + total_tokens=25, + ), + response_id="resp_final_123", + ) + + mock_get_response.side_effect = [mcp_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=False, # PII disabled + ) + + events = capture_events() + + await agents.Runner.run( + test_agent, + "Please use MCP tool", + run_config=test_run_config, + ) + + (transaction,) = events + spans = transaction["spans"] + + # Find the MCP execute_tool span + mcp_tool_span = None + for span in spans: + if ( + span.get("description") == "execute_tool test_mcp_tool" + and span.get("data", {}).get("gen_ai.tool.type") == "mcp" + ): + mcp_tool_span = span + break + + # Verify the MCP tool span was created but without input/output + assert mcp_tool_span is not None, "MCP execute_tool span was not created" + assert mcp_tool_span["description"] == "execute_tool test_mcp_tool" + assert mcp_tool_span["data"]["gen_ai.tool.type"] == "mcp" + assert mcp_tool_span["data"]["gen_ai.tool.name"] == "test_mcp_tool" + + # Verify input and output are not included when send_default_pii is False + assert "gen_ai.tool.input" not in mcp_tool_span["data"] + assert "gen_ai.tool.output" not in mcp_tool_span["data"] + + @pytest.mark.asyncio async def test_multiple_agents_asyncio( sentry_init, capture_events, test_agent, mock_model_response From ffc88f5e21688d15ea486e2af4a7b3fe5f2ea0d4 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 2 Oct 2025 16:02:22 +0000 Subject: [PATCH 516/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/02)=20(#4880)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> --- scripts/populate_tox/releases.jsonl | 29 ++++--- .../openai_agents/test_openai_agents.py | 42 +++++----- tox.ini | 76 +++++++++---------- 3 files changed, 73 insertions(+), 74 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 3532b61c75..afbb5aef09 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -5,8 +5,8 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.5", "version": "2.2.28", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.1.14", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.2.25", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.24", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.6", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.25", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0a1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Flask", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "1.1.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}} @@ -24,9 +24,9 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.33.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.50.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.68.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.34.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.52.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.69.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.40", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.43", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -93,13 +93,13 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.7", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.0", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.8", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a4", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} @@ -108,12 +108,11 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.37.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.73.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.3", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} @@ -126,7 +125,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.5.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.6.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.2", "yanked": false}} {"info": {"classifiers": ["Framework :: Pylons", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": null, "version": "1.0.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.10.8", "yanked": false}} {"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.6.5", "yanked": false}} @@ -187,7 +186,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.64.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.282.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index 1768971c99..e9a8372806 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -6,6 +6,7 @@ from sentry_sdk.integrations.openai_agents import OpenAIAgentsIntegration from sentry_sdk.integrations.openai_agents.utils import safe_serialize +from sentry_sdk.utils import parse_version import agents from agents import ( @@ -20,6 +21,7 @@ ResponseOutputText, ResponseFunctionToolCall, ) +from agents.version import __version__ as OPENAI_AGENTS_VERSION from openai.types.responses.response_usage import ( InputTokensDetails, @@ -438,24 +440,28 @@ def simple_test_tool(message: str) -> str: ai_client_span2, ) = spans - available_tools = safe_serialize( - [ - { - "name": "simple_test_tool", - "description": "A simple tool", - "params_json_schema": { - "properties": {"message": {"title": "Message", "type": "string"}}, - "required": ["message"], - "title": "simple_test_tool_args", - "type": "object", - "additionalProperties": False, - }, - "on_invoke_tool": "._create_function_tool.._on_invoke_tool>", - "strict_json_schema": True, - "is_enabled": True, - } - ] - ) + available_tools = [ + { + "name": "simple_test_tool", + "description": "A simple tool", + "params_json_schema": { + "properties": {"message": {"title": "Message", "type": "string"}}, + "required": ["message"], + "title": "simple_test_tool_args", + "type": "object", + "additionalProperties": False, + }, + "on_invoke_tool": "._create_function_tool.._on_invoke_tool>", + "strict_json_schema": True, + "is_enabled": True, + } + ] + if parse_version(OPENAI_AGENTS_VERSION) >= (0, 3, 3): + available_tools[0].update( + {"tool_input_guardrails": None, "tool_output_guardrails": None} + ) + + available_tools = safe_serialize(available_tools) assert transaction["transaction"] == "test_agent workflow" assert transaction["contexts"]["trace"]["origin"] == "auto.ai.openai_agents" diff --git a/tox.ini b/tox.ini index 2354c66c7c..1bca280a11 100644 --- a/tox.ini +++ b/tox.ini @@ -46,9 +46,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.33.1 - {py3.8,py3.11,py3.12}-anthropic-v0.50.0 - {py3.8,py3.12,py3.13}-anthropic-v0.68.1 + {py3.8,py3.11,py3.12}-anthropic-v0.34.2 + {py3.8,py3.11,py3.12}-anthropic-v0.52.2 + {py3.8,py3.12,py3.13}-anthropic-v0.69.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.9.4 @@ -64,34 +64,32 @@ envlist = {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 - {py3.8,py3.11,py3.12}-openai-base-v1.37.2 - {py3.8,py3.11,py3.12}-openai-base-v1.73.0 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 + {py3.8,py3.12,py3.13}-openai-base-v2.0.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.37.2 - {py3.8,py3.11,py3.12}-openai-notiktoken-v1.73.0 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 + {py3.8,py3.12,py3.13}-openai-notiktoken-v2.0.1 - {py3.9,py3.12,py3.13}-langgraph-v0.6.7 - {py3.10,py3.12,py3.13}-langgraph-v1.0.0a3 + {py3.9,py3.12,py3.13}-langgraph-v0.6.8 + {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.3.2 + {py3.10,py3.12,py3.13}-openai_agents-v0.3.3 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.1 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.40 + {py3.9,py3.12,py3.13}-boto3-v1.40.43 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -107,7 +105,7 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 - {py3.9,py3.12,py3.13}-pymongo-v4.15.1 + {py3.9,py3.12,py3.13}-pymongo-v4.15.2 {py3.6}-redis-v2.10.6 {py3.6,py3.7,py3.8}-redis-v3.5.3 @@ -126,13 +124,13 @@ envlist = # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 - {py3.9,py3.12,py3.13}-launchdarkly-v9.12.0 + {py3.9,py3.12,py3.13}-launchdarkly-v9.12.1 {py3.8,py3.12,py3.13}-openfeature-v0.7.5 {py3.9,py3.12,py3.13}-openfeature-v0.8.3 {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.64.0 + {py3.7,py3.12,py3.13}-statsig-v0.65.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.3.0 @@ -202,8 +200,8 @@ envlist = {py3.6,py3.7}-django-v1.11.29 {py3.6,py3.8,py3.9}-django-v2.2.28 {py3.6,py3.9,py3.10}-django-v3.2.25 - {py3.8,py3.11,py3.12}-django-v4.2.24 - {py3.10,py3.12,py3.13}-django-v5.2.6 + {py3.8,py3.11,py3.12}-django-v4.2.25 + {py3.10,py3.12,py3.13}-django-v5.2.7 {py3.12,py3.13}-django-v6.0a1 {py3.6,py3.7,py3.8}-flask-v1.1.4 @@ -332,12 +330,12 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.33.1: anthropic==0.33.1 - anthropic-v0.50.0: anthropic==0.50.0 - anthropic-v0.68.1: anthropic==0.68.1 + anthropic-v0.34.2: anthropic==0.34.2 + anthropic-v0.52.2: anthropic==0.52.2 + anthropic-v0.69.0: anthropic==0.69.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.33.1: httpx<0.28.0 + anthropic-v0.34.2: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.9.4: cohere==5.9.4 @@ -360,35 +358,31 @@ deps = langchain-notiktoken-v0.3.27: langchain-community openai-base-v1.0.1: openai==1.0.1 - openai-base-v1.37.2: openai==1.37.2 - openai-base-v1.73.0: openai==1.73.0 openai-base-v1.109.1: openai==1.109.1 + openai-base-v2.0.1: openai==2.0.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 - openai-base-v1.37.2: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 - openai-notiktoken-v1.37.2: openai==1.37.2 - openai-notiktoken-v1.73.0: openai==1.73.0 openai-notiktoken-v1.109.1: openai==1.109.1 + openai-notiktoken-v2.0.1: openai==2.0.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 - openai-notiktoken-v1.37.2: httpx<0.28 - langgraph-v0.6.7: langgraph==0.6.7 - langgraph-v1.0.0a3: langgraph==1.0.0a3 + langgraph-v0.6.8: langgraph==0.6.8 + langgraph-v1.0.0a4: langgraph==1.0.0a4 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.3.2: openai-agents==0.3.2 + openai_agents-v0.3.3: openai-agents==0.3.3 openai_agents: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 - huggingface_hub-v0.35.1: huggingface_hub==0.35.1 + huggingface_hub-v0.35.3: huggingface_hub==0.35.3 huggingface_hub: responses @@ -396,7 +390,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.40: boto3==1.40.40 + boto3-v1.40.43: boto3==1.40.43 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -415,7 +409,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 - pymongo-v4.15.1: pymongo==4.15.1 + pymongo-v4.15.2: pymongo==4.15.2 pymongo: mockupdb redis-v2.10.6: redis==2.10.6 @@ -440,13 +434,13 @@ deps = # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 - launchdarkly-v9.12.0: launchdarkly-server-sdk==9.12.0 + launchdarkly-v9.12.1: launchdarkly-server-sdk==9.12.1 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.3: openfeature-sdk==0.8.3 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.64.0: statsig==0.64.0 + statsig-v0.65.0: statsig==0.65.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -548,8 +542,8 @@ deps = django-v1.11.29: django==1.11.29 django-v2.2.28: django==2.2.28 django-v3.2.25: django==3.2.25 - django-v4.2.24: django==4.2.24 - django-v5.2.6: django==5.2.6 + django-v4.2.25: django==4.2.25 + django-v5.2.7: django==5.2.7 django-v6.0a1: django==6.0a1 django: psycopg2-binary django: djangorestframework @@ -557,13 +551,13 @@ deps = django: Werkzeug django-v2.2.28: channels[daphne] django-v3.2.25: channels[daphne] - django-v4.2.24: channels[daphne] - django-v5.2.6: channels[daphne] + django-v4.2.25: channels[daphne] + django-v5.2.7: channels[daphne] django-v6.0a1: channels[daphne] django-v2.2.28: six django-v3.2.25: pytest-asyncio - django-v4.2.24: pytest-asyncio - django-v5.2.6: pytest-asyncio + django-v4.2.25: pytest-asyncio + django-v5.2.7: pytest-asyncio django-v6.0a1: pytest-asyncio django-v1.11.29: djangorestframework>=3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 From f3e8a5ccd0a341cf139f474856163f0e5335741c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 3 Oct 2025 08:40:35 +0200 Subject: [PATCH 517/591] fix(tests): Don't assume release is set (#4879) ### Description Even though we try to figure out the current release automatically if it's not provided, it can still end up being `None`. If that's the case, it won't be attached to logs. The `test_logs_attributes` test assumes there always is a release, which is incorrect. I opted for conditionally checking for `sentry.release` in the test instead of removing the check altogether, even though the test itself is supposed to test custom user provided attributes. The reason is that there is no other generic logs test testing `sentry.release`. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4878 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- tests/test_logs.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/tests/test_logs.py b/tests/test_logs.py index 596a31922e..1e252c5bfb 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -230,7 +230,8 @@ def test_logs_attributes(sentry_init, capture_envelopes): for k, v in attrs.items(): assert logs[0]["attributes"][k] == v assert logs[0]["attributes"]["sentry.environment"] == "production" - assert "sentry.release" in logs[0]["attributes"] + if sentry_sdk.get_client().options.get("release") is not None: + assert "sentry.release" in logs[0]["attributes"] assert logs[0]["attributes"]["sentry.message.parameter.my_var"] == "some value" assert logs[0]["attributes"][SPANDATA.SERVER_ADDRESS] == "test-server" assert logs[0]["attributes"]["sentry.sdk.name"].startswith("sentry.python") From bbd2a5d8cd3725dd0d4e0840c2e126038404d1a3 Mon Sep 17 00:00:00 2001 From: Vyskorko Igor Date: Fri, 3 Oct 2025 16:00:08 +0700 Subject: [PATCH 518/591] feat(integrations): Add tracing to DramatiqIntegration (#4571) Adds tracing support to DramatiqIntegration #3454 --------- Co-authored-by: igorek Co-authored-by: Anton Pirker Co-authored-by: Ivana Kellyer --- pyproject.toml | 4 + sentry_sdk/consts.py | 1 + sentry_sdk/integrations/dramatiq.py | 120 ++++++++++++++----- tests/integrations/dramatiq/test_dramatiq.py | 57 ++++++++- 4 files changed, 146 insertions(+), 36 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 8e6fe345f4..5b86531014 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -179,6 +179,10 @@ ignore_missing_imports = true module = "agents.*" ignore_missing_imports = true +[[tool.mypy.overrides]] +module = "dramatiq.*" +ignore_missing_imports = true + # # Tool: Ruff (linting and formatting) # diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 5bcc487037..7b9e61a065 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -839,6 +839,7 @@ class OP: QUEUE_TASK_HUEY = "queue.task.huey" QUEUE_SUBMIT_RAY = "queue.submit.ray" QUEUE_TASK_RAY = "queue.task.ray" + QUEUE_TASK_DRAMATIQ = "queue.task.dramatiq" SUBPROCESS = "subprocess" SUBPROCESS_WAIT = "subprocess.wait" SUBPROCESS_COMMUNICATE = "subprocess.communicate" diff --git a/sentry_sdk/integrations/dramatiq.py b/sentry_sdk/integrations/dramatiq.py index a756b4c669..8b85831cf4 100644 --- a/sentry_sdk/integrations/dramatiq.py +++ b/sentry_sdk/integrations/dramatiq.py @@ -1,18 +1,31 @@ import json import sentry_sdk -from sentry_sdk.integrations import Integration +from sentry_sdk.consts import OP, SPANSTATUS +from sentry_sdk.api import continue_trace, get_baggage, get_traceparent +from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.integrations._wsgi_common import request_body_within_bounds +from sentry_sdk.tracing import ( + BAGGAGE_HEADER_NAME, + SENTRY_TRACE_HEADER_NAME, + TransactionSource, +) from sentry_sdk.utils import ( AnnotatedValue, capture_internal_exceptions, event_from_exception, ) +from typing import TypeVar + +R = TypeVar("R") -from dramatiq.broker import Broker # type: ignore -from dramatiq.message import Message # type: ignore -from dramatiq.middleware import Middleware, default_middleware # type: ignore -from dramatiq.errors import Retry # type: ignore +try: + from dramatiq.broker import Broker + from dramatiq.middleware import Middleware, default_middleware + from dramatiq.errors import Retry + from dramatiq.message import Message +except ImportError: + raise DidNotEnable("Dramatiq is not installed") from typing import TYPE_CHECKING @@ -34,10 +47,12 @@ class DramatiqIntegration(Integration): """ identifier = "dramatiq" + origin = f"auto.queue.{identifier}" @staticmethod def setup_once(): # type: () -> None + _patch_dramatiq_broker() @@ -85,22 +100,54 @@ class SentryMiddleware(Middleware): # type: ignore[misc] DramatiqIntegration. """ - def before_process_message(self, broker, message): - # type: (Broker, Message) -> None + SENTRY_HEADERS_NAME = "_sentry_headers" + + def before_enqueue(self, broker, message, delay): + # type: (Broker, Message[R], int) -> None integration = sentry_sdk.get_client().get_integration(DramatiqIntegration) if integration is None: return - message._scope_manager = sentry_sdk.new_scope() - message._scope_manager.__enter__() + message.options[self.SENTRY_HEADERS_NAME] = { + BAGGAGE_HEADER_NAME: get_baggage(), + SENTRY_TRACE_HEADER_NAME: get_traceparent(), + } + + def before_process_message(self, broker, message): + # type: (Broker, Message[R]) -> None + integration = sentry_sdk.get_client().get_integration(DramatiqIntegration) + if integration is None: + return - scope = sentry_sdk.get_current_scope() - scope.set_transaction_name(message.actor_name) + message._scope_manager = sentry_sdk.isolation_scope() + scope = message._scope_manager.__enter__() + scope.clear_breadcrumbs() scope.set_extra("dramatiq_message_id", message.message_id) scope.add_event_processor(_make_message_event_processor(message, integration)) + sentry_headers = message.options.get(self.SENTRY_HEADERS_NAME) or {} + if "retries" in message.options: + # start new trace in case of retrying + sentry_headers = {} + + transaction = continue_trace( + sentry_headers, + name=message.actor_name, + op=OP.QUEUE_TASK_DRAMATIQ, + source=TransactionSource.TASK, + origin=DramatiqIntegration.origin, + ) + transaction.set_status(SPANSTATUS.OK) + sentry_sdk.start_transaction( + transaction, + name=message.actor_name, + op=OP.QUEUE_TASK_DRAMATIQ, + source=TransactionSource.TASK, + ) + transaction.__enter__() + def after_process_message(self, broker, message, *, result=None, exception=None): - # type: (Broker, Message, Any, Optional[Any], Optional[Exception]) -> None + # type: (Broker, Message[R], Optional[Any], Optional[Exception]) -> None integration = sentry_sdk.get_client().get_integration(DramatiqIntegration) if integration is None: return @@ -108,27 +155,38 @@ def after_process_message(self, broker, message, *, result=None, exception=None) actor = broker.get_actor(message.actor_name) throws = message.options.get("throws") or actor.options.get("throws") - try: - if ( - exception is not None - and not (throws and isinstance(exception, throws)) - and not isinstance(exception, Retry) - ): - event, hint = event_from_exception( - exception, - client_options=sentry_sdk.get_client().options, - mechanism={ - "type": DramatiqIntegration.identifier, - "handled": False, - }, - ) - sentry_sdk.capture_event(event, hint=hint) - finally: - message._scope_manager.__exit__(None, None, None) + scope_manager = message._scope_manager + transaction = sentry_sdk.get_current_scope().transaction + if not transaction: + return None + + is_event_capture_required = ( + exception is not None + and not (throws and isinstance(exception, throws)) + and not isinstance(exception, Retry) + ) + if not is_event_capture_required: + # normal transaction finish + transaction.__exit__(None, None, None) + scope_manager.__exit__(None, None, None) + return + + event, hint = event_from_exception( + exception, # type: ignore[arg-type] + client_options=sentry_sdk.get_client().options, + mechanism={ + "type": DramatiqIntegration.identifier, + "handled": False, + }, + ) + sentry_sdk.capture_event(event, hint=hint) + # transaction error + transaction.__exit__(type(exception), exception, None) + scope_manager.__exit__(type(exception), exception, None) def _make_message_event_processor(message, integration): - # type: (Message, DramatiqIntegration) -> Callable[[Event, Hint], Optional[Event]] + # type: (Message[R], DramatiqIntegration) -> Callable[[Event, Hint], Optional[Event]] def inner(event, hint): # type: (Event, Hint) -> Optional[Event] @@ -142,7 +200,7 @@ def inner(event, hint): class DramatiqMessageExtractor: def __init__(self, message): - # type: (Message) -> None + # type: (Message[R]) -> None self.message_data = dict(message.asdict()) def content_length(self): diff --git a/tests/integrations/dramatiq/test_dramatiq.py b/tests/integrations/dramatiq/test_dramatiq.py index d7917cbd00..53c36b640c 100644 --- a/tests/integrations/dramatiq/test_dramatiq.py +++ b/tests/integrations/dramatiq/test_dramatiq.py @@ -5,12 +5,21 @@ from dramatiq.brokers.stub import StubBroker import sentry_sdk +from sentry_sdk.tracing import TransactionSource +from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANSTATUS from sentry_sdk.integrations.dramatiq import DramatiqIntegration +from sentry_sdk.integrations.logging import ignore_logger +ignore_logger("dramatiq.worker.WorkerThread") -@pytest.fixture -def broker(sentry_init): - sentry_init(integrations=[DramatiqIntegration()]) + +@pytest.fixture(scope="function") +def broker(request, sentry_init): + sentry_init( + integrations=[DramatiqIntegration()], + traces_sample_rate=getattr(request, "param", None), + ) broker = StubBroker() broker.emit_after("process_boot") dramatiq.set_broker(broker) @@ -44,19 +53,57 @@ def dummy_actor(x, y): assert exception["type"] == "ZeroDivisionError" -def test_that_actor_name_is_set_as_transaction(broker, worker, capture_events): +@pytest.mark.parametrize( + "broker,expected_span_status", + [ + (1.0, SPANSTATUS.INTERNAL_ERROR), + (1.0, SPANSTATUS.OK), + ], + ids=["error", "success"], + indirect=["broker"], +) +def test_task_transaction(broker, worker, capture_events, expected_span_status): events = capture_events() + task_fails = expected_span_status == SPANSTATUS.INTERNAL_ERROR @dramatiq.actor(max_retries=0) def dummy_actor(x, y): return x / y - dummy_actor.send(1, 0) + dummy_actor.send(1, int(not task_fails)) broker.join(dummy_actor.queue_name) worker.join() + if task_fails: + error_event = events.pop(0) + exception = error_event["exception"]["values"][0] + assert exception["type"] == "ZeroDivisionError" + assert exception["mechanism"]["type"] == DramatiqIntegration.identifier + (event,) = events + assert event["type"] == "transaction" assert event["transaction"] == "dummy_actor" + assert event["transaction_info"] == {"source": TransactionSource.TASK} + assert event["contexts"]["trace"]["status"] == expected_span_status + + +@pytest.mark.parametrize("broker", [1.0], indirect=True) +def test_dramatiq_propagate_trace(broker, worker, capture_events): + events = capture_events() + + @dramatiq.actor(max_retries=0) + def propagated_trace_task(): + pass + + with start_transaction() as outer_transaction: + propagated_trace_task.send() + broker.join(propagated_trace_task.queue_name) + worker.join() + + assert ( + events[0]["transaction"] == "propagated_trace_task" + ) # the "inner" transaction + assert events[0]["contexts"]["trace"]["trace_id"] == outer_transaction.trace_id def test_that_dramatiq_message_id_is_set_as_extra(broker, worker, capture_events): From f979abfc25b935f00b3c18d99d18888d833d0f78 Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Fri, 3 Oct 2025 11:47:57 +0200 Subject: [PATCH 519/591] feat(integrations): add litellm integration (#4864) Add a first implementation of the litellm integration, supporting completion and embeddings Closes https://linear.app/getsentry/issue/PY-1828/add-agent-monitoring-support-for-litellm Closes https://linear.app/getsentry/issue/TET-1218/litellm-testing --- > [!NOTE] > Introduce `LiteLLMIntegration` that instruments LiteLLM chat/embeddings calls with spans, token usage, optional prompt logging, and exception capture. > > - **Integrations**: > - Add `sentry_sdk/integrations/litellm.py` with `LiteLLMIntegration` registering LiteLLM `input/success/failure` callbacks. > - Start spans for `chat`/`embeddings`, set `gen_ai.*` metadata (provider/system, operation, model, params like `max_tokens`, `temperature`, `top_p`, `stream`). > - Record LiteLLM-specific fields: `api_base`, `api_version`, `custom_llm_provider`. > - Optionally capture request/response messages when `include_prompts` and PII are enabled. > - Track token usage from response `usage` and capture exceptions; always finish spans. > > Written by [Cursor Bugbot](https://cursor.com/dashboard?tab=bugbot) for commit 1ecd559392f9526c884e3cec12a152c65f098965. This will update automatically on new commits. Configure [here](https://cursor.com/dashboard?tab=bugbot). --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 + scripts/populate_tox/config.py | 3 + scripts/populate_tox/releases.jsonl | 7 +- .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/__init__.py | 1 + sentry_sdk/integrations/litellm.py | 251 ++++++++ setup.py | 1 + tests/integrations/litellm/__init__.py | 0 tests/integrations/litellm/test_litellm.py | 547 ++++++++++++++++++ tox.ini | 21 +- 10 files changed, 825 insertions(+), 11 deletions(-) create mode 100644 sentry_sdk/integrations/litellm.py create mode 100644 tests/integrations/litellm/__init__.py create mode 100644 tests/integrations/litellm/test_litellm.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index cf21720ff1..fcbb464078 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -66,6 +66,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-notiktoken" + - name: Test litellm + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-litellm" - name: Test openai-base run: | set -x # print commands that are executed diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 34ae680fad..f69e5f2f90 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -214,6 +214,9 @@ "package": "launchdarkly-server-sdk", "num_versions": 2, }, + "litellm": { + "package": "litellm", + }, "litestar": { "package": "litestar", "deps": { diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index afbb5aef09..b7cca55815 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.43", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.44", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -101,6 +101,7 @@ {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.17.0", "yanked": false}} @@ -108,7 +109,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} @@ -200,6 +201,6 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.7", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.8", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.19.2", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index a7b7c394b1..81f887ad4f 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -74,6 +74,7 @@ "cohere", "langchain-base", "langchain-notiktoken", + "litellm", "openai-base", "openai-notiktoken", "langgraph", diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index e397c9986a..3f71f0f4ba 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -146,6 +146,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "langchain": (0, 1, 0), "langgraph": (0, 6, 6), "launchdarkly": (9, 8, 0), + "litellm": (1, 77, 5), "loguru": (0, 7, 0), "openai": (1, 0, 0), "openai_agents": (0, 0, 19), diff --git a/sentry_sdk/integrations/litellm.py b/sentry_sdk/integrations/litellm.py new file mode 100644 index 0000000000..2582c2bc05 --- /dev/null +++ b/sentry_sdk/integrations/litellm.py @@ -0,0 +1,251 @@ +from typing import TYPE_CHECKING + +import sentry_sdk +from sentry_sdk import consts +from sentry_sdk.ai.monitoring import record_token_usage +from sentry_sdk.ai.utils import get_start_span_function, set_data_normalized +from sentry_sdk.consts import SPANDATA +from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import event_from_exception + +if TYPE_CHECKING: + from typing import Any, Dict + from datetime import datetime + +try: + import litellm # type: ignore[import-not-found] +except ImportError: + raise DidNotEnable("LiteLLM not installed") + + +def _get_metadata_dict(kwargs): + # type: (Dict[str, Any]) -> Dict[str, Any] + """Get the metadata dictionary from the kwargs.""" + litellm_params = kwargs.setdefault("litellm_params", {}) + + # we need this weird little dance, as metadata might be set but may be None initially + metadata = litellm_params.get("metadata") + if metadata is None: + metadata = {} + litellm_params["metadata"] = metadata + return metadata + + +def _input_callback(kwargs): + # type: (Dict[str, Any]) -> None + """Handle the start of a request.""" + integration = sentry_sdk.get_client().get_integration(LiteLLMIntegration) + + if integration is None: + return + + # Get key parameters + full_model = kwargs.get("model", "") + try: + model, provider, _, _ = litellm.get_llm_provider(full_model) + except Exception: + model = full_model + provider = "unknown" + + messages = kwargs.get("messages", []) + operation = "chat" if messages else "embeddings" + + # Start a new span/transaction + span = get_start_span_function()( + op=( + consts.OP.GEN_AI_CHAT + if operation == "chat" + else consts.OP.GEN_AI_EMBEDDINGS + ), + name=f"{operation} {model}", + origin=LiteLLMIntegration.origin, + ) + span.__enter__() + + # Store span for later + _get_metadata_dict(kwargs)["_sentry_span"] = span + + # Set basic data + set_data_normalized(span, SPANDATA.GEN_AI_SYSTEM, provider) + set_data_normalized(span, SPANDATA.GEN_AI_OPERATION_NAME, operation) + + # Record messages if allowed + if messages and should_send_default_pii() and integration.include_prompts: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + ) + + # Record other parameters + params = { + "model": SPANDATA.GEN_AI_REQUEST_MODEL, + "stream": SPANDATA.GEN_AI_RESPONSE_STREAMING, + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, + } + for key, attribute in params.items(): + value = kwargs.get(key) + if value is not None: + set_data_normalized(span, attribute, value) + + # Record LiteLLM-specific parameters + litellm_params = { + "api_base": kwargs.get("api_base"), + "api_version": kwargs.get("api_version"), + "custom_llm_provider": kwargs.get("custom_llm_provider"), + } + for key, value in litellm_params.items(): + if value is not None: + set_data_normalized(span, f"gen_ai.litellm.{key}", value) + + +def _success_callback(kwargs, completion_response, start_time, end_time): + # type: (Dict[str, Any], Any, datetime, datetime) -> None + """Handle successful completion.""" + + span = _get_metadata_dict(kwargs).get("_sentry_span") + if span is None: + return + + integration = sentry_sdk.get_client().get_integration(LiteLLMIntegration) + if integration is None: + return + + try: + # Record model information + if hasattr(completion_response, "model"): + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_MODEL, completion_response.model + ) + + # Record response content if allowed + if should_send_default_pii() and integration.include_prompts: + if hasattr(completion_response, "choices"): + response_messages = [] + for choice in completion_response.choices: + if hasattr(choice, "message"): + if hasattr(choice.message, "model_dump"): + response_messages.append(choice.message.model_dump()) + elif hasattr(choice.message, "dict"): + response_messages.append(choice.message.dict()) + else: + # Fallback for basic message objects + msg = {} + if hasattr(choice.message, "role"): + msg["role"] = choice.message.role + if hasattr(choice.message, "content"): + msg["content"] = choice.message.content + if hasattr(choice.message, "tool_calls"): + msg["tool_calls"] = choice.message.tool_calls + response_messages.append(msg) + + if response_messages: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, response_messages + ) + + # Record token usage + if hasattr(completion_response, "usage"): + usage = completion_response.usage + record_token_usage( + span, + input_tokens=getattr(usage, "prompt_tokens", None), + output_tokens=getattr(usage, "completion_tokens", None), + total_tokens=getattr(usage, "total_tokens", None), + ) + + finally: + # Always finish the span and clean up + span.__exit__(None, None, None) + + +def _failure_callback(kwargs, exception, start_time, end_time): + # type: (Dict[str, Any], Exception, datetime, datetime) -> None + """Handle request failure.""" + span = _get_metadata_dict(kwargs).get("_sentry_span") + if span is None: + return + + try: + # Capture the exception + event, hint = event_from_exception( + exception, + client_options=sentry_sdk.get_client().options, + mechanism={"type": "litellm", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + finally: + # Always finish the span and clean up + span.__exit__(type(exception), exception, None) + + +class LiteLLMIntegration(Integration): + """ + LiteLLM integration for Sentry. + + This integration automatically captures LiteLLM API calls and sends them to Sentry + for monitoring and error tracking. It supports all 100+ LLM providers that LiteLLM + supports, including OpenAI, Anthropic, Google, Cohere, and many others. + + Features: + - Automatic exception capture for all LiteLLM calls + - Token usage tracking across all providers + - Provider detection and attribution + - Input/output message capture (configurable) + - Streaming response support + - Cost tracking integration + + Usage: + + ```python + import litellm + import sentry_sdk + + # Initialize Sentry with the LiteLLM integration + sentry_sdk.init( + dsn="your-dsn", + send_default_pii=True + integrations=[ + sentry_sdk.integrations.LiteLLMIntegration( + include_prompts=True # Set to False to exclude message content + ) + ] + ) + + # All LiteLLM calls will now be monitored + response = litellm.completion( + model="gpt-3.5-turbo", + messages=[{"role": "user", "content": "Hello!"}] + ) + ``` + + Configuration: + - include_prompts (bool): Whether to include prompts and responses in spans. + Defaults to True. Set to False to exclude potentially sensitive data. + """ + + identifier = "litellm" + origin = f"auto.ai.{identifier}" + + def __init__(self, include_prompts=True): + # type: (LiteLLMIntegration, bool) -> None + self.include_prompts = include_prompts + + @staticmethod + def setup_once(): + # type: () -> None + """Set up LiteLLM callbacks for monitoring.""" + litellm.input_callback = litellm.input_callback or [] + if _input_callback not in litellm.input_callback: + litellm.input_callback.append(_input_callback) + + litellm.success_callback = litellm.success_callback or [] + if _success_callback not in litellm.success_callback: + litellm.success_callback.append(_success_callback) + + litellm.failure_callback = litellm.failure_callback or [] + if _failure_callback not in litellm.failure_callback: + litellm.failure_callback.append(_failure_callback) diff --git a/setup.py b/setup.py index 7119e20e90..6629726798 100644 --- a/setup.py +++ b/setup.py @@ -65,6 +65,7 @@ def get_file_text(file_name): "langchain": ["langchain>=0.0.210"], "langgraph": ["langgraph>=0.6.6"], "launchdarkly": ["launchdarkly-server-sdk>=9.8.0"], + "litellm": ["litellm>=1.77.5"], "litestar": ["litestar>=2.0.0"], "loguru": ["loguru>=0.5"], "openai": ["openai>=1.0.0", "tiktoken>=0.3.0"], diff --git a/tests/integrations/litellm/__init__.py b/tests/integrations/litellm/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/litellm/test_litellm.py b/tests/integrations/litellm/test_litellm.py new file mode 100644 index 0000000000..b600c32905 --- /dev/null +++ b/tests/integrations/litellm/test_litellm.py @@ -0,0 +1,547 @@ +import pytest +from unittest import mock +from datetime import datetime + +try: + from unittest.mock import AsyncMock +except ImportError: + + class AsyncMock(mock.MagicMock): + async def __call__(self, *args, **kwargs): + return super(AsyncMock, self).__call__(*args, **kwargs) + + +try: + import litellm +except ImportError: + pytest.skip("litellm not installed", allow_module_level=True) + +from sentry_sdk import start_transaction +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.integrations.litellm import ( + LiteLLMIntegration, + _input_callback, + _success_callback, + _failure_callback, +) +from sentry_sdk.utils import package_version + + +LITELLM_VERSION = package_version("litellm") + + +# Mock response objects +class MockMessage: + def __init__(self, role="assistant", content="Test response"): + self.role = role + self.content = content + self.tool_calls = None + + def model_dump(self): + return {"role": self.role, "content": self.content} + + +class MockChoice: + def __init__(self, message=None): + self.message = message or MockMessage() + self.index = 0 + self.finish_reason = "stop" + + +class MockUsage: + def __init__(self, prompt_tokens=10, completion_tokens=20, total_tokens=30): + self.prompt_tokens = prompt_tokens + self.completion_tokens = completion_tokens + self.total_tokens = total_tokens + + +class MockCompletionResponse: + def __init__( + self, + model="gpt-3.5-turbo", + choices=None, + usage=None, + ): + self.id = "chatcmpl-test" + self.model = model + self.choices = choices or [MockChoice()] + self.usage = usage or MockUsage() + self.object = "chat.completion" + self.created = 1234567890 + + +class MockEmbeddingData: + def __init__(self, embedding=None): + self.embedding = embedding or [0.1, 0.2, 0.3] + self.index = 0 + self.object = "embedding" + + +class MockEmbeddingResponse: + def __init__(self, model="text-embedding-ada-002", data=None, usage=None): + self.model = model + self.data = data or [MockEmbeddingData()] + self.usage = usage or MockUsage( + prompt_tokens=5, completion_tokens=0, total_tokens=5 + ) + self.object = "list" + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_nonstreaming_chat_completion( + sentry_init, capture_events, send_default_pii, include_prompts +): + sentry_init( + integrations=[LiteLLMIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + # Simulate what litellm does: call input callback, then success callback + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + assert len(events) == 1 + (event,) = events + + assert event["type"] == "transaction" + assert event["transaction"] == "litellm test" + + assert len(event["spans"]) == 1 + (span,) = event["spans"] + + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat gpt-3.5-turbo" + assert span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "gpt-3.5-turbo" + assert span["data"][SPANDATA.GEN_AI_RESPONSE_MODEL] == "gpt-3.5-turbo" + assert span["data"][SPANDATA.GEN_AI_SYSTEM] == "openai" + assert span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "chat" + + if send_default_pii and include_prompts: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT in span["data"] + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in span["data"] + + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + assert span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 20 + assert span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 30 + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_streaming_chat_completion( + sentry_init, capture_events, send_default_pii, include_prompts +): + sentry_init( + integrations=[LiteLLMIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + "stream": True, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + assert len(events) == 1 + (event,) = events + + assert event["type"] == "transaction" + assert len(event["spans"]) == 1 + (span,) = event["spans"] + + assert span["op"] == OP.GEN_AI_CHAT + assert span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True + + +def test_embeddings_create(sentry_init, capture_events): + sentry_init( + integrations=[LiteLLMIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + mock_response = MockEmbeddingResponse() + + with start_transaction(name="litellm test"): + # For embeddings, messages would be empty + kwargs = { + "model": "text-embedding-ada-002", + "input": "Hello!", + "messages": [], # Empty for embeddings + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + assert len(events) == 1 + (event,) = events + + assert event["type"] == "transaction" + assert len(event["spans"]) == 1 + (span,) = event["spans"] + + assert span["op"] == OP.GEN_AI_EMBEDDINGS + assert span["description"] == "embeddings text-embedding-ada-002" + assert span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "embeddings" + assert span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 5 + + +def test_exception_handling(sentry_init, capture_events): + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _failure_callback( + kwargs, + Exception("API rate limit reached"), + datetime.now(), + datetime.now(), + ) + + # Should have error event and transaction + assert len(events) >= 1 + # Find the error event + error_events = [e for e in events if e.get("level") == "error"] + assert len(error_events) == 1 + + +def test_span_origin(sentry_init, capture_events): + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + + assert event["contexts"]["trace"]["origin"] == "manual" + assert event["spans"][0]["origin"] == "auto.ai.litellm" + + +def test_multiple_providers(sentry_init, capture_events): + """Test that the integration correctly identifies different providers.""" + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + + # Test with different model prefixes + test_cases = [ + ("gpt-3.5-turbo", "openai"), + ("claude-3-opus-20240229", "anthropic"), + ("gemini/gemini-pro", "gemini"), + ] + + for model, _ in test_cases: + mock_response = MockCompletionResponse(model=model) + with start_transaction(name=f"test {model}"): + kwargs = { + "model": model, + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + assert len(events) == len(test_cases) + + for i in range(len(test_cases)): + span = events[i]["spans"][0] + # The provider should be detected by litellm.get_llm_provider + assert SPANDATA.GEN_AI_SYSTEM in span["data"] + + +def test_additional_parameters(sentry_init, capture_events): + """Test that additional parameters are captured.""" + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + "temperature": 0.7, + "max_tokens": 100, + "top_p": 0.9, + "frequency_penalty": 0.5, + "presence_penalty": 0.5, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + (span,) = event["spans"] + + assert span["data"][SPANDATA.GEN_AI_REQUEST_TEMPERATURE] == 0.7 + assert span["data"][SPANDATA.GEN_AI_REQUEST_MAX_TOKENS] == 100 + assert span["data"][SPANDATA.GEN_AI_REQUEST_TOP_P] == 0.9 + assert span["data"][SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY] == 0.5 + assert span["data"][SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY] == 0.5 + + +def test_litellm_specific_parameters(sentry_init, capture_events): + """Test that LiteLLM-specific parameters are captured.""" + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + "api_base": "https://custom-api.example.com/", + "api_version": "2023-01-01", + "custom_llm_provider": "custom_provider", + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + (span,) = event["spans"] + + assert span["data"]["gen_ai.litellm.api_base"] == "https://custom-api.example.com/" + assert span["data"]["gen_ai.litellm.api_version"] == "2023-01-01" + assert span["data"]["gen_ai.litellm.custom_llm_provider"] == "custom_provider" + + +def test_no_integration(sentry_init, capture_events): + """Test that when integration is not enabled, callbacks don't break.""" + sentry_init( + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + # When the integration isn't enabled, the callbacks should exit early + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + # These should not crash, just do nothing + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + # Should still have the transaction, but no child spans since integration is off + assert event["type"] == "transaction" + assert len(event.get("spans", [])) == 0 + + +def test_response_without_usage(sentry_init, capture_events): + """Test handling of responses without usage information.""" + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + + # Create a mock response without usage + mock_response = type( + "obj", + (object,), + { + "model": "gpt-3.5-turbo", + "choices": [MockChoice()], + }, + )() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + (span,) = event["spans"] + + # Span should still be created even without usage info + assert span["op"] == OP.GEN_AI_CHAT + assert span["description"] == "chat gpt-3.5-turbo" + + +def test_integration_setup(sentry_init): + """Test that the integration sets up the callbacks correctly.""" + sentry_init( + integrations=[LiteLLMIntegration()], + traces_sample_rate=1.0, + ) + + # Check that callbacks are registered + assert _input_callback in (litellm.input_callback or []) + assert _success_callback in (litellm.success_callback or []) + assert _failure_callback in (litellm.failure_callback or []) + + +def test_message_dict_extraction(sentry_init, capture_events): + """Test that response messages are properly extracted with dict() fallback.""" + sentry_init( + integrations=[LiteLLMIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + messages = [{"role": "user", "content": "Hello!"}] + + # Create a message that has dict() method instead of model_dump() + class DictMessage: + def __init__(self): + self.role = "assistant" + self.content = "Response" + self.tool_calls = None + + def dict(self): + return {"role": self.role, "content": self.content} + + mock_response = MockCompletionResponse(choices=[MockChoice(message=DictMessage())]) + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + (event,) = events + (span,) = event["spans"] + + # Should have extracted the response message + assert SPANDATA.GEN_AI_RESPONSE_TEXT in span["data"] diff --git a/tox.ini b/tox.ini index 1bca280a11..c74755f206 100644 --- a/tox.ini +++ b/tox.ini @@ -63,13 +63,15 @@ envlist = {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.2.17 {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 + {py3.9,py3.12,py3.13}-litellm-v1.77.5 + {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.8,py3.12,py3.13}-openai-base-v2.0.1 + {py3.8,py3.12,py3.13}-openai-base-v2.1.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.8,py3.12,py3.13}-openai-notiktoken-v2.0.1 + {py3.8,py3.12,py3.13}-openai-notiktoken-v2.1.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.8 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 @@ -89,7 +91,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.43 + {py3.9,py3.12,py3.13}-boto3-v1.40.44 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -267,7 +269,7 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.9,py3.12,py3.13}-trytond-v7.6.7 + {py3.9,py3.12,py3.13}-trytond-v7.6.8 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.8,py3.12,py3.13}-typer-v0.19.2 @@ -357,16 +359,18 @@ deps = langchain-notiktoken: langchain-openai langchain-notiktoken-v0.3.27: langchain-community + litellm-v1.77.5: litellm==1.77.5 + openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.0.1: openai==2.0.1 + openai-base-v2.1.0: openai==2.1.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.0.1: openai==2.0.1 + openai-notiktoken-v2.1.0: openai==2.1.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 @@ -390,7 +394,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.43: boto3==1.40.43 + boto3-v1.40.44: boto3==1.40.44 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -688,7 +692,7 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.6.7: trytond==7.6.7 + trytond-v7.6.8: trytond==7.6.8 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 @@ -746,6 +750,7 @@ setenv = langchain-notiktoken: TESTPATH=tests/integrations/langchain langgraph: TESTPATH=tests/integrations/langgraph launchdarkly: TESTPATH=tests/integrations/launchdarkly + litellm: TESTPATH=tests/integrations/litellm litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru openai-base: TESTPATH=tests/integrations/openai From 39cb2c53b5e896c4641d67e088e9ffca56226f35 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 6 Oct 2025 10:15:38 +0200 Subject: [PATCH 520/591] feat(huggingface): Support 1.0.0rc2 (#4873) ### Description huggingface_hub has a release candidate out and our test suite doesn't work with it. Two changes necessary: - 1.0 uses `httpx`, so our `responses` mocks don't work, we also need `pytest_httpx`. - With httpx we get additional `http.client` spans in the transaction, while before we were assuming the transaction only contains exactly one `gen_ai.*` span and nothing else. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4802 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/config.py | 3 +- scripts/populate_tox/releases.jsonl | 1 + .../huggingface_hub/test_huggingface_hub.py | 296 ++++++++++++++---- tox.ini | 3 + 4 files changed, 244 insertions(+), 59 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f69e5f2f90..0ff0e9b434 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -183,9 +183,8 @@ "huggingface_hub": { "package": "huggingface_hub", "deps": { - "*": ["responses"], + "*": ["responses", "pytest-httpx"], }, - "include": "<1.0", }, "langchain-base": { "package": "langchain", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index b7cca55815..bf95102c71 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -94,6 +94,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc2", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index 5aa3928a67..b9ab4df5bf 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -1,6 +1,8 @@ from unittest import mock import pytest +import re import responses +import httpx from huggingface_hub import InferenceClient @@ -32,17 +34,49 @@ ) +def _add_mock_response( + httpx_mock, rsps, method, url, json=None, status=200, body=None, headers=None +): + # HF v1+ uses httpx for making requests to their API, while <1 uses requests. + # Since we have to test both, we need mocks for both httpx and requests. + if HF_VERSION >= (1, 0, 0): + httpx_mock.add_response( + method=method, + url=url, + json=json, + content=body, + status_code=status, + headers=headers, + is_optional=True, + is_reusable=True, + ) + else: + rsps.add( + method=method, + url=url, + json=json, + body=body, + status=status, + headers=headers, + ) + + @pytest.fixture -def mock_hf_text_generation_api(): +def mock_hf_text_generation_api(httpx_mock): # type: () -> Any """Mock HuggingFace text generation API""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" - # Mock model info endpoint - rsps.add( - responses.GET, - MODEL_ENDPOINT.format(model_name=model_name), + _add_mock_response( + httpx_mock, + rsps, + "GET", + re.compile( + MODEL_ENDPOINT.format(model_name=model_name) + + r"(\?expand=inferenceProviderMapping)?" + ), json={ "id": model_name, "pipeline_tag": "text-generation", @@ -57,9 +91,10 @@ def mock_hf_text_generation_api(): status=200, ) - # Mock text generation endpoint - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name), json={ "generated_text": "[mocked] Hello! How can i help you?", @@ -73,61 +108,78 @@ def mock_hf_text_generation_api(): status=200, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_api_with_errors(): +def mock_hf_api_with_errors(httpx_mock): # type: () -> Any """Mock HuggingFace API that always raises errors for any request""" + with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint with error - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={"error": "Model not found"}, status=404, ) # Mock text generation endpoint with error - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name), json={"error": "Internal server error", "message": "Something went wrong"}, status=500, ) # Mock chat completion endpoint with error - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", json={"error": "Internal server error", "message": "Something went wrong"}, status=500, ) # Catch-all pattern for any other model requests - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", "https://huggingface.co/api/models/test-model-error", json={"error": "Generic model error"}, status=500, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_text_generation_api_streaming(): +def mock_hf_text_generation_api_streaming(httpx_mock): # type: () -> Any """Mock streaming HuggingFace text generation API""" with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={ "id": model_name, @@ -146,8 +198,10 @@ def mock_hf_text_generation_api_streaming(): # Mock text generation endpoint for streaming streaming_response = b'data:{"token":{"id":1, "special": false, "text": "the mocked "}}\n\ndata:{"token":{"id":2, "special": false, "text": "model response"}, "details":{"finish_reason": "length", "generated_tokens": 10, "seed": 0}}\n\n' - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name), body=streaming_response, status=200, @@ -158,19 +212,24 @@ def mock_hf_text_generation_api_streaming(): }, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_chat_completion_api(): +def mock_hf_chat_completion_api(httpx_mock): # type: () -> Any """Mock HuggingFace chat completion API""" with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={ "id": model_name, @@ -187,8 +246,10 @@ def mock_hf_chat_completion_api(): ) # Mock chat completion endpoint - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", json={ "id": "xyz-123", @@ -214,19 +275,24 @@ def mock_hf_chat_completion_api(): status=200, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_chat_completion_api_tools(): +def mock_hf_chat_completion_api_tools(httpx_mock): # type: () -> Any """Mock HuggingFace chat completion API with tool calls.""" with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={ "id": model_name, @@ -243,8 +309,10 @@ def mock_hf_chat_completion_api_tools(): ) # Mock chat completion endpoint - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", json={ "id": "xyz-123", @@ -279,19 +347,24 @@ def mock_hf_chat_completion_api_tools(): status=200, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_chat_completion_api_streaming(): +def mock_hf_chat_completion_api_streaming(httpx_mock): # type: () -> Any """Mock streaming HuggingFace chat completion API""" with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={ "id": model_name, @@ -313,8 +386,10 @@ def mock_hf_chat_completion_api_streaming(): b'data:{"id":"xyz-124","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","content":"model response"},"index":0,"finish_reason":"stop"}],"usage":{"prompt_tokens":183,"completion_tokens":14,"total_tokens":197}}\n\n' ) - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", body=streaming_chat_response, status=200, @@ -325,19 +400,24 @@ def mock_hf_chat_completion_api_streaming(): }, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps @pytest.fixture -def mock_hf_chat_completion_api_streaming_tools(): +def mock_hf_chat_completion_api_streaming_tools(httpx_mock): # type: () -> Any """Mock streaming HuggingFace chat completion API with tool calls.""" with responses.RequestsMock(assert_all_requests_are_fired=False) as rsps: model_name = "test-model" # Mock model info endpoint - rsps.add( - responses.GET, + _add_mock_response( + httpx_mock, + rsps, + "GET", MODEL_ENDPOINT.format(model_name=model_name), json={ "id": model_name, @@ -359,8 +439,10 @@ def mock_hf_chat_completion_api_streaming_tools(): b'data:{"id":"xyz-124","created":1234567890,"model":"test-model-123","system_fingerprint":"fp_123","choices":[{"delta":{"role":"assistant","tool_calls": [{"id": "call_123","type": "function","function": {"name": "get_weather", "arguments": {"location": "Paris"}}}]},"index":0,"finish_reason":"tool_calls"}],"usage":{"prompt_tokens":183,"completion_tokens":14,"total_tokens":197}}\n\n' ) - rsps.add( - responses.POST, + _add_mock_response( + httpx_mock, + rsps, + "POST", INFERENCE_ENDPOINT.format(model_name=model_name) + "/v1/chat/completions", body=streaming_chat_response, status=200, @@ -371,9 +453,13 @@ def mock_hf_chat_completion_api_streaming_tools(): }, ) - yield rsps + if HF_VERSION >= (1, 0, 0): + yield httpx_mock + else: + yield rsps +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_text_generation( @@ -401,7 +487,18 @@ def test_text_generation( ) (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.generate_text" assert span["description"] == "generate_text test-model" @@ -431,6 +528,7 @@ def test_text_generation( assert "gen_ai.response.model" not in span["data"] +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_text_generation_streaming( @@ -459,7 +557,18 @@ def test_text_generation_streaming( pass (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.generate_text" assert span["description"] == "generate_text test-model" @@ -489,6 +598,7 @@ def test_text_generation_streaming( assert "gen_ai.response.model" not in span["data"] +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_chat_completion( @@ -515,7 +625,18 @@ def test_chat_completion( ) (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.chat" assert span["description"] == "chat test-model" @@ -549,6 +670,7 @@ def test_chat_completion( assert span["data"] == expected_data +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_chat_completion_streaming( @@ -577,7 +699,18 @@ def test_chat_completion_streaming( ) (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.chat" assert span["description"] == "chat test-model" @@ -611,6 +744,7 @@ def test_chat_completion_streaming( assert span["data"] == expected_data +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) def test_chat_completion_api_error( sentry_init, capture_events, mock_hf_api_with_errors ): @@ -634,7 +768,17 @@ def test_chat_completion_api_error( assert error["exception"]["values"][0]["mechanism"]["type"] == "huggingface_hub" assert not error["exception"]["values"][0]["mechanism"]["handled"] - (span,) = transaction["spans"] + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.chat" assert span["description"] == "chat test-model" @@ -654,6 +798,7 @@ def test_chat_completion_api_error( assert span["data"] == expected_data +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) def test_span_status_error(sentry_init, capture_events, mock_hf_api_with_errors): # type: (Any, Any, Any) -> None sentry_init(traces_sample_rate=1.0) @@ -669,10 +814,24 @@ def test_span_status_error(sentry_init, capture_events, mock_hf_api_with_errors) (error, transaction) = events assert error["level"] == "error" - assert transaction["spans"][0]["tags"]["status"] == "error" + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None + assert span["tags"]["status"] == "error" + assert transaction["contexts"]["trace"]["status"] == "error" +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_chat_completion_with_tools( @@ -715,7 +874,18 @@ def test_chat_completion_with_tools( ) (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.chat" assert span["description"] == "chat test-model" @@ -750,6 +920,7 @@ def test_chat_completion_with_tools( assert span["data"] == expected_data +@pytest.mark.httpx_mock(assert_all_requests_were_expected=False) @pytest.mark.parametrize("send_default_pii", [True, False]) @pytest.mark.parametrize("include_prompts", [True, False]) def test_chat_completion_streaming_with_tools( @@ -795,7 +966,18 @@ def test_chat_completion_streaming_with_tools( ) (transaction,) = events - (span,) = transaction["spans"] + + span = None + for sp in transaction["spans"]: + if sp["op"].startswith("gen_ai"): + assert span is None, "there is exactly one gen_ai span" + span = sp + else: + # there should be no other spans, just the gen_ai span + # and optionally some http.client spans from talking to the hf api + assert sp["op"] == "http.client" + + assert span is not None assert span["op"] == "gen_ai.chat" assert span["description"] == "chat test-model" diff --git a/tox.ini b/tox.ini index c74755f206..41243ccc28 100644 --- a/tox.ini +++ b/tox.ini @@ -85,6 +85,7 @@ envlist = {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc2 # ~~~ Cloud ~~~ @@ -387,7 +388,9 @@ deps = huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 huggingface_hub-v0.35.3: huggingface_hub==0.35.3 + huggingface_hub-v1.0.0rc2: huggingface_hub==1.0.0rc2 huggingface_hub: responses + huggingface_hub: pytest-httpx # ~~~ Cloud ~~~ From 8bc196620c70e437f24128e20740dd0990579ff9 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 6 Oct 2025 08:32:12 +0000 Subject: [PATCH 521/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/06)=20(#4889)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/releases.jsonl | 7 ++++--- tox.ini | 14 ++++++++------ 2 files changed, 12 insertions(+), 9 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index bf95102c71..d52a769def 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.44", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.45", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -78,6 +78,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.75.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.76.0rc1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}} @@ -102,10 +103,10 @@ {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.5", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.17.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} diff --git a/tox.ini b/tox.ini index 41243ccc28..d02b3c73ab 100644 --- a/tox.ini +++ b/tox.ini @@ -63,7 +63,7 @@ envlist = {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.2.17 {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 - {py3.9,py3.12,py3.13}-litellm-v1.77.5 + {py3.9,py3.12,py3.13}-litellm-v1.77.7 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 @@ -92,7 +92,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.44 + {py3.9,py3.12,py3.13}-boto3-v1.40.45 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -159,6 +159,7 @@ envlist = {py3.7,py3.9,py3.10}-grpc-v1.47.5 {py3.7,py3.11,py3.12}-grpc-v1.62.3 {py3.9,py3.12,py3.13}-grpc-v1.75.1 + {py3.9,py3.12,py3.13}-grpc-v1.76.0rc1 {py3.6,py3.8,py3.9}-httpx-v0.16.1 {py3.6,py3.9,py3.10}-httpx-v0.20.0 @@ -239,7 +240,7 @@ envlist = {py3.8,py3.10,py3.11}-litestar-v2.0.1 {py3.8,py3.11,py3.12}-litestar-v2.6.4 {py3.8,py3.11,py3.12}-litestar-v2.12.1 - {py3.8,py3.12,py3.13}-litestar-v2.17.0 + {py3.8,py3.12,py3.13}-litestar-v2.18.0 {py3.6}-pyramid-v1.8.6 {py3.6,py3.8,py3.9}-pyramid-v1.10.8 @@ -360,7 +361,7 @@ deps = langchain-notiktoken: langchain-openai langchain-notiktoken-v0.3.27: langchain-community - litellm-v1.77.5: litellm==1.77.5 + litellm-v1.77.7: litellm==1.77.7 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 @@ -397,7 +398,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.44: boto3==1.40.44 + boto3-v1.40.45: boto3==1.40.45 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -484,6 +485,7 @@ deps = grpc-v1.47.5: grpcio==1.47.5 grpc-v1.62.3: grpcio==1.62.3 grpc-v1.75.1: grpcio==1.75.1 + grpc-v1.76.0rc1: grpcio==1.76.0rc1 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -633,7 +635,7 @@ deps = litestar-v2.0.1: litestar==2.0.1 litestar-v2.6.4: litestar==2.6.4 litestar-v2.12.1: litestar==2.12.1 - litestar-v2.17.0: litestar==2.17.0 + litestar-v2.18.0: litestar==2.18.0 litestar: pytest-asyncio litestar: python-multipart litestar: requests From 41f709e27683ada43513be0edbaeae6ecc099ca0 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 6 Oct 2025 09:46:47 +0000 Subject: [PATCH 522/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/06)=20(#4890)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/releases.jsonl | 2 +- tox.ini | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index d52a769def..bd04eb7c28 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -191,7 +191,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.282.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} diff --git a/tox.ini b/tox.ini index d02b3c73ab..8eb04550fb 100644 --- a/tox.ini +++ b/tox.ini @@ -151,7 +151,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.9,py3.12,py3.13}-strawberry-v0.282.0 + {py3.9,py3.12,py3.13}-strawberry-v0.283.0 # ~~~ Network ~~~ @@ -475,7 +475,7 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.282.0: strawberry-graphql[fastapi,flask]==0.282.0 + strawberry-v0.283.0: strawberry-graphql[fastapi,flask]==0.283.0 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 From bf77a86b92edc8b281b32df99efb6c438e746b18 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Mon, 6 Oct 2025 13:27:23 +0200 Subject: [PATCH 523/591] fix(litestar): Copy request info to prevent cookies mutation (#4883) Prevent mutating cookies on incoming HTTP requests if the cookie name is in the scrubbers denylist. Cookies like `token=...` were replaced with `AnnotatedValue` because a shallow reference of the request information was held by the client. A deep copy is introduced so scrubbing does not interfere with Litestar, and in particular does not break `JWTCookieAuth`. Closes https://github.com/getsentry/sentry-python/issues/4882 --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/integrations/litestar.py | 4 +++- sentry_sdk/integrations/starlite.py | 4 +++- 2 files changed, 6 insertions(+), 2 deletions(-) diff --git a/sentry_sdk/integrations/litestar.py b/sentry_sdk/integrations/litestar.py index 745a00bcba..0cb9f4b972 100644 --- a/sentry_sdk/integrations/litestar.py +++ b/sentry_sdk/integrations/litestar.py @@ -1,4 +1,6 @@ from collections.abc import Set +from copy import deepcopy + import sentry_sdk from sentry_sdk.consts import OP from sentry_sdk.integrations import ( @@ -260,7 +262,7 @@ def event_processor(event, _): event.update( { - "request": request_info, + "request": deepcopy(request_info), "transaction": tx_name, "transaction_info": tx_info, } diff --git a/sentry_sdk/integrations/starlite.py b/sentry_sdk/integrations/starlite.py index daab82d642..855b87ad60 100644 --- a/sentry_sdk/integrations/starlite.py +++ b/sentry_sdk/integrations/starlite.py @@ -1,3 +1,5 @@ +from copy import deepcopy + import sentry_sdk from sentry_sdk.consts import OP from sentry_sdk.integrations import DidNotEnable, Integration @@ -237,7 +239,7 @@ def event_processor(event, _): event.update( { - "request": request_info, + "request": deepcopy(request_info), "transaction": tx_name, "transaction_info": tx_info, } From 91cc2bcd40bd0ce68e30a24d5201357d0ca321c8 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 6 Oct 2025 11:40:44 +0000 Subject: [PATCH 524/591] release: 2.40.0 --- CHANGELOG.md | 22 ++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 25 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 941aec99e6..bb147b2d8b 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,27 @@ # Changelog +## 2.40.0 + +### Various fixes & improvements + +- fix(litestar): Copy request info to prevent cookies mutation (#4883) by @alexander-alderman-webb +- ci: 🤖 Update test matrix with new releases (10/06) (#4890) by @github-actions +- ci: 🤖 Update test matrix with new releases (10/06) (#4889) by @github-actions +- feat(huggingface): Support 1.0.0rc2 (#4873) by @sentrivana +- feat(integrations): add litellm integration (#4864) by @constantinius +- feat(integrations): Add tracing to DramatiqIntegration (#4571) by @Igreh +- fix(tests): Don't assume release is set (#4879) by @sentrivana +- ci: 🤖 Update test matrix with new releases (10/02) (#4880) by @github-actions +- fix(openai-agents): also emit spans for MCP tool calls done by the LLM (#4875) by @constantinius +- feat: Option to not trace HTTP requests based on status codes (#4869) by @alexander-alderman-webb +- fix(openai-agents): Move _set_agent_data call to ai_client_span function (#4876) by @constantinius +- ci: 🤖 Update test matrix with new releases (09/29) (#4872) by @github-actions +- build(deps): update shibuya requirement from <2025.9.22 (#4871) by @dependabot +- feat: Add script to determine lowest supported versions (#4867) by @sentrivana +- docs: Update CONTRIBUTING.md (#4870) by @sentrivana +- ci: Replace `black` and `flake8` with `ruff`. (#4866) by @antonpirker +- feat(toxgen): Generate `TESTPATH` for integrated test suites (#4863) by @sentrivana + ## 2.39.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 292e0971e2..2f630c382b 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.39.0" +release = "2.40.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 7b9e61a065..43c7e857ac 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1343,4 +1343,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.39.0" +VERSION = "2.40.0" diff --git a/setup.py b/setup.py index 6629726798..fbb8694e5e 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.39.0", + version="2.40.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 070ecd045014c31b79ef131bcfb7f9a8f1ff8771 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 6 Oct 2025 13:50:13 +0200 Subject: [PATCH 525/591] Update CHANGELOG.md --- CHANGELOG.md | 54 +++++++++++++++++++++++++++++++++++----------------- 1 file changed, 37 insertions(+), 17 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index bb147b2d8b..0aef8e9681 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,23 +4,43 @@ ### Various fixes & improvements -- fix(litestar): Copy request info to prevent cookies mutation (#4883) by @alexander-alderman-webb -- ci: 🤖 Update test matrix with new releases (10/06) (#4890) by @github-actions -- ci: 🤖 Update test matrix with new releases (10/06) (#4889) by @github-actions -- feat(huggingface): Support 1.0.0rc2 (#4873) by @sentrivana -- feat(integrations): add litellm integration (#4864) by @constantinius -- feat(integrations): Add tracing to DramatiqIntegration (#4571) by @Igreh -- fix(tests): Don't assume release is set (#4879) by @sentrivana -- ci: 🤖 Update test matrix with new releases (10/02) (#4880) by @github-actions -- fix(openai-agents): also emit spans for MCP tool calls done by the LLM (#4875) by @constantinius -- feat: Option to not trace HTTP requests based on status codes (#4869) by @alexander-alderman-webb -- fix(openai-agents): Move _set_agent_data call to ai_client_span function (#4876) by @constantinius -- ci: 🤖 Update test matrix with new releases (09/29) (#4872) by @github-actions -- build(deps): update shibuya requirement from <2025.9.22 (#4871) by @dependabot -- feat: Add script to determine lowest supported versions (#4867) by @sentrivana -- docs: Update CONTRIBUTING.md (#4870) by @sentrivana -- ci: Replace `black` and `flake8` with `ruff`. (#4866) by @antonpirker -- feat(toxgen): Generate `TESTPATH` for integrated test suites (#4863) by @sentrivana +- Add LiteLLM integration (#4864) by @constantinius + Once you've enabled the new LiteLLM integration, you can use the Sentry AI Agents Monitoring, a Sentry dashboard that helps you understand what's going on with your AI requests: + + ```python + import sentry_sdk + from sentry_sdk.integrations.litellm import LiteLLMIntegration + sentry_sdk.init( + dsn="", + # Set traces_sample_rate to 1.0 to capture 100% + # of transactions for tracing. + traces_sample_rate=1.0, + # Add data like inputs and responses; + # see https://docs.sentry.io/platforms/python/data-management/data-collected/ for more info + send_default_pii=True, + integrations=[ + LiteLLMIntegration(), + ], + ) + ``` + +- Litestar: Copy request info to prevent cookies mutation (#4883) by @alexander-alderman-webb +- Add tracing to `DramatiqIntegration` (#4571) by @Igreh +- Also emit spans for MCP tool calls done by the LLM (#4875) by @constantinius +- Option to not trace HTTP requests based on status codes (#4869) by @alexander-alderman-webb + You can now disable transactions for incoming requests with specific HTTP status codes. The new `trace_ignore_status_codes` option accepts a `set` of status codes as integers. If a transaction wraps a request that results in one of the provided status codes, the transaction will be unsampled. + + ```python + import sentry_sdk + + sentry_sdk.init( + trace_ignore_status_codes={301, 302, 303, *range(305, 400), 404}, + ) + ``` + +- Move `_set_agent_data` call to `ai_client_span` function (#4876) by @constantinius +- Add script to determine lowest supported versions (#4867) by @sentrivana +- Update `CONTRIBUTING.md` (#4870) by @sentrivana ## 2.39.0 From 04968c4dc2923cf66635378bf3787b9c8a167625 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Mon, 6 Oct 2025 13:55:01 +0200 Subject: [PATCH 526/591] Add links to CHANGELOG.md --- CHANGELOG.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 0aef8e9681..0bcd623611 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,7 +5,7 @@ ### Various fixes & improvements - Add LiteLLM integration (#4864) by @constantinius - Once you've enabled the new LiteLLM integration, you can use the Sentry AI Agents Monitoring, a Sentry dashboard that helps you understand what's going on with your AI requests: + Once you've enabled the [new LiteLLM integration](https://docs.sentry.io/platforms/python/integrations/litellm/), you can use the Sentry AI Agents Monitoring, a Sentry dashboard that helps you understand what's going on with your AI requests: ```python import sentry_sdk @@ -28,7 +28,7 @@ - Add tracing to `DramatiqIntegration` (#4571) by @Igreh - Also emit spans for MCP tool calls done by the LLM (#4875) by @constantinius - Option to not trace HTTP requests based on status codes (#4869) by @alexander-alderman-webb - You can now disable transactions for incoming requests with specific HTTP status codes. The new `trace_ignore_status_codes` option accepts a `set` of status codes as integers. If a transaction wraps a request that results in one of the provided status codes, the transaction will be unsampled. + You can now disable transactions for incoming requests with specific HTTP status codes. The [new `trace_ignore_status_codes` option](https://docs.sentry.io/platforms/python/configuration/options/#trace_ignore_status_codes) accepts a `set` of status codes as integers. If a transaction wraps a request that results in one of the provided status codes, the transaction will be unsampled. ```python import sentry_sdk From a879d8292b90becc582d5c0ee1e5edf1072dea10 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 8 Oct 2025 08:32:40 +0200 Subject: [PATCH 527/591] ci: Remove toxgen check (#4892) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit ### Description Removing the Check CI Config step altogether as well as associated parts of the toxgen script (`fail_on_changes`). Added a BIG ALL CAPS WARNING to `tox.ini` instead. Also updated the toxgen readme a bit. Removing the check should be fine because we haven't actually seen cases of people trying to edit `tox.ini` directly -- if this happens in the future it's easy to notice in the PR. If we don't notice it then, we can notice it during the weekly toxgen update. And if don't notice it then, the file simply gets overwritten. 🤷🏻‍♀️ ### The Problem With Checking `tox.ini`: The Long Read In order to check manual changes to `tox.ini` on a PR, we hash the committed file, then run toxgen, hash the result, and compare. If the hashes differ, we fail the check. This works fine as long as there have been no new releases between the two points in time when `tox.ini` was last committed and when we ran the check. This is usually not the case. There are new releases all the time. When we then rerun toxgen, the resulting `tox.ini` is different from the committed one because it contains the new releases. So the hashes are different without any manual changes to the file. One solution to this is always saving the timestamp of the last time `tox.ini` was generated, and then when rerunning toxgen for the purposes of the check, ignoring all new releases past the timestamp. This means any changes we detect were actually made by the user. However, the explicit timestamp is prone to merge conflicts. Anytime `master` has had a toxgen update, and a PR is made that also ran toxgen, the PR will have a merge conflict on the timestamp field that needs to be sorted out manually. This is annoying and unnecessary. (An attempt was made to use an implicit timestamp instead in the form of the commit timestamp, but this doesn't work since we squash commits on master, so the timestamp of the last commit that touched `tox.ini` is actually much later than the change was made. There are also other problems, like someone running toxgen but committing the change much later, etc.) ### Solutions considered - using a custom merge driver to resolve the timestamp conflict automatically (doesn't work on GH PRs) - running toxgen in CI on each PR and committing the change (would work but we're essentially already doing this with the cron job every week) - not checking in `tox.ini` at all, but running toxgen on each PR (introduces new package releases unrelated to the PR, no test setup committed -- contributors and package index maintainers also need to run our tests) - finding a different commit to use as the implicit timestamp (doesn't work because we squash commits on `master`) - ... In the end I decided to just get rid of the check. If people modifying `tox.ini` manually becomes a problem, we can deal with it then. I've added a big warning to `tox.ini` to discourage this. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4886 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/ci.yml | 22 -------- scripts/populate_tox/README.md | 37 +++++++------ scripts/populate_tox/config.py | 5 +- scripts/populate_tox/populate_tox.py | 81 +++++----------------------- scripts/populate_tox/releases.jsonl | 8 +-- scripts/populate_tox/tox.jinja | 22 ++++---- tox.ini | 44 +++++++-------- 7 files changed, 74 insertions(+), 145 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 50ab1d39ce..8ea11db711 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -33,28 +33,6 @@ jobs: pip install tox tox -e linters - check-ci-config: - name: Check CI config - runs-on: ubuntu-latest - timeout-minutes: 10 - - steps: - - uses: actions/checkout@v5.0.0 - with: - ref: ${{ github.event.pull_request.head.sha }} - fetch-depth: 0 - - uses: actions/setup-python@v6 - with: - python-version: 3.12 - - - name: Detect unexpected changes to tox.ini or CI - run: | - pip install -e . - pip install -r scripts/populate_tox/requirements.txt - python scripts/populate_tox/populate_tox.py --fail-on-changes - pip install -r scripts/split_tox_gh_actions/requirements.txt - python scripts/split_tox_gh_actions/split_tox_gh_actions.py --fail-on-changes - build_lambda_layer: name: Build Package runs-on: ubuntu-latest diff --git a/scripts/populate_tox/README.md b/scripts/populate_tox/README.md index 9bdb3567b8..d6c4e52147 100644 --- a/scripts/populate_tox/README.md +++ b/scripts/populate_tox/README.md @@ -14,7 +14,7 @@ combination of hardcoded and generated entries. The `populate_tox.py` script fills out the auto-generated part of that template. It does this by querying PyPI for each framework's package and its metadata and -then determining which versions make sense to test to get good coverage. +then determining which versions it makes sense to test to get good coverage. By default, the lowest supported and latest version of a framework are always tested, with a number of releases in between: @@ -22,17 +22,16 @@ tested, with a number of releases in between: - If the package doesn't have multiple majors, we pick two versions in between lowest and highest. -#### Caveats +Each test suite requires at least some configuration to be added to +`TEST_SUITE_CONFIG` in `scripts/populate_tox/config.py`. If you're adding a new +integration, check out the [Add a new test suite](#add-a-new-test-suite) section. -- Make sure the integration name is the same everywhere. If it consists of - multiple words, use an underscore instead of a hyphen. +## Test suite config -## Defining constraints - -The `TEST_SUITE_CONFIG` dictionary defines, for each integration test suite, -the main package (framework, library) to test with; any additional test -dependencies, optionally gated behind specific conditions; and optionally -the Python versions to test on. +The `TEST_SUITE_CONFIG` dictionary in `scripts/populate_tox/config.py` defines, +for each integration test suite, the main package (framework, library) to test +with; any additional test dependencies, optionally gated behind specific +conditions; and optionally the Python versions to test on. Constraints are defined using the format specified below. The following sections describe each key. @@ -58,7 +57,7 @@ in [packaging.specifiers](https://packaging.pypa.io/en/stable/specifiers.html). ### `package` -The name of the third party package as it's listed on PyPI. The script will +The name of the third-party package as it's listed on PyPI. The script will be picking different versions of this package to test. This key is mandatory. @@ -69,7 +68,7 @@ The test dependencies of the test suite. They're defined as a dictionary of `rule: [package1, package2, ...]` key-value pairs. All packages in the package list of a rule will be installed as long as the rule applies. -`rule`s are predefined. Each `rule` must be one of the following: +Each `rule` must be one of the following: - `*`: packages will be always installed - a version specifier on the main package (e.g. `<=0.32`): packages will only be installed if the main package falls into the version bounds specified @@ -77,7 +76,7 @@ in the package list of a rule will be installed as long as the rule applies. installed if the Python version matches one from the list Rules can be used to specify version bounds on older versions of the main -package's dependencies, for example. If e.g. Flask tests generally need +package's dependencies, for example. If Flask tests generally need Werkzeug and don't care about its version, but Flask older than 3.0 needs a specific Werkzeug version to work, you can say: @@ -176,7 +175,7 @@ be expressed like so: ### `integration_name` Sometimes, the name of the test suite doesn't match the name of the integration. -For example, we have the `openai_base` and `openai_notiktoken` test suites, both +For example, we have the `openai-base` and `openai-notiktoken` test suites, both of which are actually testing the `openai` integration. If this is the case, you can use the `integration_name` key to define the name of the integration. If not provided, it will default to the name of the test suite. @@ -193,6 +192,11 @@ greater than 2, as the oldest and latest supported versions will always be picked. Additionally, if there is a recent prerelease, it'll also always be picked (this doesn't count towards `num_versions`). +For instance, `num_versions` set to `2` will only test the first supported and +the last release of the package. `num_versions` equal to `3` will test the first +supported, the last release, and one release in between; `num_versions` set to `4` +will test an additional release in between. In all these cases, if there is +a recent prerelease, it'll be picked as well in addition to the picked versions. ## How-Tos @@ -202,9 +206,10 @@ picked (this doesn't count towards `num_versions`). in `integrations/__init__.py`. This should be the lowest version of the framework that we can guarantee works with the SDK. If you've just added the integration, you should generally set this to the latest version of the framework - at the time. + at the time, unless you've verified the integration works for earlier versions + as well. 2. Add the integration and any constraints to `TEST_SUITE_CONFIG`. See the - "Defining constraints" section for the format. + [Test suite config](#test-suite-config) section for the format. 3. Add the integration to one of the groups in the `GROUPS` dictionary in `scripts/split_tox_gh_actions/split_tox_gh_actions.py`. 4. Run `scripts/generate-test-files.sh` and commit the changes. diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 0ff0e9b434..f6b90e75e6 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -1,7 +1,6 @@ # The TEST_SUITE_CONFIG dictionary defines, for each integration test suite, -# the main package (framework, library) to test with; any additional test -# dependencies, optionally gated behind specific conditions; and optionally -# the Python versions to test on. +# at least the main package (framework, library) to test with. Additional +# test dependencies, Python versions to test on, etc. can also be defined here. # # See scripts/populate_tox/README.md for more info on the format and examples. diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index c0bf7f1a9f..453823f39d 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -130,7 +130,8 @@ def _save_to_cache(package: str, version: Version, release: Optional[dict]) -> N def _prefilter_releases( - integration: str, releases: dict[str, dict], older_than: Optional[datetime] = None + integration: str, + releases: dict[str, dict], ) -> tuple[list[Version], Optional[Version]]: """ Filter `releases`, removing releases that are for sure unsupported. @@ -178,9 +179,6 @@ def _prefilter_releases( uploaded = datetime.fromisoformat(meta["upload_time_iso_8601"]) - if older_than is not None and uploaded > older_than: - continue - if CUTOFF is not None and uploaded < CUTOFF: continue @@ -224,7 +222,7 @@ def _prefilter_releases( def get_supported_releases( - integration: str, pypi_data: dict, older_than: Optional[datetime] = None + integration: str, pypi_data: dict ) -> tuple[list[Version], Optional[Version]]: """ Get a list of releases that are currently supported by the SDK. @@ -236,9 +234,6 @@ def get_supported_releases( We return the list of supported releases and optionally also the newest prerelease, if it should be tested (meaning it's for a version higher than the current stable version). - - If an `older_than` timestamp is provided, no release newer than that will be - considered. """ package = pypi_data["info"]["name"] @@ -246,7 +241,8 @@ def get_supported_releases( # (because that might require an additional API call for some # of the releases) releases, latest_prerelease = _prefilter_releases( - integration, pypi_data["releases"], older_than + integration, + pypi_data["releases"], ) def _supports_lowest(release: Version) -> bool: @@ -665,32 +661,10 @@ def _normalize_release(release: dict) -> dict: return normalized -def main(fail_on_changes: bool = False) -> dict[str, list]: +def main() -> dict[str, list]: """ Generate tox.ini from the tox.jinja template. - - The script has two modes of operation: - - fail on changes mode (if `fail_on_changes` is True) - - normal mode (if `fail_on_changes` is False) - - Fail on changes mode is run on every PR to make sure that `tox.ini`, - `tox.jinja` and this script don't go out of sync because of manual changes - in one place but not the other. - - Normal mode is meant to be run as a cron job, regenerating tox.ini and - proposing the changes via a PR. """ - print(f"Running in {'fail_on_changes' if fail_on_changes else 'normal'} mode.") - last_updated = get_last_updated() - if fail_on_changes: - # We need to make the script ignore any new releases after the last updated - # timestamp so that we don't fail CI on a PR just because a new package - # version was released, leading to unrelated changes in tox.ini. - print( - f"Since we're in fail_on_changes mode, we're only considering " - f"releases before the last tox.ini update at {last_updated.isoformat()}." - ) - global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION meta = _fetch_sdk_metadata() sdk_python_versions = _parse_python_versions_from_classifiers( @@ -736,12 +710,7 @@ def main(fail_on_changes: bool = False) -> dict[str, list]: # Get the list of all supported releases - # If in fail-on-changes mode, ignore releases newer than `last_updated` - older_than = last_updated if fail_on_changes else None - - releases, latest_prerelease = get_supported_releases( - integration, pypi_data, older_than - ) + releases, latest_prerelease = get_supported_releases(integration, pypi_data) if not releases: print(" Found no supported releases.") @@ -778,9 +747,6 @@ def main(fail_on_changes: bool = False) -> dict[str, list]: } ) - if fail_on_changes: - old_file_hash = get_file_hash() - write_tox_file(packages) # Sort the release cache file @@ -798,36 +764,13 @@ def main(fail_on_changes: bool = False) -> dict[str, list]: ): releases_cache.write(json.dumps(release) + "\n") - if fail_on_changes: - new_file_hash = get_file_hash() - if old_file_hash != new_file_hash: - raise RuntimeError( - dedent( - """ - Detected that `tox.ini` is out of sync with - `scripts/populate_tox/tox.jinja` and/or - `scripts/populate_tox/populate_tox.py`. This might either mean - that `tox.ini` was changed manually, or the `tox.jinja` - template and/or the `populate_tox.py` script were changed without - regenerating `tox.ini`. - - Please don't make manual changes to `tox.ini`. Instead, make the - changes to the `tox.jinja` template and/or the `populate_tox.py` - script (as applicable) and regenerate the `tox.ini` file by - running scripts/generate-test-files.sh - """ - ) - ) - print("Done checking tox.ini. Looking good!") - else: - print( - "Done generating tox.ini. Make sure to also update the CI YAML " - "files to reflect the new test targets." - ) + print( + "Done generating tox.ini. Make sure to also update the CI YAML " + "files to reflect the new test targets." + ) return packages if __name__ == "__main__": - fail_on_changes = len(sys.argv) == 2 and sys.argv[1] == "--fail-on-changes" - main(fail_on_changes) + main() diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index bd04eb7c28..9f937e5e77 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -20,7 +20,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.12.15", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.13.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.45", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.46", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -111,7 +111,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.1.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.2.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} @@ -191,7 +191,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 2a33e7790d..40ea309b08 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -1,14 +1,16 @@ -# Tox (http://codespeak.net/~hpk/tox/) is a tool for running tests -# in multiple virtualenvs. This configuration file will run the -# test suite on all supported python versions. To use it, "pip install tox" -# and then run "tox" from this directory. +# DON'T EDIT THIS FILE BY HAND. This file has been generated from a template by +# `scripts/populate_tox/populate_tox.py`. # -# This file has been generated from a template -# by "scripts/populate_tox/populate_tox.py". Any changes to the file should -# be made in the template (if you want to change a hardcoded part of the file) -# or in the script (if you want to change the auto-generated part). -# The file (and all resulting CI YAMLs) then needs to be regenerated via -# "scripts/generate-test-files.sh". +# Any changes to the test matrix should be made +# - either in the script config in `scripts/populate_tox/config.py` (if you want +# to change the auto-generated part) +# - or in the template in `scripts/populate_tox/tox.jinja` (if you want to change +# a hardcoded part of the file) +# +# This file (and all resulting CI YAMLs) then needs to be regenerated via +# `scripts/generate-test-files.sh`. +# +# See also `scripts/populate_tox/README.md` for more info. [tox] requires = diff --git a/tox.ini b/tox.ini index 8eb04550fb..2c77edd07c 100644 --- a/tox.ini +++ b/tox.ini @@ -1,14 +1,16 @@ -# Tox (http://codespeak.net/~hpk/tox/) is a tool for running tests -# in multiple virtualenvs. This configuration file will run the -# test suite on all supported python versions. To use it, "pip install tox" -# and then run "tox" from this directory. +# DON'T EDIT THIS FILE BY HAND. This file has been generated from a template by +# `scripts/populate_tox/populate_tox.py`. # -# This file has been generated from a template -# by "scripts/populate_tox/populate_tox.py". Any changes to the file should -# be made in the template (if you want to change a hardcoded part of the file) -# or in the script (if you want to change the auto-generated part). -# The file (and all resulting CI YAMLs) then needs to be regenerated via -# "scripts/generate-test-files.sh". +# Any changes to the test matrix should be made +# - either in the script config in `scripts/populate_tox/config.py` (if you want +# to change the auto-generated part) +# - or in the template in `scripts/populate_tox/tox.jinja` (if you want to change +# a hardcoded part of the file) +# +# This file (and all resulting CI YAMLs) then needs to be regenerated via +# `scripts/generate-test-files.sh`. +# +# See also `scripts/populate_tox/README.md` for more info. [tox] requires = @@ -67,11 +69,11 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.8,py3.12,py3.13}-openai-base-v2.1.0 + {py3.8,py3.12,py3.13}-openai-base-v2.2.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.8,py3.12,py3.13}-openai-notiktoken-v2.1.0 + {py3.8,py3.12,py3.13}-openai-notiktoken-v2.2.0 {py3.9,py3.12,py3.13}-langgraph-v0.6.8 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 @@ -92,7 +94,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.45 + {py3.9,py3.12,py3.13}-boto3-v1.40.46 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -151,7 +153,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.9,py3.12,py3.13}-strawberry-v0.283.0 + {py3.9,py3.12,py3.13}-strawberry-v0.283.1 # ~~~ Network ~~~ @@ -227,7 +229,7 @@ envlist = {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.12.15 + {py3.9,py3.12,py3.13}-aiohttp-v3.13.0 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -365,14 +367,14 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.1.0: openai==2.1.0 + openai-base-v2.2.0: openai==2.2.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.1.0: openai==2.1.0 + openai-notiktoken-v2.2.0: openai==2.2.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 @@ -398,7 +400,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.45: boto3==1.40.45 + boto3-v1.40.46: boto3==1.40.46 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -475,7 +477,7 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.283.0: strawberry-graphql[fastapi,flask]==0.283.0 + strawberry-v0.283.1: strawberry-graphql[fastapi,flask]==0.283.1 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 @@ -618,10 +620,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.12.15: aiohttp==3.12.15 + aiohttp-v3.13.0: aiohttp==3.13.0 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.12.15: pytest-asyncio + aiohttp-v3.13.0: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.4: bottle==0.13.4 From b1dd2dcf3b63c73ab91f25ca217a28cf3ea7f6e9 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Wed, 8 Oct 2025 12:41:55 +0200 Subject: [PATCH 528/591] fix(ai): add mapping for gen_ai message roles (#4884) - Add a constant that contains the allowed message roles according to OTEL and a mapping - Apply that mapping to all gen_ai integrations - We will track input roles that do not conform to expectations via a Sentry issue in agent monitoring to make sure we continually update the mappings --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/ai/__init__.py | 7 + sentry_sdk/ai/utils.py | 48 ++++++ sentry_sdk/integrations/anthropic.py | 12 +- sentry_sdk/integrations/langchain.py | 33 +++- sentry_sdk/integrations/langgraph.py | 8 +- sentry_sdk/integrations/openai.py | 5 +- .../openai_agents/spans/invoke_agent.py | 12 +- .../integrations/openai_agents/utils.py | 53 ++++--- .../integrations/anthropic/test_anthropic.py | 67 +++++++++ .../integrations/langchain/test_langchain.py | 141 ++++++++++++++++++ .../integrations/langgraph/test_langgraph.py | 71 +++++++++ tests/integrations/openai/test_openai.py | 53 +++++++ .../openai_agents/test_openai_agents.py | 46 ++++++ 13 files changed, 525 insertions(+), 31 deletions(-) diff --git a/sentry_sdk/ai/__init__.py b/sentry_sdk/ai/__init__.py index e69de29bb2..fbcb9c061d 100644 --- a/sentry_sdk/ai/__init__.py +++ b/sentry_sdk/ai/__init__.py @@ -0,0 +1,7 @@ +from .utils import ( + set_data_normalized, + GEN_AI_MESSAGE_ROLE_MAPPING, + GEN_AI_MESSAGE_ROLE_REVERSE_MAPPING, + normalize_message_role, + normalize_message_roles, +) # noqa: F401 diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index d0ccf1bed3..0c0b937006 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -10,6 +10,26 @@ from sentry_sdk.utils import logger +class GEN_AI_ALLOWED_MESSAGE_ROLES: + SYSTEM = "system" + USER = "user" + ASSISTANT = "assistant" + TOOL = "tool" + + +GEN_AI_MESSAGE_ROLE_REVERSE_MAPPING = { + GEN_AI_ALLOWED_MESSAGE_ROLES.SYSTEM: ["system"], + GEN_AI_ALLOWED_MESSAGE_ROLES.USER: ["user", "human"], + GEN_AI_ALLOWED_MESSAGE_ROLES.ASSISTANT: ["assistant", "ai"], + GEN_AI_ALLOWED_MESSAGE_ROLES.TOOL: ["tool", "tool_call"], +} + +GEN_AI_MESSAGE_ROLE_MAPPING = {} +for target_role, source_roles in GEN_AI_MESSAGE_ROLE_REVERSE_MAPPING.items(): + for source_role in source_roles: + GEN_AI_MESSAGE_ROLE_MAPPING[source_role] = target_role + + def _normalize_data(data, unpack=True): # type: (Any, bool) -> Any # convert pydantic data (e.g. OpenAI v1+) to json compatible format @@ -40,6 +60,34 @@ def set_data_normalized(span, key, value, unpack=True): span.set_data(key, json.dumps(normalized)) +def normalize_message_role(role): + # type: (str) -> str + """ + Normalize a message role to one of the 4 allowed gen_ai role values. + Maps "ai" -> "assistant" and keeps other standard roles unchanged. + """ + return GEN_AI_MESSAGE_ROLE_MAPPING.get(role, role) + + +def normalize_message_roles(messages): + # type: (list[dict[str, Any]]) -> list[dict[str, Any]] + """ + Normalize roles in a list of messages to use standard gen_ai role values. + Creates a deep copy to avoid modifying the original messages. + """ + normalized_messages = [] + for message in messages: + if not isinstance(message, dict): + normalized_messages.append(message) + continue + normalized_message = message.copy() + if "role" in message: + normalized_message["role"] = normalize_message_role(message["role"]) + normalized_messages.append(normalized_message) + + return normalized_messages + + def get_start_span_function(): # type: () -> Callable[..., Any] current_span = sentry_sdk.get_current_span() diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index d9898fa1d1..46c6b2a766 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -3,7 +3,11 @@ import sentry_sdk from sentry_sdk.ai.monitoring import record_token_usage -from sentry_sdk.ai.utils import set_data_normalized, get_start_span_function +from sentry_sdk.ai.utils import ( + set_data_normalized, + normalize_message_roles, + get_start_span_function, +) from sentry_sdk.consts import OP, SPANDATA, SPANSTATUS from sentry_sdk.integrations import _check_minimum_version, DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -140,8 +144,12 @@ def _set_input_data(span, kwargs, integration): else: normalized_messages.append(message) + role_normalized_messages = normalize_message_roles(normalized_messages) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, normalized_messages, unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + role_normalized_messages, + unpack=False, ) set_data_normalized( diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index fdba26569d..724d908665 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -4,7 +4,12 @@ import sentry_sdk from sentry_sdk.ai.monitoring import set_ai_pipeline_name -from sentry_sdk.ai.utils import set_data_normalized, get_start_span_function +from sentry_sdk.ai.utils import ( + GEN_AI_ALLOWED_MESSAGE_ROLES, + normalize_message_roles, + set_data_normalized, + get_start_span_function, +) from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -209,8 +214,18 @@ def on_llm_start( _set_tools_on_span(span, all_params.get("tools")) if should_send_default_pii() and self.include_prompts: + normalized_messages = [ + { + "role": GEN_AI_ALLOWED_MESSAGE_ROLES.USER, + "content": {"type": "text", "text": prompt}, + } + for prompt in prompts + ] set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, prompts, unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + normalized_messages, + unpack=False, ) def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): @@ -262,6 +277,8 @@ def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): normalized_messages.append( self._normalize_langchain_message(message) ) + normalized_messages = normalize_message_roles(normalized_messages) + set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, @@ -740,8 +757,12 @@ def new_invoke(self, *args, **kwargs): and should_send_default_pii() and integration.include_prompts ): + normalized_messages = normalize_message_roles([input]) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, [input], unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + normalized_messages, + unpack=False, ) output = result.get("output") @@ -791,8 +812,12 @@ def new_stream(self, *args, **kwargs): and should_send_default_pii() and integration.include_prompts ): + normalized_messages = normalize_message_roles([input]) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, [input], unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + normalized_messages, + unpack=False, ) # Run the agent diff --git a/sentry_sdk/integrations/langgraph.py b/sentry_sdk/integrations/langgraph.py index df3941bb13..11aa1facf4 100644 --- a/sentry_sdk/integrations/langgraph.py +++ b/sentry_sdk/integrations/langgraph.py @@ -2,7 +2,7 @@ from typing import Any, Callable, List, Optional import sentry_sdk -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import set_data_normalized, normalize_message_roles from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -180,10 +180,11 @@ def new_invoke(self, *args, **kwargs): ): input_messages = _parse_langgraph_messages(args[0]) if input_messages: + normalized_input_messages = normalize_message_roles(input_messages) set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, - input_messages, + normalized_input_messages, unpack=False, ) @@ -230,10 +231,11 @@ async def new_ainvoke(self, *args, **kwargs): ): input_messages = _parse_langgraph_messages(args[0]) if input_messages: + normalized_input_messages = normalize_message_roles(input_messages) set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, - input_messages, + normalized_input_messages, unpack=False, ) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index e8b3b30ab2..e9bd2efa23 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -3,7 +3,7 @@ import sentry_sdk from sentry_sdk import consts from sentry_sdk.ai.monitoring import record_token_usage -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import set_data_normalized, normalize_message_roles from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -182,8 +182,9 @@ def _set_input_data(span, kwargs, operation, integration): and should_send_default_pii() and integration.include_prompts ): + normalized_messages = normalize_message_roles(messages) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, normalized_messages, unpack=False ) # Input attributes: Common diff --git a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py index cf06120625..2a9c5ebe66 100644 --- a/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py +++ b/sentry_sdk/integrations/openai_agents/spans/invoke_agent.py @@ -1,5 +1,9 @@ import sentry_sdk -from sentry_sdk.ai.utils import get_start_span_function, set_data_normalized +from sentry_sdk.ai.utils import ( + get_start_span_function, + set_data_normalized, + normalize_message_roles, +) from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import safe_serialize @@ -56,8 +60,12 @@ def invoke_agent_span(context, agent, kwargs): ) if len(messages) > 0: + normalized_messages = normalize_message_roles(messages) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + normalized_messages, + unpack=False, ) _set_agent_data(span, agent) diff --git a/sentry_sdk/integrations/openai_agents/utils.py b/sentry_sdk/integrations/openai_agents/utils.py index b0ad6bf903..125ff1175b 100644 --- a/sentry_sdk/integrations/openai_agents/utils.py +++ b/sentry_sdk/integrations/openai_agents/utils.py @@ -1,5 +1,10 @@ import sentry_sdk -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import ( + GEN_AI_ALLOWED_MESSAGE_ROLES, + normalize_message_roles, + set_data_normalized, + normalize_message_role, +) from sentry_sdk.consts import SPANDATA, SPANSTATUS, OP from sentry_sdk.integrations import DidNotEnable from sentry_sdk.scope import should_send_default_pii @@ -94,35 +99,47 @@ def _set_input_data(span, get_response_kwargs): # type: (sentry_sdk.tracing.Span, dict[str, Any]) -> None if not should_send_default_pii(): return + request_messages = [] - messages_by_role = { - "system": [], - "user": [], - "assistant": [], - "tool": [], - } # type: (dict[str, list[Any]]) system_instructions = get_response_kwargs.get("system_instructions") if system_instructions: - messages_by_role["system"].append({"type": "text", "text": system_instructions}) + request_messages.append( + { + "role": GEN_AI_ALLOWED_MESSAGE_ROLES.SYSTEM, + "content": [{"type": "text", "text": system_instructions}], + } + ) for message in get_response_kwargs.get("input", []): if "role" in message: - messages_by_role[message.get("role")].append( - {"type": "text", "text": message.get("content")} + normalized_role = normalize_message_role(message.get("role")) + request_messages.append( + { + "role": normalized_role, + "content": [{"type": "text", "text": message.get("content")}], + } ) else: if message.get("type") == "function_call": - messages_by_role["assistant"].append(message) + request_messages.append( + { + "role": GEN_AI_ALLOWED_MESSAGE_ROLES.ASSISTANT, + "content": [message], + } + ) elif message.get("type") == "function_call_output": - messages_by_role["tool"].append(message) - - request_messages = [] - for role, messages in messages_by_role.items(): - if len(messages) > 0: - request_messages.append({"role": role, "content": messages}) + request_messages.append( + { + "role": GEN_AI_ALLOWED_MESSAGE_ROLES.TOOL, + "content": [message], + } + ) set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, request_messages, unpack=False + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + normalize_message_roles(request_messages), + unpack=False, ) diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index 04ff12eb8b..e9065e2d32 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -1,5 +1,6 @@ import pytest from unittest import mock +import json try: from unittest.mock import AsyncMock @@ -878,3 +879,69 @@ def test_set_output_data_with_input_json_delta(sentry_init): assert span._data.get(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS) == 10 assert span._data.get(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS) == 20 assert span._data.get(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS) == 30 + + +def test_anthropic_message_role_mapping(sentry_init, capture_events): + """Test that Anthropic integration properly maps message roles like 'ai' to 'assistant'""" + sentry_init( + integrations=[AnthropicIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = Anthropic(api_key="z") + + def mock_messages_create(*args, **kwargs): + return Message( + id="msg_1", + content=[TextBlock(text="Hi there!", type="text")], + model="claude-3-opus", + role="assistant", + stop_reason="end_turn", + stop_sequence=None, + type="message", + usage=Usage(input_tokens=10, output_tokens=5), + ) + + client.messages._post = mock.Mock(return_value=mock_messages_create()) + + # Test messages with mixed roles including "ai" that should be mapped to "assistant" + test_messages = [ + {"role": "system", "content": "You are helpful."}, + {"role": "user", "content": "Hello"}, + {"role": "ai", "content": "Hi there!"}, # Should be mapped to "assistant" + {"role": "assistant", "content": "How can I help?"}, # Should stay "assistant" + ] + + with start_transaction(name="anthropic tx"): + client.messages.create( + model="claude-3-opus", max_tokens=10, messages=test_messages + ) + + (event,) = events + span = event["spans"][0] + + # Verify that the span was created correctly + assert span["op"] == "gen_ai.chat" + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"] + + # Parse the stored messages + stored_messages = json.loads(span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]) + + # Verify that "ai" role was mapped to "assistant" + assert len(stored_messages) == 4 + assert stored_messages[0]["role"] == "system" + assert stored_messages[1]["role"] == "user" + assert ( + stored_messages[2]["role"] == "assistant" + ) # "ai" should be mapped to "assistant" + assert stored_messages[3]["role"] == "assistant" # should stay "assistant" + + # Verify content is preserved + assert stored_messages[2]["content"] == "Hi there!" + assert stored_messages[3]["content"] == "How can I help?" + + # Verify no "ai" roles remain + roles = [msg["role"] for msg in stored_messages] + assert "ai" not in roles diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index ba49b2e508..661208432f 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -817,3 +817,144 @@ def test_langchain_integration_with_langchain_core_only(sentry_init, capture_eve assert llm_span["data"]["gen_ai.usage.total_tokens"] == 25 assert llm_span["data"]["gen_ai.usage.input_tokens"] == 10 assert llm_span["data"]["gen_ai.usage.output_tokens"] == 15 + + +def test_langchain_message_role_mapping(sentry_init, capture_events): + """Test that message roles are properly normalized in langchain integration.""" + global llm_type + llm_type = "openai-chat" + + sentry_init( + integrations=[LangchainIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + prompt = ChatPromptTemplate.from_messages( + [ + ("system", "You are a helpful assistant"), + ("human", "{input}"), + MessagesPlaceholder(variable_name="agent_scratchpad"), + ] + ) + + global stream_result_mock + stream_result_mock = Mock( + side_effect=[ + [ + ChatGenerationChunk( + type="ChatGenerationChunk", + message=AIMessageChunk(content="Test response"), + ), + ] + ] + ) + + llm = MockOpenAI( + model_name="gpt-3.5-turbo", + temperature=0, + openai_api_key="badkey", + ) + agent = create_openai_tools_agent(llm, [get_word_length], prompt) + agent_executor = AgentExecutor(agent=agent, tools=[get_word_length], verbose=True) + + # Test input that should trigger message role normalization + test_input = "Hello, how are you?" + + with start_transaction(): + list(agent_executor.stream({"input": test_input})) + + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + # Find spans with gen_ai operation that should have message data + gen_ai_spans = [ + span for span in tx.get("spans", []) if span.get("op", "").startswith("gen_ai") + ] + + # Check if any span has message data with normalized roles + message_data_found = False + for span in gen_ai_spans: + span_data = span.get("data", {}) + if SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data: + message_data_found = True + messages_data = span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] + + # Parse the message data (might be JSON string) + if isinstance(messages_data, str): + import json + + try: + messages = json.loads(messages_data) + except json.JSONDecodeError: + # If not valid JSON, skip this assertion + continue + else: + messages = messages_data + + # Verify that the input message is present and contains the test input + assert isinstance(messages, list) + assert len(messages) > 0 + + # The test input should be in one of the messages + input_found = False + for msg in messages: + if isinstance(msg, dict) and test_input in str(msg.get("content", "")): + input_found = True + break + elif isinstance(msg, str) and test_input in msg: + input_found = True + break + + assert input_found, ( + f"Test input '{test_input}' not found in messages: {messages}" + ) + break + + # The message role mapping functionality is primarily tested through the normalization + # that happens in the integration code. The fact that we can capture and process + # the messages without errors indicates the role mapping is working correctly. + assert message_data_found, "No span found with gen_ai request messages data" + + +def test_langchain_message_role_normalization_units(): + """Test the message role normalization functions directly.""" + from sentry_sdk.ai.utils import normalize_message_role, normalize_message_roles + + # Test individual role normalization + assert normalize_message_role("ai") == "assistant" + assert normalize_message_role("human") == "user" + assert normalize_message_role("tool_call") == "tool" + assert normalize_message_role("system") == "system" + assert normalize_message_role("user") == "user" + assert normalize_message_role("assistant") == "assistant" + assert normalize_message_role("tool") == "tool" + + # Test unknown role (should remain unchanged) + assert normalize_message_role("unknown_role") == "unknown_role" + + # Test message list normalization + test_messages = [ + {"role": "human", "content": "Hello"}, + {"role": "ai", "content": "Hi there!"}, + {"role": "tool_call", "content": "function_call"}, + {"role": "system", "content": "You are helpful"}, + {"content": "Message without role"}, + "string message", + ] + + normalized = normalize_message_roles(test_messages) + + # Verify the original messages are not modified + assert test_messages[0]["role"] == "human" # Original unchanged + assert test_messages[1]["role"] == "ai" # Original unchanged + + # Verify the normalized messages have correct roles + assert normalized[0]["role"] == "user" # human -> user + assert normalized[1]["role"] == "assistant" # ai -> assistant + assert normalized[2]["role"] == "tool" # tool_call -> tool + assert normalized[3]["role"] == "system" # system unchanged + assert "role" not in normalized[4] # Message without role unchanged + assert normalized[5] == "string message" # String message unchanged diff --git a/tests/integrations/langgraph/test_langgraph.py b/tests/integrations/langgraph/test_langgraph.py index 1510305b06..6ec6d9a96d 100644 --- a/tests/integrations/langgraph/test_langgraph.py +++ b/tests/integrations/langgraph/test_langgraph.py @@ -625,3 +625,74 @@ def original_invoke(self, *args, **kwargs): assert tool_calls_data[0]["function"]["name"] == "search" assert tool_calls_data[1]["id"] == "call_multi_2" assert tool_calls_data[1]["function"]["name"] == "calculate" + + +def test_langgraph_message_role_mapping(sentry_init, capture_events): + """Test that Langgraph integration properly maps message roles like 'ai' to 'assistant'""" + sentry_init( + integrations=[LanggraphIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + # Mock a langgraph message with mixed roles + class MockMessage: + def __init__(self, content, message_type="human"): + self.content = content + self.type = message_type + + # Create mock state with messages having different roles + state_data = { + "messages": [ + MockMessage("System prompt", "system"), + MockMessage("Hello", "human"), + MockMessage("Hi there!", "ai"), # Should be mapped to "assistant" + MockMessage("How can I help?", "assistant"), # Should stay "assistant" + ] + } + + compiled_graph = MockCompiledGraph("test_graph") + pregel = MockPregelInstance(compiled_graph) + + with start_transaction(name="langgraph tx"): + # Use the wrapped invoke function directly + from sentry_sdk.integrations.langgraph import _wrap_pregel_invoke + + wrapped_invoke = _wrap_pregel_invoke( + lambda self, state_data: {"result": "success"} + ) + wrapped_invoke(pregel, state_data) + + (event,) = events + span = event["spans"][0] + + # Verify that the span was created correctly + assert span["op"] == "gen_ai.invoke_agent" + + # If messages were captured, verify role mapping + if SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"]: + import json + + stored_messages = json.loads(span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]) + + # Find messages with specific content to verify role mapping + ai_message = next( + (msg for msg in stored_messages if msg.get("content") == "Hi there!"), None + ) + assistant_message = next( + (msg for msg in stored_messages if msg.get("content") == "How can I help?"), + None, + ) + + if ai_message: + # "ai" should have been mapped to "assistant" + assert ai_message["role"] == "assistant" + + if assistant_message: + # "assistant" should stay "assistant" + assert assistant_message["role"] == "assistant" + + # Verify no "ai" roles remain + roles = [msg["role"] for msg in stored_messages if "role" in msg] + assert "ai" not in roles diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index e7fbf8a7d8..06e0a09fcf 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1447,3 +1447,56 @@ def test_empty_tools_in_chat_completion(sentry_init, capture_events, tools): span = event["spans"][0] assert "gen_ai.request.available_tools" not in span["data"] + + +def test_openai_message_role_mapping(sentry_init, capture_events): + """Test that OpenAI integration properly maps message roles like 'ai' to 'assistant'""" + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.chat.completions._post = mock.Mock(return_value=EXAMPLE_CHAT_COMPLETION) + + # Test messages with mixed roles including "ai" that should be mapped to "assistant" + test_messages = [ + {"role": "system", "content": "You are helpful."}, + {"role": "user", "content": "Hello"}, + {"role": "ai", "content": "Hi there!"}, # Should be mapped to "assistant" + {"role": "assistant", "content": "How can I help?"}, # Should stay "assistant" + ] + + with start_transaction(name="openai tx"): + client.chat.completions.create(model="test-model", messages=test_messages) + + (event,) = events + span = event["spans"][0] + + # Verify that the span was created correctly + assert span["op"] == "gen_ai.chat" + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"] + + # Parse the stored messages + import json + + stored_messages = json.loads(span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]) + + # Verify that "ai" role was mapped to "assistant" + assert len(stored_messages) == 4 + assert stored_messages[0]["role"] == "system" + assert stored_messages[1]["role"] == "user" + assert ( + stored_messages[2]["role"] == "assistant" + ) # "ai" should be mapped to "assistant" + assert stored_messages[3]["role"] == "assistant" # should stay "assistant" + + # Verify content is preserved + assert stored_messages[2]["content"] == "Hi there!" + assert stored_messages[3]["content"] == "How can I help?" + + # Verify no "ai" roles remain + roles = [msg["role"] for msg in stored_messages] + assert "ai" not in roles diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index e9a8372806..e647ce9fad 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -1031,3 +1031,49 @@ async def run(): assert txn2["transaction"] == "test_agent workflow" assert txn3["type"] == "transaction" assert txn3["transaction"] == "test_agent workflow" + + +def test_openai_agents_message_role_mapping(sentry_init, capture_events): + """Test that OpenAI Agents integration properly maps message roles like 'ai' to 'assistant'""" + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + # Test input messages with mixed roles including "ai" + test_input = [ + {"role": "system", "content": "You are helpful."}, + {"role": "user", "content": "Hello"}, + {"role": "ai", "content": "Hi there!"}, # Should be mapped to "assistant" + {"role": "assistant", "content": "How can I help?"}, # Should stay "assistant" + ] + + get_response_kwargs = {"input": test_input} + + from sentry_sdk.integrations.openai_agents.utils import _set_input_data + from sentry_sdk import start_span + + with start_span(op="test") as span: + _set_input_data(span, get_response_kwargs) + + # Verify that messages were processed and roles were mapped + from sentry_sdk.consts import SPANDATA + + if SPANDATA.GEN_AI_REQUEST_MESSAGES in span._data: + import json + + stored_messages = json.loads(span._data[SPANDATA.GEN_AI_REQUEST_MESSAGES]) + + # Verify roles were properly mapped + found_assistant_roles = 0 + for message in stored_messages: + if message["role"] == "assistant": + found_assistant_roles += 1 + + # Should have 2 assistant roles (1 from original "assistant", 1 from mapped "ai") + assert found_assistant_roles == 2 + + # Verify no "ai" roles remain in any message + for message in stored_messages: + assert message["role"] != "ai" From f32e391a6fe22cc1d26c093e63ac4499e52b9d68 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 8 Oct 2025 14:12:24 +0200 Subject: [PATCH 529/591] feat: Add concurrent.futures patch to threading integration (#4770) Automatically fork isolation and current scopes when running tasks with `concurrent.future`. Packages the implementation from https://github.com/getsentry/sentry-python/issues/4508#issuecomment-3003526348 as an integration. Closes https://github.com/getsentry/sentry-python/issues/4565 --------- Co-authored-by: Anton Pirker --- sentry_sdk/integrations/threading.py | 60 +++++++++++++++--- .../integrations/threading/test_threading.py | 61 +++++++++++++++++++ 2 files changed, 113 insertions(+), 8 deletions(-) diff --git a/sentry_sdk/integrations/threading.py b/sentry_sdk/integrations/threading.py index c031c51f50..cfe54c829c 100644 --- a/sentry_sdk/integrations/threading.py +++ b/sentry_sdk/integrations/threading.py @@ -2,6 +2,7 @@ import warnings from functools import wraps from threading import Thread, current_thread +from concurrent.futures import ThreadPoolExecutor, Future import sentry_sdk from sentry_sdk.integrations import Integration @@ -24,6 +25,7 @@ from sentry_sdk._types import ExcInfo F = TypeVar("F", bound=Callable[..., Any]) + T = TypeVar("T", bound=Any) class ThreadingIntegration(Integration): @@ -59,6 +61,15 @@ def setup_once(): django_version = None channels_version = None + is_async_emulated_with_threads = ( + sys.version_info < (3, 9) + and channels_version is not None + and channels_version < "4.0.0" + and django_version is not None + and django_version >= (3, 0) + and django_version < (4, 0) + ) + @wraps(old_start) def sentry_start(self, *a, **kw): # type: (Thread, *Any, **Any) -> Any @@ -67,14 +78,7 @@ def sentry_start(self, *a, **kw): return old_start(self, *a, **kw) if integration.propagate_scope: - if ( - sys.version_info < (3, 9) - and channels_version is not None - and channels_version < "4.0.0" - and django_version is not None - and django_version >= (3, 0) - and django_version < (4, 0) - ): + if is_async_emulated_with_threads: warnings.warn( "There is a known issue with Django channels 2.x and 3.x when using Python 3.8 or older. " "(Async support is emulated using threads and some Sentry data may be leaked between those threads.) " @@ -109,6 +113,9 @@ def sentry_start(self, *a, **kw): return old_start(self, *a, **kw) Thread.start = sentry_start # type: ignore + ThreadPoolExecutor.submit = _wrap_threadpool_executor_submit( # type: ignore + ThreadPoolExecutor.submit, is_async_emulated_with_threads + ) def _wrap_run(isolation_scope_to_use, current_scope_to_use, old_run_func): @@ -134,6 +141,43 @@ def _run_old_run_func(): return run # type: ignore +def _wrap_threadpool_executor_submit(func, is_async_emulated_with_threads): + # type: (Callable[..., Future[T]], bool) -> Callable[..., Future[T]] + """ + Wrap submit call to propagate scopes on task submission. + """ + + @wraps(func) + def sentry_submit(self, fn, *args, **kwargs): + # type: (ThreadPoolExecutor, Callable[..., T], *Any, **Any) -> Future[T] + integration = sentry_sdk.get_client().get_integration(ThreadingIntegration) + if integration is None: + return func(self, fn, *args, **kwargs) + + if integration.propagate_scope and is_async_emulated_with_threads: + isolation_scope = sentry_sdk.get_isolation_scope() + current_scope = sentry_sdk.get_current_scope() + elif integration.propagate_scope: + isolation_scope = sentry_sdk.get_isolation_scope().fork() + current_scope = sentry_sdk.get_current_scope().fork() + else: + isolation_scope = None + current_scope = None + + def wrapped_fn(*args, **kwargs): + # type: (*Any, **Any) -> Any + if isolation_scope is not None and current_scope is not None: + with use_isolation_scope(isolation_scope): + with use_scope(current_scope): + return fn(*args, **kwargs) + + return fn(*args, **kwargs) + + return func(self, wrapped_fn, *args, **kwargs) + + return sentry_submit + + def _capture_exception(): # type: () -> ExcInfo exc_info = sys.exc_info() diff --git a/tests/integrations/threading/test_threading.py b/tests/integrations/threading/test_threading.py index 799298910b..9c9a24aa63 100644 --- a/tests/integrations/threading/test_threading.py +++ b/tests/integrations/threading/test_threading.py @@ -276,3 +276,64 @@ def do_some_work(number): - op="outer-submit-4": description="Thread: main"\ """ ) + + +@pytest.mark.parametrize( + "propagate_scope", + (True, False), + ids=["propagate_scope=True", "propagate_scope=False"], +) +def test_spans_from_threadpool( + sentry_init, capture_events, render_span_tree, propagate_scope +): + sentry_init( + traces_sample_rate=1.0, + integrations=[ThreadingIntegration(propagate_scope=propagate_scope)], + ) + events = capture_events() + + def do_some_work(number): + with sentry_sdk.start_span( + op=f"inner-run-{number}", name=f"Thread: child-{number}" + ): + pass + + with sentry_sdk.start_transaction(op="outer-trx"): + with futures.ThreadPoolExecutor(max_workers=1) as executor: + for number in range(5): + with sentry_sdk.start_span( + op=f"outer-submit-{number}", name="Thread: main" + ): + future = executor.submit(do_some_work, number) + future.result() + + (event,) = events + + if propagate_scope: + assert render_span_tree(event) == dedent( + """\ + - op="outer-trx": description=null + - op="outer-submit-0": description="Thread: main" + - op="inner-run-0": description="Thread: child-0" + - op="outer-submit-1": description="Thread: main" + - op="inner-run-1": description="Thread: child-1" + - op="outer-submit-2": description="Thread: main" + - op="inner-run-2": description="Thread: child-2" + - op="outer-submit-3": description="Thread: main" + - op="inner-run-3": description="Thread: child-3" + - op="outer-submit-4": description="Thread: main" + - op="inner-run-4": description="Thread: child-4"\ +""" + ) + + elif not propagate_scope: + assert render_span_tree(event) == dedent( + """\ + - op="outer-trx": description=null + - op="outer-submit-0": description="Thread: main" + - op="outer-submit-1": description="Thread: main" + - op="outer-submit-2": description="Thread: main" + - op="outer-submit-3": description="Thread: main" + - op="outer-submit-4": description="Thread: main"\ +""" + ) From 55e903e2c52b3026d434e67a4b18b04878894a0f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 8 Oct 2025 15:35:05 +0200 Subject: [PATCH 530/591] ci: Bump Python version for linting (#4897) ### Description Python 3.14 is out, let's use it for linting. #### Issues Ref https://github.com/getsentry/sentry-python/issues/4895 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/ci.yml | 2 +- scripts/populate_tox/tox.jinja | 2 +- tox.ini | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 8ea11db711..43f9d296a1 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -27,7 +27,7 @@ jobs: - uses: actions/checkout@v5.0.0 - uses: actions/setup-python@v6 with: - python-version: 3.12 + python-version: 3.14 - run: | pip install tox diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 40ea309b08..b86da57c24 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -181,7 +181,7 @@ basepython = # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python # version (configured in pyproject.toml), ensuring consistent behavior. - linters: python3.12 + linters: python3.14 commands = {py3.7,py3.8}-boto3: pip install urllib3<2.0.0 diff --git a/tox.ini b/tox.ini index 2c77edd07c..4bfc90cee9 100644 --- a/tox.ini +++ b/tox.ini @@ -812,7 +812,7 @@ basepython = # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python # version (configured in pyproject.toml), ensuring consistent behavior. - linters: python3.12 + linters: python3.14 commands = {py3.7,py3.8}-boto3: pip install urllib3<2.0.0 From 79973687822f0af0d7cda10e92a79aac7393a097 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 9 Oct 2025 10:27:36 +0200 Subject: [PATCH 531/591] chore: Remove old metrics code (#4899) ### Description Remove old metrics code to make way for https://github.com/getsentry/sentry-python/pull/4898 Metrics was always an experimental feature and Sentry stopped accepting metrics a year ago. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- sentry_sdk/_types.py | 22 - sentry_sdk/client.py | 27 -- sentry_sdk/consts.py | 7 - sentry_sdk/envelope.py | 4 +- sentry_sdk/metrics.py | 971 ---------------------------------------- sentry_sdk/tracing.py | 26 -- sentry_sdk/transport.py | 18 +- tests/test_envelope.py | 1 - tests/test_metrics.py | 971 ---------------------------------------- tests/test_transport.py | 111 ----- 10 files changed, 2 insertions(+), 2156 deletions(-) delete mode 100644 sentry_sdk/metrics.py delete mode 100644 tests/test_metrics.py diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index b28c7260ce..d057f215e4 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -210,7 +210,6 @@ class SDKInfo(TypedDict): "type": Literal["check_in", "transaction"], "user": dict[str, object], "_dropped_spans": int, - "_metrics_summary": dict[str, object], }, total=False, ) @@ -266,7 +265,6 @@ class SDKInfo(TypedDict): "internal", "profile", "profile_chunk", - "metric_bucket", "monitor", "span", "log_item", @@ -276,26 +274,6 @@ class SDKInfo(TypedDict): ContinuousProfilerMode = Literal["thread", "gevent", "unknown"] ProfilerMode = Union[ContinuousProfilerMode, Literal["sleep"]] - # Type of the metric. - MetricType = Literal["d", "s", "g", "c"] - - # Value of the metric. - MetricValue = Union[int, float, str] - - # Internal representation of tags as a tuple of tuples (this is done in order to allow for the same key to exist - # multiple times). - MetricTagsInternal = Tuple[Tuple[str, str], ...] - - # External representation of tags as a dictionary. - MetricTagValue = Union[str, int, float, None] - MetricTags = Mapping[str, MetricTagValue] - - # Value inside the generator for the metric value. - FlushedMetricValue = Union[int, float] - - BucketKey = Tuple[MetricType, str, MeasurementUnit, MetricTagsInternal] - MetricMetaKey = Tuple[MetricType, str, MeasurementUnit] - MonitorConfigScheduleType = Literal["crontab", "interval"] MonitorConfigScheduleUnit = Literal[ "year", diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index c06043ebe2..59a5013783 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -61,7 +61,6 @@ from sentry_sdk._types import Event, Hint, SDKInfo, Log from sentry_sdk.integrations import Integration - from sentry_sdk.metrics import MetricsAggregator from sentry_sdk.scope import Scope from sentry_sdk.session import Session from sentry_sdk.spotlight import SpotlightClient @@ -182,7 +181,6 @@ def __init__(self, options=None): self.transport = None # type: Optional[Transport] self.monitor = None # type: Optional[Monitor] - self.metrics_aggregator = None # type: Optional[MetricsAggregator] self.log_batcher = None # type: Optional[LogBatcher] def __getstate__(self, *args, **kwargs): @@ -361,26 +359,6 @@ def _capture_envelope(envelope): self.session_flusher = SessionFlusher(capture_func=_capture_envelope) - self.metrics_aggregator = None # type: Optional[MetricsAggregator] - experiments = self.options.get("_experiments", {}) - if experiments.get("enable_metrics", True): - # Context vars are not working correctly on Python <=3.6 - # with gevent. - metrics_supported = not is_gevent() or PY37 - if metrics_supported: - from sentry_sdk.metrics import MetricsAggregator - - self.metrics_aggregator = MetricsAggregator( - capture_func=_capture_envelope, - enable_code_locations=bool( - experiments.get("metric_code_locations", True) - ), - ) - else: - logger.info( - "Metrics not supported on Python 3.6 and lower with gevent." - ) - self.log_batcher = None if has_logs_enabled(self.options): @@ -467,7 +445,6 @@ def _capture_envelope(envelope): if ( self.monitor - or self.metrics_aggregator or self.log_batcher or has_profiling_enabled(self.options) or isinstance(self.transport, BaseHttpTransport) @@ -1019,8 +996,6 @@ def close( if self.transport is not None: self.flush(timeout=timeout, callback=callback) self.session_flusher.kill() - if self.metrics_aggregator is not None: - self.metrics_aggregator.kill() if self.log_batcher is not None: self.log_batcher.kill() if self.monitor: @@ -1045,8 +1020,6 @@ def flush( if timeout is None: timeout = self.options["shutdown_timeout"] self.session_flusher.flush() - if self.metrics_aggregator is not None: - self.metrics_aggregator.flush() if self.log_batcher is not None: self.log_batcher.flush() self.transport.flush(timeout=timeout, callback=callback) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 43c7e857ac..0f71a0d460 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -55,8 +55,6 @@ class CompressionAlgo(Enum): ProfilerMode, TracesSampler, TransactionProcessor, - MetricTags, - MetricValue, ) # Experiments are feature flags to enable and disable certain unstable SDK @@ -77,11 +75,6 @@ class CompressionAlgo(Enum): "transport_compression_algo": Optional[CompressionAlgo], "transport_num_pools": Optional[int], "transport_http2": Optional[bool], - "enable_metrics": Optional[bool], - "before_emit_metric": Optional[ - Callable[[str, MetricValue, MeasurementUnit, MetricTags], bool] - ], - "metric_code_locations": Optional[bool], "enable_logs": Optional[bool], "before_send_log": Optional[Callable[[Log, Hint], Optional[Log]]], }, diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index d9b2c1629a..b26c458d41 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -291,8 +291,6 @@ def data_category(self): return "profile" elif ty == "profile_chunk": return "profile_chunk" - elif ty == "statsd": - return "metric_bucket" elif ty == "check_in": return "monitor" else: @@ -354,7 +352,7 @@ def deserialize_from( # if no length was specified we need to read up to the end of line # and remove it (if it is present, i.e. not the very last char in an eof terminated envelope) payload = f.readline().rstrip(b"\n") - if headers.get("type") in ("event", "transaction", "metric_buckets"): + if headers.get("type") in ("event", "transaction"): rv = cls(headers=headers, payload=PayloadRef(json=parse_json(payload))) else: rv = cls(headers=headers, payload=payload) diff --git a/sentry_sdk/metrics.py b/sentry_sdk/metrics.py deleted file mode 100644 index d0041114ce..0000000000 --- a/sentry_sdk/metrics.py +++ /dev/null @@ -1,971 +0,0 @@ -import io -import os -import random -import re -import sys -import threading -import time -import warnings -import zlib -from abc import ABC, abstractmethod -from contextlib import contextmanager -from datetime import datetime, timezone -from functools import wraps, partial - -import sentry_sdk -from sentry_sdk.utils import ( - ContextVar, - now, - nanosecond_time, - to_timestamp, - serialize_frame, - json_dumps, -) -from sentry_sdk.envelope import Envelope, Item -from sentry_sdk.tracing import TransactionSource - -from typing import TYPE_CHECKING - -if TYPE_CHECKING: - from typing import Any - from typing import Callable - from typing import Dict - from typing import Generator - from typing import Iterable - from typing import List - from typing import Optional - from typing import Set - from typing import Tuple - from typing import Union - - from sentry_sdk._types import BucketKey - from sentry_sdk._types import DurationUnit - from sentry_sdk._types import FlushedMetricValue - from sentry_sdk._types import MeasurementUnit - from sentry_sdk._types import MetricMetaKey - from sentry_sdk._types import MetricTagValue - from sentry_sdk._types import MetricTags - from sentry_sdk._types import MetricTagsInternal - from sentry_sdk._types import MetricType - from sentry_sdk._types import MetricValue - - -warnings.warn( - "The sentry_sdk.metrics module is deprecated and will be removed in the next major release. " - "Sentry will reject all metrics sent after October 7, 2024. " - "Learn more: https://sentry.zendesk.com/hc/en-us/articles/26369339769883-Upcoming-API-Changes-to-Metrics", - DeprecationWarning, - stacklevel=2, -) - -_in_metrics = ContextVar("in_metrics", default=False) -_set = set # set is shadowed below - -GOOD_TRANSACTION_SOURCES = frozenset( - [ - TransactionSource.ROUTE, - TransactionSource.VIEW, - TransactionSource.COMPONENT, - TransactionSource.TASK, - ] -) - -_sanitize_unit = partial(re.compile(r"[^a-zA-Z0-9_]+").sub, "") -_sanitize_metric_key = partial(re.compile(r"[^a-zA-Z0-9_\-.]+").sub, "_") -_sanitize_tag_key = partial(re.compile(r"[^a-zA-Z0-9_\-.\/]+").sub, "") - - -def _sanitize_tag_value(value): - # type: (str) -> str - table = str.maketrans( - { - "\n": "\\n", - "\r": "\\r", - "\t": "\\t", - "\\": "\\\\", - "|": "\\u{7c}", - ",": "\\u{2c}", - } - ) - return value.translate(table) - - -def get_code_location(stacklevel): - # type: (int) -> Optional[Dict[str, Any]] - try: - frm = sys._getframe(stacklevel) - except Exception: - return None - - return serialize_frame( - frm, include_local_variables=False, include_source_context=True - ) - - -@contextmanager -def recursion_protection(): - # type: () -> Generator[bool, None, None] - """Enters recursion protection and returns the old flag.""" - old_in_metrics = _in_metrics.get() - _in_metrics.set(True) - try: - yield old_in_metrics - finally: - _in_metrics.set(old_in_metrics) - - -def metrics_noop(func): - # type: (Any) -> Any - """Convenient decorator that uses `recursion_protection` to - make a function a noop. - """ - - @wraps(func) - def new_func(*args, **kwargs): - # type: (*Any, **Any) -> Any - with recursion_protection() as in_metrics: - if not in_metrics: - return func(*args, **kwargs) - - return new_func - - -class Metric(ABC): - __slots__ = () - - @abstractmethod - def __init__(self, first): - # type: (MetricValue) -> None - pass - - @property - @abstractmethod - def weight(self): - # type: () -> int - pass - - @abstractmethod - def add(self, value): - # type: (MetricValue) -> None - pass - - @abstractmethod - def serialize_value(self): - # type: () -> Iterable[FlushedMetricValue] - pass - - -class CounterMetric(Metric): - __slots__ = ("value",) - - def __init__( - self, - first, # type: MetricValue - ): - # type: (...) -> None - self.value = float(first) - - @property - def weight(self): - # type: (...) -> int - return 1 - - def add( - self, - value, # type: MetricValue - ): - # type: (...) -> None - self.value += float(value) - - def serialize_value(self): - # type: (...) -> Iterable[FlushedMetricValue] - return (self.value,) - - -class GaugeMetric(Metric): - __slots__ = ( - "last", - "min", - "max", - "sum", - "count", - ) - - def __init__( - self, - first, # type: MetricValue - ): - # type: (...) -> None - first = float(first) - self.last = first - self.min = first - self.max = first - self.sum = first - self.count = 1 - - @property - def weight(self): - # type: (...) -> int - # Number of elements. - return 5 - - def add( - self, - value, # type: MetricValue - ): - # type: (...) -> None - value = float(value) - self.last = value - self.min = min(self.min, value) - self.max = max(self.max, value) - self.sum += value - self.count += 1 - - def serialize_value(self): - # type: (...) -> Iterable[FlushedMetricValue] - return ( - self.last, - self.min, - self.max, - self.sum, - self.count, - ) - - -class DistributionMetric(Metric): - __slots__ = ("value",) - - def __init__( - self, - first, # type: MetricValue - ): - # type(...) -> None - self.value = [float(first)] - - @property - def weight(self): - # type: (...) -> int - return len(self.value) - - def add( - self, - value, # type: MetricValue - ): - # type: (...) -> None - self.value.append(float(value)) - - def serialize_value(self): - # type: (...) -> Iterable[FlushedMetricValue] - return self.value - - -class SetMetric(Metric): - __slots__ = ("value",) - - def __init__( - self, - first, # type: MetricValue - ): - # type: (...) -> None - self.value = {first} - - @property - def weight(self): - # type: (...) -> int - return len(self.value) - - def add( - self, - value, # type: MetricValue - ): - # type: (...) -> None - self.value.add(value) - - def serialize_value(self): - # type: (...) -> Iterable[FlushedMetricValue] - def _hash(x): - # type: (MetricValue) -> int - if isinstance(x, str): - return zlib.crc32(x.encode("utf-8")) & 0xFFFFFFFF - return int(x) - - return (_hash(value) for value in self.value) - - -def _encode_metrics(flushable_buckets): - # type: (Iterable[Tuple[int, Dict[BucketKey, Metric]]]) -> bytes - out = io.BytesIO() - _write = out.write - - # Note on sanitization: we intentionally sanitize in emission (serialization) - # and not during aggregation for performance reasons. This means that the - # envelope can in fact have duplicate buckets stored. This is acceptable for - # relay side emission and should not happen commonly. - - for timestamp, buckets in flushable_buckets: - for bucket_key, metric in buckets.items(): - metric_type, metric_name, metric_unit, metric_tags = bucket_key - metric_name = _sanitize_metric_key(metric_name) - metric_unit = _sanitize_unit(metric_unit) - _write(metric_name.encode("utf-8")) - _write(b"@") - _write(metric_unit.encode("utf-8")) - - for serialized_value in metric.serialize_value(): - _write(b":") - _write(str(serialized_value).encode("utf-8")) - - _write(b"|") - _write(metric_type.encode("ascii")) - - if metric_tags: - _write(b"|#") - first = True - for tag_key, tag_value in metric_tags: - tag_key = _sanitize_tag_key(tag_key) - if not tag_key: - continue - if first: - first = False - else: - _write(b",") - _write(tag_key.encode("utf-8")) - _write(b":") - _write(_sanitize_tag_value(tag_value).encode("utf-8")) - - _write(b"|T") - _write(str(timestamp).encode("ascii")) - _write(b"\n") - - return out.getvalue() - - -def _encode_locations(timestamp, code_locations): - # type: (int, Iterable[Tuple[MetricMetaKey, Dict[str, Any]]]) -> bytes - mapping = {} # type: Dict[str, List[Any]] - - for key, loc in code_locations: - metric_type, name, unit = key - mri = "{}:{}@{}".format( - metric_type, _sanitize_metric_key(name), _sanitize_unit(unit) - ) - - loc["type"] = "location" - mapping.setdefault(mri, []).append(loc) - - return json_dumps({"timestamp": timestamp, "mapping": mapping}) - - -METRIC_TYPES = { - "c": CounterMetric, - "g": GaugeMetric, - "d": DistributionMetric, - "s": SetMetric, -} # type: dict[MetricType, type[Metric]] - -# some of these are dumb -TIMING_FUNCTIONS = { - "nanosecond": nanosecond_time, - "microsecond": lambda: nanosecond_time() / 1000.0, - "millisecond": lambda: nanosecond_time() / 1000000.0, - "second": now, - "minute": lambda: now() / 60.0, - "hour": lambda: now() / 3600.0, - "day": lambda: now() / 3600.0 / 24.0, - "week": lambda: now() / 3600.0 / 24.0 / 7.0, -} - - -class LocalAggregator: - __slots__ = ("_measurements",) - - def __init__(self): - # type: (...) -> None - self._measurements = {} # type: Dict[Tuple[str, MetricTagsInternal], Tuple[float, float, int, float]] - - def add( - self, - ty, # type: MetricType - key, # type: str - value, # type: float - unit, # type: MeasurementUnit - tags, # type: MetricTagsInternal - ): - # type: (...) -> None - export_key = "%s:%s@%s" % (ty, key, unit) - bucket_key = (export_key, tags) - - old = self._measurements.get(bucket_key) - if old is not None: - v_min, v_max, v_count, v_sum = old - v_min = min(v_min, value) - v_max = max(v_max, value) - v_count += 1 - v_sum += value - else: - v_min = v_max = v_sum = value - v_count = 1 - self._measurements[bucket_key] = (v_min, v_max, v_count, v_sum) - - def to_json(self): - # type: (...) -> Dict[str, Any] - rv = {} # type: Any - for (export_key, tags), ( - v_min, - v_max, - v_count, - v_sum, - ) in self._measurements.items(): - rv.setdefault(export_key, []).append( - { - "tags": _tags_to_dict(tags), - "min": v_min, - "max": v_max, - "count": v_count, - "sum": v_sum, - } - ) - return rv - - -class MetricsAggregator: - ROLLUP_IN_SECONDS = 10.0 - MAX_WEIGHT = 100000 - FLUSHER_SLEEP_TIME = 5.0 - - def __init__( - self, - capture_func, # type: Callable[[Envelope], None] - enable_code_locations=False, # type: bool - ): - # type: (...) -> None - self.buckets = {} # type: Dict[int, Any] - self._enable_code_locations = enable_code_locations - self._seen_locations = _set() # type: Set[Tuple[int, MetricMetaKey]] - self._pending_locations = {} # type: Dict[int, List[Tuple[MetricMetaKey, Any]]] - self._buckets_total_weight = 0 - self._capture_func = capture_func - self._running = True - self._lock = threading.Lock() - - self._flush_event = threading.Event() # type: threading.Event - self._force_flush = False - - # The aggregator shifts its flushing by up to an entire rollup window to - # avoid multiple clients trampling on end of a 10 second window as all the - # buckets are anchored to multiples of ROLLUP seconds. We randomize this - # number once per aggregator boot to achieve some level of offsetting - # across a fleet of deployed SDKs. Relay itself will also apply independent - # jittering. - self._flush_shift = random.random() * self.ROLLUP_IN_SECONDS - - self._flusher = None # type: Optional[threading.Thread] - self._flusher_pid = None # type: Optional[int] - - def _ensure_thread(self): - # type: (...) -> bool - """For forking processes we might need to restart this thread. - This ensures that our process actually has that thread running. - """ - if not self._running: - return False - - pid = os.getpid() - if self._flusher_pid == pid: - return True - - with self._lock: - # Recheck to make sure another thread didn't get here and start the - # the flusher in the meantime - if self._flusher_pid == pid: - return True - - self._flusher_pid = pid - - self._flusher = threading.Thread(target=self._flush_loop) - self._flusher.daemon = True - - try: - self._flusher.start() - except RuntimeError: - # Unfortunately at this point the interpreter is in a state that no - # longer allows us to spawn a thread and we have to bail. - self._running = False - return False - - return True - - def _flush_loop(self): - # type: (...) -> None - _in_metrics.set(True) - while self._running or self._force_flush: - if self._running: - self._flush_event.wait(self.FLUSHER_SLEEP_TIME) - self._flush() - - def _flush(self): - # type: (...) -> None - self._emit(self._flushable_buckets(), self._flushable_locations()) - - def _flushable_buckets(self): - # type: (...) -> (Iterable[Tuple[int, Dict[BucketKey, Metric]]]) - with self._lock: - force_flush = self._force_flush - cutoff = time.time() - self.ROLLUP_IN_SECONDS - self._flush_shift - flushable_buckets = () # type: Iterable[Tuple[int, Dict[BucketKey, Metric]]] - weight_to_remove = 0 - - if force_flush: - flushable_buckets = self.buckets.items() - self.buckets = {} - self._buckets_total_weight = 0 - self._force_flush = False - else: - flushable_buckets = [] - for buckets_timestamp, buckets in self.buckets.items(): - # If the timestamp of the bucket is newer that the rollup we want to skip it. - if buckets_timestamp <= cutoff: - flushable_buckets.append((buckets_timestamp, buckets)) - - # We will clear the elements while holding the lock, in order to avoid requesting it downstream again. - for buckets_timestamp, buckets in flushable_buckets: - for metric in buckets.values(): - weight_to_remove += metric.weight - del self.buckets[buckets_timestamp] - - self._buckets_total_weight -= weight_to_remove - - return flushable_buckets - - def _flushable_locations(self): - # type: (...) -> Dict[int, List[Tuple[MetricMetaKey, Dict[str, Any]]]] - with self._lock: - locations = self._pending_locations - self._pending_locations = {} - return locations - - @metrics_noop - def add( - self, - ty, # type: MetricType - key, # type: str - value, # type: MetricValue - unit, # type: MeasurementUnit - tags, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - local_aggregator=None, # type: Optional[LocalAggregator] - stacklevel=0, # type: Optional[int] - ): - # type: (...) -> None - if not self._ensure_thread() or self._flusher is None: - return None - - if timestamp is None: - timestamp = time.time() - elif isinstance(timestamp, datetime): - timestamp = to_timestamp(timestamp) - - bucket_timestamp = int( - (timestamp // self.ROLLUP_IN_SECONDS) * self.ROLLUP_IN_SECONDS - ) - serialized_tags = _serialize_tags(tags) - bucket_key = ( - ty, - key, - unit, - serialized_tags, - ) - - with self._lock: - local_buckets = self.buckets.setdefault(bucket_timestamp, {}) - metric = local_buckets.get(bucket_key) - if metric is not None: - previous_weight = metric.weight - metric.add(value) - else: - metric = local_buckets[bucket_key] = METRIC_TYPES[ty](value) - previous_weight = 0 - - added = metric.weight - previous_weight - - if stacklevel is not None: - self.record_code_location(ty, key, unit, stacklevel + 2, timestamp) - - # Given the new weight we consider whether we want to force flush. - self._consider_force_flush() - - # For sets, we only record that a value has been added to the set but not which one. - # See develop docs: https://develop.sentry.dev/sdk/metrics/#sets - if local_aggregator is not None: - local_value = float(added if ty == "s" else value) - local_aggregator.add(ty, key, local_value, unit, serialized_tags) - - def record_code_location( - self, - ty, # type: MetricType - key, # type: str - unit, # type: MeasurementUnit - stacklevel, # type: int - timestamp=None, # type: Optional[float] - ): - # type: (...) -> None - if not self._enable_code_locations: - return - if timestamp is None: - timestamp = time.time() - meta_key = (ty, key, unit) - start_of_day = datetime.fromtimestamp(timestamp, timezone.utc).replace( - hour=0, minute=0, second=0, microsecond=0, tzinfo=None - ) - start_of_day = int(to_timestamp(start_of_day)) - - if (start_of_day, meta_key) not in self._seen_locations: - self._seen_locations.add((start_of_day, meta_key)) - loc = get_code_location(stacklevel + 3) - if loc is not None: - # Group metadata by day to make flushing more efficient. - # There needs to be one envelope item per timestamp. - self._pending_locations.setdefault(start_of_day, []).append( - (meta_key, loc) - ) - - @metrics_noop - def need_code_location( - self, - ty, # type: MetricType - key, # type: str - unit, # type: MeasurementUnit - timestamp, # type: float - ): - # type: (...) -> bool - if self._enable_code_locations: - return False - meta_key = (ty, key, unit) - start_of_day = datetime.fromtimestamp(timestamp, timezone.utc).replace( - hour=0, minute=0, second=0, microsecond=0, tzinfo=None - ) - start_of_day = int(to_timestamp(start_of_day)) - return (start_of_day, meta_key) not in self._seen_locations - - def kill(self): - # type: (...) -> None - if self._flusher is None: - return - - self._running = False - self._flush_event.set() - self._flusher = None - - @metrics_noop - def flush(self): - # type: (...) -> None - self._force_flush = True - self._flush() - - def _consider_force_flush(self): - # type: (...) -> None - # It's important to acquire a lock around this method, since it will touch shared data structures. - total_weight = len(self.buckets) + self._buckets_total_weight - if total_weight >= self.MAX_WEIGHT: - self._force_flush = True - self._flush_event.set() - - def _emit( - self, - flushable_buckets, # type: (Iterable[Tuple[int, Dict[BucketKey, Metric]]]) - code_locations, # type: Dict[int, List[Tuple[MetricMetaKey, Dict[str, Any]]]] - ): - # type: (...) -> Optional[Envelope] - envelope = Envelope() - - if flushable_buckets: - encoded_metrics = _encode_metrics(flushable_buckets) - envelope.add_item(Item(payload=encoded_metrics, type="statsd")) - - for timestamp, locations in code_locations.items(): - encoded_locations = _encode_locations(timestamp, locations) - envelope.add_item(Item(payload=encoded_locations, type="metric_meta")) - - if envelope.items: - self._capture_func(envelope) - return envelope - return None - - -def _serialize_tags( - tags, # type: Optional[MetricTags] -): - # type: (...) -> MetricTagsInternal - if not tags: - return () - - rv = [] - for key, value in tags.items(): - # If the value is a collection, we want to flatten it. - if isinstance(value, (list, tuple)): - for inner_value in value: - if inner_value is not None: - rv.append((key, str(inner_value))) - elif value is not None: - rv.append((key, str(value))) - - # It's very important to sort the tags in order to obtain the - # same bucket key. - return tuple(sorted(rv)) - - -def _tags_to_dict(tags): - # type: (MetricTagsInternal) -> Dict[str, Any] - rv = {} # type: Dict[str, Any] - for tag_name, tag_value in tags: - old_value = rv.get(tag_name) - if old_value is not None: - if isinstance(old_value, list): - old_value.append(tag_value) - else: - rv[tag_name] = [old_value, tag_value] - else: - rv[tag_name] = tag_value - return rv - - -def _get_aggregator(): - # type: () -> Optional[MetricsAggregator] - client = sentry_sdk.get_client() - return ( - client.metrics_aggregator - if client.is_active() and client.metrics_aggregator is not None - else None - ) - - -def _get_aggregator_and_update_tags(key, value, unit, tags): - # type: (str, Optional[MetricValue], MeasurementUnit, Optional[MetricTags]) -> Tuple[Optional[MetricsAggregator], Optional[LocalAggregator], Optional[MetricTags]] - client = sentry_sdk.get_client() - if not client.is_active() or client.metrics_aggregator is None: - return None, None, tags - - updated_tags = dict(tags or ()) # type: Dict[str, MetricTagValue] - updated_tags.setdefault("release", client.options["release"]) - updated_tags.setdefault("environment", client.options["environment"]) - - scope = sentry_sdk.get_current_scope() - local_aggregator = None - - # We go with the low-level API here to access transaction information as - # this one is the same between just errors and errors + performance - transaction_source = scope._transaction_info.get("source") - if transaction_source in GOOD_TRANSACTION_SOURCES: - transaction_name = scope._transaction - if transaction_name: - updated_tags.setdefault("transaction", transaction_name) - if scope._span is not None: - local_aggregator = scope._span._get_local_aggregator() - - experiments = client.options.get("_experiments", {}) - before_emit_callback = experiments.get("before_emit_metric") - if before_emit_callback is not None: - with recursion_protection() as in_metrics: - if not in_metrics: - if not before_emit_callback(key, value, unit, updated_tags): - return None, None, updated_tags - - return client.metrics_aggregator, local_aggregator, updated_tags - - -def increment( - key, # type: str - value=1.0, # type: float - unit="none", # type: MeasurementUnit - tags=None, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - stacklevel=0, # type: int -): - # type: (...) -> None - """Increments a counter.""" - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - key, value, unit, tags - ) - if aggregator is not None: - aggregator.add( - "c", key, value, unit, tags, timestamp, local_aggregator, stacklevel - ) - - -# alias as incr is relatively common in python -incr = increment - - -class _Timing: - def __init__( - self, - key, # type: str - tags, # type: Optional[MetricTags] - timestamp, # type: Optional[Union[float, datetime]] - value, # type: Optional[float] - unit, # type: DurationUnit - stacklevel, # type: int - ): - # type: (...) -> None - self.key = key - self.tags = tags - self.timestamp = timestamp - self.value = value - self.unit = unit - self.entered = None # type: Optional[float] - self._span = None # type: Optional[sentry_sdk.tracing.Span] - self.stacklevel = stacklevel - - def _validate_invocation(self, context): - # type: (str) -> None - if self.value is not None: - raise TypeError( - "cannot use timing as %s when a value is provided" % context - ) - - def __enter__(self): - # type: (...) -> _Timing - self.entered = TIMING_FUNCTIONS[self.unit]() - self._validate_invocation("context-manager") - self._span = sentry_sdk.start_span(op="metric.timing", name=self.key) - if self.tags: - for key, value in self.tags.items(): - if isinstance(value, (tuple, list)): - value = ",".join(sorted(map(str, value))) - self._span.set_tag(key, value) - self._span.__enter__() - - # report code locations here for better accuracy - aggregator = _get_aggregator() - if aggregator is not None: - aggregator.record_code_location("d", self.key, self.unit, self.stacklevel) - - return self - - def __exit__(self, exc_type, exc_value, tb): - # type: (Any, Any, Any) -> None - assert self._span, "did not enter" - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - self.key, - self.value, - self.unit, - self.tags, - ) - if aggregator is not None: - elapsed = TIMING_FUNCTIONS[self.unit]() - self.entered # type: ignore - aggregator.add( - "d", - self.key, - elapsed, - self.unit, - tags, - self.timestamp, - local_aggregator, - None, # code locations are reported in __enter__ - ) - - self._span.__exit__(exc_type, exc_value, tb) - self._span = None - - def __call__(self, f): - # type: (Any) -> Any - self._validate_invocation("decorator") - - @wraps(f) - def timed_func(*args, **kwargs): - # type: (*Any, **Any) -> Any - with timing( - key=self.key, - tags=self.tags, - timestamp=self.timestamp, - unit=self.unit, - stacklevel=self.stacklevel + 1, - ): - return f(*args, **kwargs) - - return timed_func - - -def timing( - key, # type: str - value=None, # type: Optional[float] - unit="second", # type: DurationUnit - tags=None, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - stacklevel=0, # type: int -): - # type: (...) -> _Timing - """Emits a distribution with the time it takes to run the given code block. - - This method supports three forms of invocation: - - - when a `value` is provided, it functions similar to `distribution` but with - - it can be used as a context manager - - it can be used as a decorator - """ - if value is not None: - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - key, value, unit, tags - ) - if aggregator is not None: - aggregator.add( - "d", key, value, unit, tags, timestamp, local_aggregator, stacklevel - ) - return _Timing(key, tags, timestamp, value, unit, stacklevel) - - -def distribution( - key, # type: str - value, # type: float - unit="none", # type: MeasurementUnit - tags=None, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - stacklevel=0, # type: int -): - # type: (...) -> None - """Emits a distribution.""" - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - key, value, unit, tags - ) - if aggregator is not None: - aggregator.add( - "d", key, value, unit, tags, timestamp, local_aggregator, stacklevel - ) - - -def set( - key, # type: str - value, # type: Union[int, str] - unit="none", # type: MeasurementUnit - tags=None, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - stacklevel=0, # type: int -): - # type: (...) -> None - """Emits a set.""" - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - key, value, unit, tags - ) - if aggregator is not None: - aggregator.add( - "s", key, value, unit, tags, timestamp, local_aggregator, stacklevel - ) - - -def gauge( - key, # type: str - value, # type: float - unit="none", # type: MeasurementUnit - tags=None, # type: Optional[MetricTags] - timestamp=None, # type: Optional[Union[float, datetime]] - stacklevel=0, # type: int -): - # type: (...) -> None - """Emits a gauge.""" - aggregator, local_aggregator, tags = _get_aggregator_and_update_tags( - key, value, unit, tags - ) - if aggregator is not None: - aggregator.add( - "g", key, value, unit, tags, timestamp, local_aggregator, stacklevel - ) diff --git a/sentry_sdk/tracing.py b/sentry_sdk/tracing.py index 1697df1f22..0d652e490a 100644 --- a/sentry_sdk/tracing.py +++ b/sentry_sdk/tracing.py @@ -276,7 +276,6 @@ class Span: "hub", "_context_manager_state", "_containing_transaction", - "_local_aggregator", "scope", "origin", "name", @@ -345,7 +344,6 @@ def __init__( self.timestamp = None # type: Optional[datetime] self._span_recorder = None # type: Optional[_SpanRecorder] - self._local_aggregator = None # type: Optional[LocalAggregator] self.update_active_thread() self.set_profiler_id(get_profiler_id()) @@ -383,13 +381,6 @@ def span_id(self, value): # type: (str) -> None self._span_id = value - def _get_local_aggregator(self): - # type: (...) -> LocalAggregator - rv = self._local_aggregator - if rv is None: - rv = self._local_aggregator = LocalAggregator() - return rv - def __repr__(self): # type: () -> str return ( @@ -741,11 +732,6 @@ def to_json(self): if self.status: self._tags["status"] = self.status - if self._local_aggregator is not None: - metrics_summary = self._local_aggregator.to_json() - if metrics_summary: - rv["_metrics_summary"] = metrics_summary - if len(self._measurements) > 0: rv["measurements"] = self._measurements @@ -1122,13 +1108,6 @@ def finish( event["measurements"] = self._measurements - # This is here since `to_json` is not invoked. This really should - # be gone when we switch to onlyspans. - if self._local_aggregator is not None: - metrics_summary = self._local_aggregator.to_json() - if metrics_summary: - event["_metrics_summary"] = metrics_summary - return scope.capture_event(event) def set_measurement(self, name, value, unit=""): @@ -1505,8 +1484,3 @@ def calculate_interest_rate(amount, rate, years): has_tracing_enabled, maybe_create_breadcrumbs_from_span, ) - -with warnings.catch_warnings(): - # The code in this file which uses `LocalAggregator` is only called from the deprecated `metrics` module. - warnings.simplefilter("ignore", DeprecationWarning) - from sentry_sdk.metrics import LocalAggregator diff --git a/sentry_sdk/transport.py b/sentry_sdk/transport.py index 75384519e9..645bfead19 100644 --- a/sentry_sdk/transport.py +++ b/sentry_sdk/transport.py @@ -171,17 +171,7 @@ def _parse_rate_limits(header, now=None): retry_after = now + timedelta(seconds=int(retry_after_val)) for category in categories and categories.split(";") or (None,): - if category == "metric_bucket": - try: - namespaces = parameters[4].split(";") - except IndexError: - namespaces = [] - - if not namespaces or "custom" in namespaces: - yield category, retry_after # type: ignore - - else: - yield category, retry_after # type: ignore + yield category, retry_after # type: ignore except (LookupError, ValueError): continue @@ -417,12 +407,6 @@ def _check_disabled(self, category): # type: (str) -> bool def _disabled(bucket): # type: (Any) -> bool - - # The envelope item type used for metrics is statsd - # whereas the rate limit category is metric_bucket - if bucket == "statsd": - bucket = "metric_bucket" - ts = self._disabled_until.get(bucket) return ts is not None and ts > datetime.now(timezone.utc) diff --git a/tests/test_envelope.py b/tests/test_envelope.py index 06f8971dc3..d66cd9460a 100644 --- a/tests/test_envelope.py +++ b/tests/test_envelope.py @@ -252,7 +252,6 @@ def test_envelope_item_data_category_mapping(): ("client_report", "internal"), ("profile", "profile"), ("profile_chunk", "profile_chunk"), - ("statsd", "metric_bucket"), ("check_in", "monitor"), ("unknown_type", "default"), ] diff --git a/tests/test_metrics.py b/tests/test_metrics.py deleted file mode 100644 index c02f075288..0000000000 --- a/tests/test_metrics.py +++ /dev/null @@ -1,971 +0,0 @@ -import sys -import time -import linecache -from unittest import mock - -import pytest - -import sentry_sdk -from sentry_sdk import metrics -from sentry_sdk.tracing import TransactionSource -from sentry_sdk.envelope import parse_json - -try: - import gevent -except ImportError: - gevent = None - - -minimum_python_37_with_gevent = pytest.mark.skipif( - gevent and sys.version_info < (3, 7), - reason="Require Python 3.7 or higher with gevent", -) - - -def parse_metrics(bytes): - rv = [] - for line in bytes.splitlines(): - pieces = line.decode("utf-8").split("|") - payload = pieces[0].split(":") - name = payload[0] - values = payload[1:] - ty = pieces[1] - ts = None - tags = {} - for piece in pieces[2:]: - if piece[0] == "#": - for pair in piece[1:].split(","): - k, v = pair.split(":", 1) - old = tags.get(k) - if old is not None: - if isinstance(old, list): - old.append(v) - else: - tags[k] = [old, v] - else: - tags[k] = v - elif piece[0] == "T": - ts = int(piece[1:]) - else: - raise ValueError("unknown piece %r" % (piece,)) - rv.append((ts, name, ty, values, tags)) - rv.sort(key=lambda x: (x[0], x[1], tuple(sorted(tags.items())))) - return rv - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_increment(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.increment("foobar", 1.0, tags={"foo": "bar", "blub": "blah"}, timestamp=ts) - # python specific alias - metrics.incr("foobar", 2.0, tags={"foo": "bar", "blub": "blah"}, timestamp=ts) - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "foobar@none" - assert m[0][2] == "c" - assert m[0][3] == ["3.0"] - assert m[0][4] == { - "blub": "blah", - "foo": "bar", - "release": "fun-release", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - assert parse_json(meta_item.payload.get_bytes()) == { - "timestamp": mock.ANY, - "mapping": { - "c:foobar@none": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ] - }, - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_timing(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - ts = time.time() - envelopes = capture_envelopes() - - with metrics.timing("whatever", tags={"blub": "blah"}, timestamp=ts): - time.sleep(0.1) - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "whatever@second" - assert m[0][2] == "d" - assert len(m[0][3]) == 1 - assert float(m[0][3][0]) >= 0.1 - assert m[0][4] == { - "blub": "blah", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - json = parse_json(meta_item.payload.get_bytes()) - assert json == { - "timestamp": mock.ANY, - "mapping": { - "d:whatever@second": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ] - }, - } - - loc = json["mapping"]["d:whatever@second"][0] - line = linecache.getline(loc["abs_path"], loc["lineno"]) - assert ( - line.strip() - == 'with metrics.timing("whatever", tags={"blub": "blah"}, timestamp=ts):' - ) - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_timing_decorator( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - envelopes = capture_envelopes() - - @metrics.timing("whatever-1", tags={"x": "y"}) - def amazing(): - time.sleep(0.1) - return 42 - - @metrics.timing("whatever-2", tags={"x": "y"}, unit="nanosecond") - def amazing_nano(): - time.sleep(0.01) - return 23 - - assert amazing() == 42 - assert amazing_nano() == 23 - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 2 - assert m[0][1] == "whatever-1@second" - assert m[0][2] == "d" - assert len(m[0][3]) == 1 - assert float(m[0][3][0]) >= 0.1 - assert m[0][4] == { - "x": "y", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert m[1][1] == "whatever-2@nanosecond" - assert m[1][2] == "d" - assert len(m[1][3]) == 1 - assert float(m[1][3][0]) >= 10000000.0 - assert m[1][4] == { - "x": "y", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - json = parse_json(meta_item.payload.get_bytes()) - assert json == { - "timestamp": mock.ANY, - "mapping": { - "d:whatever-1@second": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ], - "d:whatever-2@nanosecond": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ], - }, - } - - # XXX: this is not the best location. It would probably be better to - # report the location in the function, however that is quite a bit - # tricker to do since we report from outside the function so we really - # only see the callsite. - loc = json["mapping"]["d:whatever-1@second"][0] - line = linecache.getline(loc["abs_path"], loc["lineno"]) - assert line.strip() == "assert amazing() == 42" - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_timing_basic(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.timing("timing", 1.0, tags={"a": "b"}, timestamp=ts) - metrics.timing("timing", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.timing("timing", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.timing("timing", 3.0, tags={"a": "b"}, timestamp=ts) - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "timing@second" - assert m[0][2] == "d" - assert len(m[0][3]) == 4 - assert sorted(map(float, m[0][3])) == [1.0, 2.0, 2.0, 3.0] - assert m[0][4] == { - "a": "b", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - assert parse_json(meta_item.payload.get_bytes()) == { - "timestamp": mock.ANY, - "mapping": { - "d:timing@second": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ] - }, - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_distribution(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.distribution("dist", 1.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 3.0, tags={"a": "b"}, timestamp=ts) - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "dist@none" - assert m[0][2] == "d" - assert len(m[0][3]) == 4 - assert sorted(map(float, m[0][3])) == [1.0, 2.0, 2.0, 3.0] - assert m[0][4] == { - "a": "b", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - json = parse_json(meta_item.payload.get_bytes()) - assert json == { - "timestamp": mock.ANY, - "mapping": { - "d:dist@none": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ] - }, - } - - loc = json["mapping"]["d:dist@none"][0] - line = linecache.getline(loc["abs_path"], loc["lineno"]) - assert ( - line.strip() - == 'metrics.distribution("dist", 1.0, tags={"a": "b"}, timestamp=ts)' - ) - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_set(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": True}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.set("my-set", "peter", tags={"magic": "puff"}, timestamp=ts) - metrics.set("my-set", "paul", tags={"magic": "puff"}, timestamp=ts) - metrics.set("my-set", "mary", tags={"magic": "puff"}, timestamp=ts) - sentry_sdk.flush() - - (envelope,) = envelopes - statsd_item, meta_item = envelope.items - - assert statsd_item.headers["type"] == "statsd" - m = parse_metrics(statsd_item.payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "my-set@none" - assert m[0][2] == "s" - assert len(m[0][3]) == 3 - assert sorted(map(int, m[0][3])) == [354582103, 2513273657, 3329318813] - assert m[0][4] == { - "magic": "puff", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert meta_item.headers["type"] == "metric_meta" - assert parse_json(meta_item.payload.get_bytes()) == { - "timestamp": mock.ANY, - "mapping": { - "s:my-set@none": [ - { - "type": "location", - "filename": "tests/test_metrics.py", - "abs_path": __file__, - "function": sys._getframe().f_code.co_name, - "module": __name__, - "lineno": mock.ANY, - "pre_context": mock.ANY, - "context_line": mock.ANY, - "post_context": mock.ANY, - } - ] - }, - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_gauge(sentry_init, capture_envelopes, maybe_monkeypatched_threading): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.gauge("my-gauge", 10.0, tags={"x": "y"}, timestamp=ts) - metrics.gauge("my-gauge", 20.0, tags={"x": "y"}, timestamp=ts) - metrics.gauge("my-gauge", 30.0, tags={"x": "y"}, timestamp=ts) - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "my-gauge@none" - assert m[0][2] == "g" - assert len(m[0][3]) == 5 - assert list(map(float, m[0][3])) == [30.0, 10.0, 30.0, 60.0, 3.0] - assert m[0][4] == { - "x": "y", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_multiple(sentry_init, capture_envelopes): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - ts = time.time() - envelopes = capture_envelopes() - - metrics.gauge("my-gauge", 10.0, tags={"x": "y"}, timestamp=ts) - metrics.gauge("my-gauge", 20.0, tags={"x": "y"}, timestamp=ts) - metrics.gauge("my-gauge", 30.0, tags={"x": "y"}, timestamp=ts) - for _ in range(10): - metrics.increment("counter-1", 1.0, timestamp=ts) - metrics.increment("counter-2", 1.0, timestamp=ts) - - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 3 - - assert m[0][1] == "counter-1@none" - assert m[0][2] == "c" - assert list(map(float, m[0][3])) == [10.0] - assert m[0][4] == { - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert m[1][1] == "counter-2@none" - assert m[1][2] == "c" - assert list(map(float, m[1][3])) == [1.0] - assert m[1][4] == { - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert m[2][1] == "my-gauge@none" - assert m[2][2] == "g" - assert len(m[2][3]) == 5 - assert list(map(float, m[2][3])) == [30.0, 10.0, 30.0, 60.0, 3.0] - assert m[2][4] == { - "x": "y", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_transaction_name( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - ts = time.time() - envelopes = capture_envelopes() - - sentry_sdk.get_current_scope().set_transaction_name( - "/user/{user_id}", source=TransactionSource.ROUTE - ) - metrics.distribution("dist", 1.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 2.0, tags={"a": "b"}, timestamp=ts) - metrics.distribution("dist", 3.0, tags={"a": "b"}, timestamp=ts) - - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "dist@none" - assert m[0][2] == "d" - assert len(m[0][3]) == 4 - assert sorted(map(float, m[0][3])) == [1.0, 2.0, 2.0, 3.0] - assert m[0][4] == { - "a": "b", - "transaction": "/user/{user_id}", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_metric_summaries( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - enable_tracing=True, - ) - ts = time.time() - envelopes = capture_envelopes() - - with sentry_sdk.start_transaction( - op="stuff", name="/foo", source=TransactionSource.ROUTE - ) as transaction: - metrics.increment("root-counter", timestamp=ts) - with metrics.timing("my-timer-metric", tags={"a": "b"}, timestamp=ts): - for x in range(10): - metrics.distribution("my-dist", float(x), timestamp=ts) - - sentry_sdk.flush() - - (transaction, envelope) = envelopes - - # Metrics Emission - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 3 - - assert m[0][1] == "my-dist@none" - assert m[0][2] == "d" - assert len(m[0][3]) == 10 - assert sorted(m[0][3]) == list(map(str, map(float, range(10)))) - assert m[0][4] == { - "transaction": "/foo", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert m[1][1] == "my-timer-metric@second" - assert m[1][2] == "d" - assert len(m[1][3]) == 1 - assert m[1][4] == { - "a": "b", - "transaction": "/foo", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - assert m[2][1] == "root-counter@none" - assert m[2][2] == "c" - assert m[2][3] == ["1.0"] - assert m[2][4] == { - "transaction": "/foo", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - } - - # Measurement Attachment - t = transaction.items[0].get_transaction_event() - - assert t["_metrics_summary"] == { - "c:root-counter@none": [ - { - "count": 1, - "min": 1.0, - "max": 1.0, - "sum": 1.0, - "tags": { - "transaction": "/foo", - "release": "fun-release@1.0.0", - "environment": "not-fun-env", - }, - } - ] - } - - assert t["spans"][0]["_metrics_summary"]["d:my-dist@none"] == [ - { - "count": 10, - "min": 0.0, - "max": 9.0, - "sum": 45.0, - "tags": { - "environment": "not-fun-env", - "release": "fun-release@1.0.0", - "transaction": "/foo", - }, - } - ] - - assert t["spans"][0]["tags"] == {"a": "b"} - (timer,) = t["spans"][0]["_metrics_summary"]["d:my-timer-metric@second"] - assert timer["count"] == 1 - assert timer["max"] == timer["min"] == timer["sum"] - assert timer["sum"] > 0 - assert timer["tags"] == { - "a": "b", - "environment": "not-fun-env", - "release": "fun-release@1.0.0", - "transaction": "/foo", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -@pytest.mark.parametrize( - "metric_name,metric_unit,expected_name", - [ - ("first-metric", "nano-second", "first-metric@nanosecond"), - ("another_metric?", "nano second", "another_metric_@nanosecond"), - ( - "metric", - "nanosecond", - "metric@nanosecond", - ), - ( - "my.amaze.metric I guess", - "nano|\nsecond", - "my.amaze.metric_I_guess@nanosecond", - ), - ("métríc", "nanöseconď", "m_tr_c@nansecon"), - ], -) -def test_metric_name_normalization( - sentry_init, - capture_envelopes, - metric_name, - metric_unit, - expected_name, - maybe_monkeypatched_threading, -): - sentry_init( - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - envelopes = capture_envelopes() - - metrics.distribution(metric_name, 1.0, unit=metric_unit) - - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - - parsed_metrics = parse_metrics(envelope.items[0].payload.get_bytes()) - assert len(parsed_metrics) == 1 - - name = parsed_metrics[0][1] - assert name == expected_name - - -@minimum_python_37_with_gevent -@pytest.mark.forked -@pytest.mark.parametrize( - "metric_tag,expected_tag", - [ - ({"f-oo|bar": "%$foo/"}, {"f-oobar": "%$foo/"}), - ({"foo$.$.$bar": "blah{}"}, {"foo..bar": "blah{}"}), - ( - {"foö-bar": "snöwmän"}, - {"fo-bar": "snöwmän"}, - ), - ({"route": "GET /foo"}, {"route": "GET /foo"}), - ({"__bar__": "this | or , that"}, {"__bar__": "this \\u{7c} or \\u{2c} that"}), - ({"foo/": "hello!\n\r\t\\"}, {"foo/": "hello!\\n\\r\\t\\\\"}), - ], -) -def test_metric_tag_normalization( - sentry_init, - capture_envelopes, - metric_tag, - expected_tag, - maybe_monkeypatched_threading, -): - sentry_init( - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - envelopes = capture_envelopes() - - metrics.distribution("a", 1.0, tags=metric_tag) - - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - - parsed_metrics = parse_metrics(envelope.items[0].payload.get_bytes()) - assert len(parsed_metrics) == 1 - - tags = parsed_metrics[0][4] - - expected_tag_key, expected_tag_value = expected_tag.popitem() - assert expected_tag_key in tags - assert tags[expected_tag_key] == expected_tag_value - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_before_emit_metric( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - def before_emit(key, value, unit, tags): - if key == "removed-metric" or value == 47 or unit == "unsupported": - return False - - tags["extra"] = "foo" - del tags["release"] - # this better be a noop! - metrics.increment("shitty-recursion") - return True - - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={ - "enable_metrics": True, - "metric_code_locations": False, - "before_emit_metric": before_emit, - }, - ) - envelopes = capture_envelopes() - - metrics.increment("removed-metric", 1.0) - metrics.increment("another-removed-metric", 47) - metrics.increment("yet-another-removed-metric", 1.0, unit="unsupported") - metrics.increment("actual-metric", 1.0) - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 1 - assert m[0][1] == "actual-metric@none" - assert m[0][3] == ["1.0"] - assert m[0][4] == { - "extra": "foo", - "environment": "not-fun-env", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_aggregator_flush( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release@1.0.0", - environment="not-fun-env", - _experiments={ - "enable_metrics": True, - }, - ) - envelopes = capture_envelopes() - - metrics.increment("a-metric", 1.0) - sentry_sdk.flush() - - assert len(envelopes) == 1 - assert sentry_sdk.get_client().metrics_aggregator.buckets == {} - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_tag_serialization( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release", - environment="not-fun-env", - _experiments={"enable_metrics": True, "metric_code_locations": False}, - ) - envelopes = capture_envelopes() - - metrics.increment( - "counter", - tags={ - "no-value": None, - "an-int": 42, - "a-float": 23.0, - "a-string": "blah", - "more-than-one": [1, "zwei", "3.0", None], - }, - ) - sentry_sdk.flush() - - (envelope,) = envelopes - - assert len(envelope.items) == 1 - assert envelope.items[0].headers["type"] == "statsd" - m = parse_metrics(envelope.items[0].payload.get_bytes()) - - assert len(m) == 1 - assert m[0][4] == { - "an-int": "42", - "a-float": "23.0", - "a-string": "blah", - "more-than-one": ["1", "3.0", "zwei"], - "release": "fun-release", - "environment": "not-fun-env", - } - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_flush_recursion_protection( - sentry_init, capture_envelopes, monkeypatch, maybe_monkeypatched_threading -): - sentry_init( - release="fun-release", - environment="not-fun-env", - _experiments={"enable_metrics": True}, - ) - envelopes = capture_envelopes() - test_client = sentry_sdk.get_client() - - real_capture_envelope = test_client.transport.capture_envelope - - def bad_capture_envelope(*args, **kwargs): - metrics.increment("bad-metric") - return real_capture_envelope(*args, **kwargs) - - monkeypatch.setattr(test_client.transport, "capture_envelope", bad_capture_envelope) - - metrics.increment("counter") - - # flush twice to see the inner metric - sentry_sdk.flush() - sentry_sdk.flush() - - (envelope,) = envelopes - m = parse_metrics(envelope.items[0].payload.get_bytes()) - assert len(m) == 1 - assert m[0][1] == "counter@none" - - -@minimum_python_37_with_gevent -@pytest.mark.forked -def test_flush_recursion_protection_background_flush( - sentry_init, capture_envelopes, monkeypatch, maybe_monkeypatched_threading -): - monkeypatch.setattr(metrics.MetricsAggregator, "FLUSHER_SLEEP_TIME", 0.01) - sentry_init( - release="fun-release", - environment="not-fun-env", - _experiments={"enable_metrics": True}, - ) - envelopes = capture_envelopes() - test_client = sentry_sdk.get_client() - - real_capture_envelope = test_client.transport.capture_envelope - - def bad_capture_envelope(*args, **kwargs): - metrics.increment("bad-metric") - return real_capture_envelope(*args, **kwargs) - - monkeypatch.setattr(test_client.transport, "capture_envelope", bad_capture_envelope) - - metrics.increment("counter") - - # flush via sleep and flag - sentry_sdk.get_client().metrics_aggregator._force_flush = True - time.sleep(0.5) - - (envelope,) = envelopes - m = parse_metrics(envelope.items[0].payload.get_bytes()) - assert len(m) == 1 - assert m[0][1] == "counter@none" - - -@pytest.mark.skipif( - not gevent or sys.version_info >= (3, 7), - reason="Python 3.6 or lower and gevent required", -) -@pytest.mark.forked -def test_disable_metrics_for_old_python_with_gevent( - sentry_init, capture_envelopes, maybe_monkeypatched_threading -): - if maybe_monkeypatched_threading != "greenlet": - pytest.skip("Test specifically for gevent/greenlet") - - sentry_init( - release="fun-release", - environment="not-fun-env", - _experiments={"enable_metrics": True}, - ) - envelopes = capture_envelopes() - - metrics.incr("counter") - - sentry_sdk.flush() - - assert sentry_sdk.get_client().metrics_aggregator is None - assert not envelopes diff --git a/tests/test_transport.py b/tests/test_transport.py index 68669fa24d..804105b010 100644 --- a/tests/test_transport.py +++ b/tests/test_transport.py @@ -590,43 +590,6 @@ def test_complex_limits_without_data_category( assert len(capturing_server.captured) == 0 -@pytest.mark.parametrize("response_code", [200, 429]) -def test_metric_bucket_limits(capturing_server, response_code, make_client): - client = make_client() - capturing_server.respond_with( - code=response_code, - headers={ - "X-Sentry-Rate-Limits": "4711:metric_bucket:organization:quota_exceeded:custom" - }, - ) - - envelope = Envelope() - envelope.add_item(Item(payload=b"{}", type="statsd")) - client.transport.capture_envelope(envelope) - client.flush() - - assert len(capturing_server.captured) == 1 - assert capturing_server.captured[0].path == "/api/132/envelope/" - capturing_server.clear_captured() - - assert set(client.transport._disabled_until) == {"metric_bucket"} - - client.transport.capture_envelope(envelope) - client.capture_event({"type": "transaction"}) - client.flush() - - assert len(capturing_server.captured) == 2 - - envelope = capturing_server.captured[0].envelope - assert envelope.items[0].type == "transaction" - envelope = capturing_server.captured[1].envelope - assert envelope.items[0].type == "client_report" - report = parse_json(envelope.items[0].get_bytes()) - assert report["discarded_events"] == [ - {"category": "metric_bucket", "reason": "ratelimit_backoff", "quantity": 1}, - ] - - @pytest.mark.parametrize("response_code", [200, 429]) def test_log_item_limits(capturing_server, response_code, make_client): client = make_client() @@ -664,80 +627,6 @@ def test_log_item_limits(capturing_server, response_code, make_client): ] -@pytest.mark.parametrize("response_code", [200, 429]) -def test_metric_bucket_limits_with_namespace( - capturing_server, response_code, make_client -): - client = make_client() - capturing_server.respond_with( - code=response_code, - headers={ - "X-Sentry-Rate-Limits": "4711:metric_bucket:organization:quota_exceeded:foo" - }, - ) - - envelope = Envelope() - envelope.add_item(Item(payload=b"{}", type="statsd")) - client.transport.capture_envelope(envelope) - client.flush() - - assert len(capturing_server.captured) == 1 - assert capturing_server.captured[0].path == "/api/132/envelope/" - capturing_server.clear_captured() - - assert set(client.transport._disabled_until) == set([]) - - client.transport.capture_envelope(envelope) - client.capture_event({"type": "transaction"}) - client.flush() - - assert len(capturing_server.captured) == 2 - - envelope = capturing_server.captured[0].envelope - assert envelope.items[0].type == "statsd" - envelope = capturing_server.captured[1].envelope - assert envelope.items[0].type == "transaction" - - -@pytest.mark.parametrize("response_code", [200, 429]) -def test_metric_bucket_limits_with_all_namespaces( - capturing_server, response_code, make_client -): - client = make_client() - capturing_server.respond_with( - code=response_code, - headers={ - "X-Sentry-Rate-Limits": "4711:metric_bucket:organization:quota_exceeded" - }, - ) - - envelope = Envelope() - envelope.add_item(Item(payload=b"{}", type="statsd")) - client.transport.capture_envelope(envelope) - client.flush() - - assert len(capturing_server.captured) == 1 - assert capturing_server.captured[0].path == "/api/132/envelope/" - capturing_server.clear_captured() - - assert set(client.transport._disabled_until) == set(["metric_bucket"]) - - client.transport.capture_envelope(envelope) - client.capture_event({"type": "transaction"}) - client.flush() - - assert len(capturing_server.captured) == 2 - - envelope = capturing_server.captured[0].envelope - assert envelope.items[0].type == "transaction" - envelope = capturing_server.captured[1].envelope - assert envelope.items[0].type == "client_report" - report = parse_json(envelope.items[0].get_bytes()) - assert report["discarded_events"] == [ - {"category": "metric_bucket", "reason": "ratelimit_backoff", "quantity": 1}, - ] - - def test_hub_cls_backwards_compat(): class TestCustomHubClass(Hub): pass From a04974744e54c333733648152b530d19c457205f Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 9 Oct 2025 13:34:35 +0200 Subject: [PATCH 532/591] ref: Remove "experimental" from log func name (#4901) ### Description Logs are not experimental anymore, but one of the internal log-related functions still had "experimental" in the name. #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- sentry_sdk/client.py | 4 ++-- sentry_sdk/integrations/logging.py | 2 +- sentry_sdk/integrations/loguru.py | 2 +- sentry_sdk/logger.py | 2 +- 4 files changed, 5 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 59a5013783..9401c3f0b0 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -213,7 +213,7 @@ def capture_event(self, *args, **kwargs): # type: (*Any, **Any) -> Optional[str] return None - def _capture_experimental_log(self, log): + def _capture_log(self, log): # type: (Log) -> None pass @@ -877,7 +877,7 @@ def capture_event( return return_value - def _capture_experimental_log(self, log): + def _capture_log(self, log): # type: (Optional[Log]) -> None if not has_logs_enabled(self.options) or log is None: return diff --git a/sentry_sdk/integrations/logging.py b/sentry_sdk/integrations/logging.py index bfb30fc67b..7e16943b28 100644 --- a/sentry_sdk/integrations/logging.py +++ b/sentry_sdk/integrations/logging.py @@ -409,7 +409,7 @@ def _capture_log_from_record(self, client, record): attrs["logger.name"] = record.name # noinspection PyProtectedMember - client._capture_experimental_log( + client._capture_log( { "severity_text": otel_severity_text, "severity_number": otel_severity_number, diff --git a/sentry_sdk/integrations/loguru.py b/sentry_sdk/integrations/loguru.py index b910b9a407..2c0279d0ce 100644 --- a/sentry_sdk/integrations/loguru.py +++ b/sentry_sdk/integrations/loguru.py @@ -193,7 +193,7 @@ def loguru_sentry_logs_handler(message): if record.get("name"): attrs["logger.name"] = record["name"] - client._capture_experimental_log( + client._capture_log( { "severity_text": otel_severity_text, "severity_number": otel_severity_number, diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index bc98f35155..0ea7218e01 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -46,7 +46,7 @@ def _capture_log(severity_text, severity_number, template, **kwargs): } # noinspection PyProtectedMember - client._capture_experimental_log( + client._capture_log( { "severity_text": severity_text, "severity_number": severity_number, From 1f8c008e9368874d6cea289283da61e236062b34 Mon Sep 17 00:00:00 2001 From: Kev <6111995+k-fish@users.noreply.github.com> Date: Thu, 9 Oct 2025 09:58:52 -0400 Subject: [PATCH 533/591] feat(metrics): Add trace metrics behind an experiments flag (#4898) ### Summary Similar to https://github.com/getsentry/sentry-javascript/pull/17883, this allows the py sdk to send in new trace metric protocol items, although this code is experimental since the schema may still change. Most of this code has been copied from logs (eg. log batcher -> metrics batcher) in order to dogfood, once we're more sure of our approach we can refactor. Closes LOGS-367 --------- Co-authored-by: Ivana Kellyer --- sentry_sdk/_metrics.py | 81 +++++++++++++ sentry_sdk/_metrics_batcher.py | 156 +++++++++++++++++++++++++ sentry_sdk/_types.py | 27 +++++ sentry_sdk/client.py | 80 ++++++++++++- sentry_sdk/consts.py | 3 + sentry_sdk/envelope.py | 2 + sentry_sdk/types.py | 3 + sentry_sdk/utils.py | 18 ++- tests/test_metrics.py | 208 +++++++++++++++++++++++++++++++++ 9 files changed, 576 insertions(+), 2 deletions(-) create mode 100644 sentry_sdk/_metrics.py create mode 100644 sentry_sdk/_metrics_batcher.py create mode 100644 tests/test_metrics.py diff --git a/sentry_sdk/_metrics.py b/sentry_sdk/_metrics.py new file mode 100644 index 0000000000..03bde137bd --- /dev/null +++ b/sentry_sdk/_metrics.py @@ -0,0 +1,81 @@ +""" +NOTE: This file contains experimental code that may be changed or removed at any +time without prior notice. +""" + +import time +from typing import Any, Optional, TYPE_CHECKING, Union + +import sentry_sdk +from sentry_sdk.utils import safe_repr + +if TYPE_CHECKING: + from sentry_sdk._types import Metric, MetricType + + +def _capture_metric( + name, # type: str + metric_type, # type: MetricType + value, # type: float + unit=None, # type: Optional[str] + attributes=None, # type: Optional[dict[str, Any]] +): + # type: (...) -> None + client = sentry_sdk.get_client() + + attrs = {} # type: dict[str, Union[str, bool, float, int]] + if attributes: + for k, v in attributes.items(): + attrs[k] = ( + v + if ( + isinstance(v, str) + or isinstance(v, int) + or isinstance(v, bool) + or isinstance(v, float) + ) + else safe_repr(v) + ) + + metric = { + "timestamp": time.time(), + "trace_id": None, + "span_id": None, + "name": name, + "type": metric_type, + "value": float(value), + "unit": unit, + "attributes": attrs, + } # type: Metric + + client._capture_metric(metric) + + +def count( + name, # type: str + value, # type: float + unit=None, # type: Optional[str] + attributes=None, # type: Optional[dict[str, Any]] +): + # type: (...) -> None + _capture_metric(name, "counter", value, unit, attributes) + + +def gauge( + name, # type: str + value, # type: float + unit=None, # type: Optional[str] + attributes=None, # type: Optional[dict[str, Any]] +): + # type: (...) -> None + _capture_metric(name, "gauge", value, unit, attributes) + + +def distribution( + name, # type: str + value, # type: float + unit=None, # type: Optional[str] + attributes=None, # type: Optional[dict[str, Any]] +): + # type: (...) -> None + _capture_metric(name, "distribution", value, unit, attributes) diff --git a/sentry_sdk/_metrics_batcher.py b/sentry_sdk/_metrics_batcher.py new file mode 100644 index 0000000000..fd9a5d732b --- /dev/null +++ b/sentry_sdk/_metrics_batcher.py @@ -0,0 +1,156 @@ +import os +import random +import threading +from datetime import datetime, timezone +from typing import Optional, List, Callable, TYPE_CHECKING, Any, Union + +from sentry_sdk.utils import format_timestamp, safe_repr +from sentry_sdk.envelope import Envelope, Item, PayloadRef + +if TYPE_CHECKING: + from sentry_sdk._types import Metric + + +class MetricsBatcher: + MAX_METRICS_BEFORE_FLUSH = 100 + FLUSH_WAIT_TIME = 5.0 + + def __init__( + self, + capture_func, # type: Callable[[Envelope], None] + ): + # type: (...) -> None + self._metric_buffer = [] # type: List[Metric] + self._capture_func = capture_func + self._running = True + self._lock = threading.Lock() + + self._flush_event = threading.Event() # type: threading.Event + + self._flusher = None # type: Optional[threading.Thread] + self._flusher_pid = None # type: Optional[int] + + def _ensure_thread(self): + # type: (...) -> bool + if not self._running: + return False + + pid = os.getpid() + if self._flusher_pid == pid: + return True + + with self._lock: + if self._flusher_pid == pid: + return True + + self._flusher_pid = pid + + self._flusher = threading.Thread(target=self._flush_loop) + self._flusher.daemon = True + + try: + self._flusher.start() + except RuntimeError: + self._running = False + return False + + return True + + def _flush_loop(self): + # type: (...) -> None + while self._running: + self._flush_event.wait(self.FLUSH_WAIT_TIME + random.random()) + self._flush_event.clear() + self._flush() + + def add( + self, + metric, # type: Metric + ): + # type: (...) -> None + if not self._ensure_thread() or self._flusher is None: + return None + + with self._lock: + self._metric_buffer.append(metric) + if len(self._metric_buffer) >= self.MAX_METRICS_BEFORE_FLUSH: + self._flush_event.set() + + def kill(self): + # type: (...) -> None + if self._flusher is None: + return + + self._running = False + self._flush_event.set() + self._flusher = None + + def flush(self): + # type: (...) -> None + self._flush() + + @staticmethod + def _metric_to_transport_format(metric): + # type: (Metric) -> Any + def format_attribute(val): + # type: (Union[int, float, str, bool]) -> Any + if isinstance(val, bool): + return {"value": val, "type": "boolean"} + if isinstance(val, int): + return {"value": val, "type": "integer"} + if isinstance(val, float): + return {"value": val, "type": "double"} + if isinstance(val, str): + return {"value": val, "type": "string"} + return {"value": safe_repr(val), "type": "string"} + + res = { + "timestamp": metric["timestamp"], + "trace_id": metric["trace_id"], + "name": metric["name"], + "type": metric["type"], + "value": metric["value"], + "attributes": { + k: format_attribute(v) for (k, v) in metric["attributes"].items() + }, + } + + if metric.get("span_id") is not None: + res["span_id"] = metric["span_id"] + + if metric.get("unit") is not None: + res["unit"] = metric["unit"] + + return res + + def _flush(self): + # type: (...) -> Optional[Envelope] + + envelope = Envelope( + headers={"sent_at": format_timestamp(datetime.now(timezone.utc))} + ) + with self._lock: + if len(self._metric_buffer) == 0: + return None + + envelope.add_item( + Item( + type="trace_metric", + content_type="application/vnd.sentry.items.trace-metric+json", + headers={ + "item_count": len(self._metric_buffer), + }, + payload=PayloadRef( + json={ + "items": [ + self._metric_to_transport_format(metric) + for metric in self._metric_buffer + ] + } + ), + ) + ) + self._metric_buffer.clear() + + self._capture_func(envelope) + return envelope diff --git a/sentry_sdk/_types.py b/sentry_sdk/_types.py index d057f215e4..66ed7df4f7 100644 --- a/sentry_sdk/_types.py +++ b/sentry_sdk/_types.py @@ -234,6 +234,32 @@ class SDKInfo(TypedDict): }, ) + MetricType = Literal["counter", "gauge", "distribution"] + + MetricAttributeValue = TypedDict( + "MetricAttributeValue", + { + "value": Union[str, bool, float, int], + "type": Literal["string", "boolean", "double", "integer"], + }, + ) + + Metric = TypedDict( + "Metric", + { + "timestamp": float, + "trace_id": Optional[str], + "span_id": Optional[str], + "name": str, + "type": MetricType, + "value": float, + "unit": Optional[str], + "attributes": dict[str, str | bool | float | int], + }, + ) + + MetricProcessor = Callable[[Metric, Hint], Optional[Metric]] + # TODO: Make a proper type definition for this (PRs welcome!) Breadcrumb = Dict[str, Any] @@ -268,6 +294,7 @@ class SDKInfo(TypedDict): "monitor", "span", "log_item", + "trace_metric", ] SessionStatus = Literal["ok", "exited", "crashed", "abnormal"] diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 9401c3f0b0..d17f922642 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -24,7 +24,9 @@ is_gevent, logger, get_before_send_log, + get_before_send_metric, has_logs_enabled, + has_metrics_enabled, ) from sentry_sdk.serializer import serialize from sentry_sdk.tracing import trace @@ -59,13 +61,14 @@ from typing import Union from typing import TypeVar - from sentry_sdk._types import Event, Hint, SDKInfo, Log + from sentry_sdk._types import Event, Hint, SDKInfo, Log, Metric from sentry_sdk.integrations import Integration from sentry_sdk.scope import Scope from sentry_sdk.session import Session from sentry_sdk.spotlight import SpotlightClient from sentry_sdk.transport import Transport from sentry_sdk._log_batcher import LogBatcher + from sentry_sdk._metrics_batcher import MetricsBatcher I = TypeVar("I", bound=Integration) # noqa: E741 @@ -182,6 +185,7 @@ def __init__(self, options=None): self.transport = None # type: Optional[Transport] self.monitor = None # type: Optional[Monitor] self.log_batcher = None # type: Optional[LogBatcher] + self.metrics_batcher = None # type: Optional[MetricsBatcher] def __getstate__(self, *args, **kwargs): # type: (*Any, **Any) -> Any @@ -217,6 +221,10 @@ def _capture_log(self, log): # type: (Log) -> None pass + def _capture_metric(self, metric): + # type: (Metric) -> None + pass + def capture_session(self, *args, **kwargs): # type: (*Any, **Any) -> None return None @@ -366,6 +374,13 @@ def _capture_envelope(envelope): self.log_batcher = LogBatcher(capture_func=_capture_envelope) + self.metrics_batcher = None + + if has_metrics_enabled(self.options): + from sentry_sdk._metrics_batcher import MetricsBatcher + + self.metrics_batcher = MetricsBatcher(capture_func=_capture_envelope) + max_request_body_size = ("always", "never", "small", "medium") if self.options["max_request_body_size"] not in max_request_body_size: raise ValueError( @@ -944,6 +959,65 @@ def _capture_log(self, log): if self.log_batcher: self.log_batcher.add(log) + def _capture_metric(self, metric): + # type: (Optional[Metric]) -> None + if not has_metrics_enabled(self.options) or metric is None: + return + + isolation_scope = sentry_sdk.get_isolation_scope() + + metric["attributes"]["sentry.sdk.name"] = SDK_INFO["name"] + metric["attributes"]["sentry.sdk.version"] = SDK_INFO["version"] + + environment = self.options.get("environment") + if environment is not None and "sentry.environment" not in metric["attributes"]: + metric["attributes"]["sentry.environment"] = environment + + release = self.options.get("release") + if release is not None and "sentry.release" not in metric["attributes"]: + metric["attributes"]["sentry.release"] = release + + span = sentry_sdk.get_current_span() + metric["trace_id"] = "00000000-0000-0000-0000-000000000000" + + if span: + metric["trace_id"] = span.trace_id + metric["span_id"] = span.span_id + else: + propagation_context = isolation_scope.get_active_propagation_context() + if propagation_context and propagation_context.trace_id: + metric["trace_id"] = propagation_context.trace_id + + if isolation_scope._user is not None: + for metric_attribute, user_attribute in ( + ("user.id", "id"), + ("user.name", "username"), + ("user.email", "email"), + ): + if ( + user_attribute in isolation_scope._user + and metric_attribute not in metric["attributes"] + ): + metric["attributes"][metric_attribute] = isolation_scope._user[ + user_attribute + ] + + debug = self.options.get("debug", False) + if debug: + logger.debug( + f"[Sentry Metrics] [{metric.get('type')}] {metric.get('name')}: {metric.get('value')}" + ) + + before_send_metric = get_before_send_metric(self.options) + if before_send_metric is not None: + metric = before_send_metric(metric, {}) + + if metric is None: + return + + if self.metrics_batcher: + self.metrics_batcher.add(metric) + def capture_session( self, session, # type: Session @@ -998,6 +1072,8 @@ def close( self.session_flusher.kill() if self.log_batcher is not None: self.log_batcher.kill() + if self.metrics_batcher is not None: + self.metrics_batcher.kill() if self.monitor: self.monitor.kill() self.transport.kill() @@ -1022,6 +1098,8 @@ def flush( self.session_flusher.flush() if self.log_batcher is not None: self.log_batcher.flush() + if self.metrics_batcher is not None: + self.metrics_batcher.flush() self.transport.flush(timeout=timeout, callback=callback) def __enter__(self): diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 0f71a0d460..12654cc76d 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -52,6 +52,7 @@ class CompressionAlgo(Enum): Hint, Log, MeasurementUnit, + Metric, ProfilerMode, TracesSampler, TransactionProcessor, @@ -77,6 +78,8 @@ class CompressionAlgo(Enum): "transport_http2": Optional[bool], "enable_logs": Optional[bool], "before_send_log": Optional[Callable[[Log, Hint], Optional[Log]]], + "enable_metrics": Optional[bool], + "before_send_metric": Optional[Callable[[Metric, Hint], Optional[Metric]]], }, total=False, ) diff --git a/sentry_sdk/envelope.py b/sentry_sdk/envelope.py index b26c458d41..56bb5fde73 100644 --- a/sentry_sdk/envelope.py +++ b/sentry_sdk/envelope.py @@ -285,6 +285,8 @@ def data_category(self): return "error" elif ty == "log": return "log_item" + elif ty == "trace_metric": + return "trace_metric" elif ty == "client_report": return "internal" elif ty == "profile": diff --git a/sentry_sdk/types.py b/sentry_sdk/types.py index 1a65247584..8b28166462 100644 --- a/sentry_sdk/types.py +++ b/sentry_sdk/types.py @@ -21,6 +21,7 @@ Log, MonitorConfig, SamplingContext, + Metric, ) else: from typing import Any @@ -35,6 +36,7 @@ Log = Any MonitorConfig = Any SamplingContext = Any + Metric = Any __all__ = ( @@ -46,4 +48,5 @@ "Log", "MonitorConfig", "SamplingContext", + "Metric", ) diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 2083fd296c..cd825b29e2 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -59,7 +59,7 @@ from gevent.hub import Hub - from sentry_sdk._types import Event, ExcInfo, Log, Hint + from sentry_sdk._types import Event, ExcInfo, Log, Hint, Metric P = ParamSpec("P") R = TypeVar("R") @@ -2013,3 +2013,19 @@ def get_before_send_log(options): return options.get("before_send_log") or options["_experiments"].get( "before_send_log" ) + + +def has_metrics_enabled(options): + # type: (Optional[dict[str, Any]]) -> bool + if options is None: + return False + + return bool(options["_experiments"].get("enable_metrics", False)) + + +def get_before_send_metric(options): + # type: (Optional[dict[str, Any]]) -> Optional[Callable[[Metric, Hint], Optional[Metric]]] + if options is None: + return None + + return options["_experiments"].get("before_send_metric") diff --git a/tests/test_metrics.py b/tests/test_metrics.py new file mode 100644 index 0000000000..5e774227fd --- /dev/null +++ b/tests/test_metrics.py @@ -0,0 +1,208 @@ +import json +import sys +from typing import List, Any, Mapping +import pytest + +import sentry_sdk +from sentry_sdk import _metrics +from sentry_sdk import get_client +from sentry_sdk.envelope import Envelope +from sentry_sdk.types import Metric + + +def envelopes_to_metrics(envelopes): + # type: (List[Envelope]) -> List[Metric] + res = [] # type: List[Metric] + for envelope in envelopes: + for item in envelope.items: + if item.type == "trace_metric": + for metric_json in item.payload.json["items"]: + metric = { + "timestamp": metric_json["timestamp"], + "trace_id": metric_json["trace_id"], + "span_id": metric_json.get("span_id"), + "name": metric_json["name"], + "type": metric_json["type"], + "value": metric_json["value"], + "unit": metric_json.get("unit"), + "attributes": { + k: v["value"] + for (k, v) in metric_json["attributes"].items() + }, + } # type: Metric + res.append(metric) + return res + + +def test_metrics_disabled_by_default(sentry_init, capture_envelopes): + sentry_init() + + envelopes = capture_envelopes() + + _metrics.count("test.counter", 1) + _metrics.gauge("test.gauge", 42) + _metrics.distribution("test.distribution", 200) + + assert len(envelopes) == 0 + + +def test_metrics_basics(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_metrics": True}) + envelopes = capture_envelopes() + + _metrics.count("test.counter", 1) + _metrics.gauge("test.gauge", 42, unit="millisecond") + _metrics.distribution("test.distribution", 200, unit="second") + + get_client().flush() + metrics = envelopes_to_metrics(envelopes) + + assert len(metrics) == 3 + + assert metrics[0]["name"] == "test.counter" + assert metrics[0]["type"] == "counter" + assert metrics[0]["value"] == 1.0 + assert metrics[0]["unit"] is None + assert "sentry.sdk.name" in metrics[0]["attributes"] + assert "sentry.sdk.version" in metrics[0]["attributes"] + + assert metrics[1]["name"] == "test.gauge" + assert metrics[1]["type"] == "gauge" + assert metrics[1]["value"] == 42.0 + assert metrics[1]["unit"] == "millisecond" + + assert metrics[2]["name"] == "test.distribution" + assert metrics[2]["type"] == "distribution" + assert metrics[2]["value"] == 200.0 + assert metrics[2]["unit"] == "second" + + +def test_metrics_experimental_option(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_metrics": True}) + envelopes = capture_envelopes() + + _metrics.count("test.counter", 5) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + + assert metrics[0]["name"] == "test.counter" + assert metrics[0]["type"] == "counter" + assert metrics[0]["value"] == 5.0 + + +def test_metrics_with_attributes(sentry_init, capture_envelopes): + sentry_init( + _experiments={"enable_metrics": True}, release="1.0.0", environment="test" + ) + envelopes = capture_envelopes() + + _metrics.count( + "test.counter", 1, attributes={"endpoint": "/api/test", "status": "success"} + ) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + + assert metrics[0]["attributes"]["endpoint"] == "/api/test" + assert metrics[0]["attributes"]["status"] == "success" + assert metrics[0]["attributes"]["sentry.release"] == "1.0.0" + assert metrics[0]["attributes"]["sentry.environment"] == "test" + + +def test_metrics_with_user(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_metrics": True}) + envelopes = capture_envelopes() + + sentry_sdk.set_user( + {"id": "user-123", "email": "test@example.com", "username": "testuser"} + ) + _metrics.count("test.user.counter", 1) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + + assert metrics[0]["attributes"]["user.id"] == "user-123" + assert metrics[0]["attributes"]["user.email"] == "test@example.com" + assert metrics[0]["attributes"]["user.name"] == "testuser" + + +def test_metrics_with_span(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_metrics": True}, traces_sample_rate=1.0) + envelopes = capture_envelopes() + + with sentry_sdk.start_transaction(op="test", name="test-span"): + _metrics.count("test.span.counter", 1) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + + assert metrics[0]["trace_id"] is not None + assert metrics[0]["trace_id"] != "00000000-0000-0000-0000-000000000000" + assert metrics[0]["span_id"] is not None + + +def test_metrics_tracing_without_performance(sentry_init, capture_envelopes): + sentry_init(_experiments={"enable_metrics": True}) + envelopes = capture_envelopes() + + _metrics.count("test.span.counter", 1) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + + assert metrics[0]["trace_id"] is not None + assert metrics[0]["trace_id"] != "00000000-0000-0000-0000-000000000000" + assert metrics[0]["span_id"] is None + + +def test_metrics_before_send(sentry_init, capture_envelopes): + before_metric_called = False + + def _before_metric(record, hint): + nonlocal before_metric_called + + assert set(record.keys()) == { + "timestamp", + "trace_id", + "span_id", + "name", + "type", + "value", + "unit", + "attributes", + } + + if record["name"] == "test.skip": + return None + + before_metric_called = True + return record + + sentry_init( + _experiments={ + "enable_metrics": True, + "before_send_metric": _before_metric, + }, + ) + envelopes = capture_envelopes() + + _metrics.count("test.skip", 1) + _metrics.count("test.keep", 1) + + get_client().flush() + + metrics = envelopes_to_metrics(envelopes) + assert len(metrics) == 1 + assert metrics[0]["name"] == "test.keep" + assert before_metric_called From 272af1b1380db38fbf70087e561555c558bf8308 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Thu, 9 Oct 2025 14:00:37 +0000 Subject: [PATCH 534/591] release: 2.41.0 --- CHANGELOG.md | 12 ++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 15 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 0bcd623611..63efefd543 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,17 @@ # Changelog +## 2.41.0 + +### Various fixes & improvements + +- feat(metrics): Add trace metrics behind an experiments flag (#4898) by @k-fish +- ref: Remove "experimental" from log func name (#4901) by @sentrivana +- chore: Remove old metrics code (#4899) by @sentrivana +- ci: Bump Python version for linting (#4897) by @sentrivana +- feat: Add concurrent.futures patch to threading integration (#4770) by @alexander-alderman-webb +- fix(ai): add mapping for gen_ai message roles (#4884) by @shellmayr +- ci: Remove toxgen check (#4892) by @sentrivana + ## 2.40.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 2f630c382b..b3522a913e 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.40.0" +release = "2.41.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 12654cc76d..2158276f9b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1339,4 +1339,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.40.0" +VERSION = "2.41.0" diff --git a/setup.py b/setup.py index fbb8694e5e..274e343be7 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.40.0", + version="2.41.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From 685287d36da8f9423d2cfb03dc7c13bb7d6c57dd Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 9 Oct 2025 16:05:08 +0200 Subject: [PATCH 535/591] Update CHANGELOG.md --- CHANGELOG.md | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 63efefd543..8f59545d6b 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,13 +4,15 @@ ### Various fixes & improvements -- feat(metrics): Add trace metrics behind an experiments flag (#4898) by @k-fish -- ref: Remove "experimental" from log func name (#4901) by @sentrivana +- feat: Add `concurrent.futures` patch to threading integration (#4770) by @alexander-alderman-webb + + The SDK now makes sure to automatically preserve span relationships when using `ThreadPoolExecutor`. - chore: Remove old metrics code (#4899) by @sentrivana -- ci: Bump Python version for linting (#4897) by @sentrivana -- feat: Add concurrent.futures patch to threading integration (#4770) by @alexander-alderman-webb -- fix(ai): add mapping for gen_ai message roles (#4884) by @shellmayr -- ci: Remove toxgen check (#4892) by @sentrivana + + Removed all code related to the deprecated experimental metrics feature (`sentry_sdk.metrics`). +- ref: Remove "experimental" from log function name (#4901) by @sentrivana +- fix(ai): Add mapping for gen_ai message roles (#4884) by @shellmayr +- feat(metrics): Add trace metrics behind an experiments flag (#4898) by @k-fish ## 2.40.0 From 149a7daac8c24c6901c8ac876d0e19cb77ba3ec6 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Vjeran=20Grozdani=C4=87?= Date: Fri, 10 Oct 2025 12:09:08 +0200 Subject: [PATCH 536/591] feat(ai): Add `python-genai` integration (#4891) Adds support for `python-genai` integrations. It supports both sync and async clients, and both regular and streaming modes for interacting with models and building agents. Closes [PY-1733: Add agent monitoring support for `google-genai`](https://linear.app/getsentry/issue/PY-1733/add-agent-monitoring-support-for-google-genai) --- .github/workflows/test-integrations-ai.yml | 4 + pyproject.toml | 4 + scripts/populate_tox/config.py | 7 + scripts/populate_tox/releases.jsonl | 16 +- .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/__init__.py | 1 + .../integrations/google_genai/__init__.py | 298 ++++++ .../integrations/google_genai/consts.py | 16 + .../integrations/google_genai/streaming.py | 155 +++ sentry_sdk/integrations/google_genai/utils.py | 566 +++++++++++ setup.py | 1 + tests/integrations/google_genai/__init__.py | 4 + .../google_genai/test_google_genai.py | 907 ++++++++++++++++++ tox.ini | 36 +- 14 files changed, 1998 insertions(+), 18 deletions(-) create mode 100644 sentry_sdk/integrations/google_genai/__init__.py create mode 100644 sentry_sdk/integrations/google_genai/consts.py create mode 100644 sentry_sdk/integrations/google_genai/streaming.py create mode 100644 sentry_sdk/integrations/google_genai/utils.py create mode 100644 tests/integrations/google_genai/__init__.py create mode 100644 tests/integrations/google_genai/test_google_genai.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index fcbb464078..65cc636cd9 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -82,6 +82,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-langgraph" + - name: Test google_genai + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-google_genai" - name: Test openai_agents run: | set -x # print commands that are executed diff --git a/pyproject.toml b/pyproject.toml index 5b86531014..4441660c50 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -118,6 +118,10 @@ ignore_missing_imports = true module = "langgraph.*" ignore_missing_imports = true +[[tool.mypy.overrides]] +module = "google.genai.*" +ignore_missing_imports = true + [[tool.mypy.overrides]] module = "executing.*" ignore_missing_imports = true diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index f6b90e75e6..85988ac3cf 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -142,6 +142,13 @@ "package": "gql[all]", "num_versions": 2, }, + "google_genai": { + "package": "google-genai", + "deps": { + "*": ["pytest-asyncio"], + }, + "python": ">=3.9", + }, "graphene": { "package": "graphene", "deps": { diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 9f937e5e77..5ee83b65bc 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.46", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.48", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -66,9 +66,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.33.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.37.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.42.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} @@ -95,7 +99,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} @@ -128,7 +132,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.5.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.6.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.3", "yanked": false}} {"info": {"classifiers": ["Framework :: Pylons", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": null, "version": "1.0.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.10.8", "yanked": false}} {"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.6.5", "yanked": false}} @@ -155,7 +159,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}} @@ -191,7 +195,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.2", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 81f887ad4f..abfa1b63cc 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -78,6 +78,7 @@ "openai-base", "openai-notiktoken", "langgraph", + "google_genai", "openai_agents", "huggingface_hub", ], diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 3f71f0f4ba..9e279b8345 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -140,6 +140,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "flask": (1, 1, 4), "gql": (3, 4, 1), "graphene": (3, 3), + "google_genai": (1, 29, 0), # google-genai "grpc": (1, 32, 0), # grpcio "httpx": (0, 16, 0), "huggingface_hub": (0, 24, 7), diff --git a/sentry_sdk/integrations/google_genai/__init__.py b/sentry_sdk/integrations/google_genai/__init__.py new file mode 100644 index 0000000000..7175b64340 --- /dev/null +++ b/sentry_sdk/integrations/google_genai/__init__.py @@ -0,0 +1,298 @@ +from functools import wraps +from typing import ( + Any, + AsyncIterator, + Callable, + Iterator, + List, +) + +import sentry_sdk +from sentry_sdk.ai.utils import get_start_span_function +from sentry_sdk.integrations import DidNotEnable, Integration +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.tracing import SPANSTATUS + + +try: + from google.genai.models import Models, AsyncModels +except ImportError: + raise DidNotEnable("google-genai not installed") + + +from .consts import IDENTIFIER, ORIGIN, GEN_AI_SYSTEM +from .utils import ( + set_span_data_for_request, + set_span_data_for_response, + _capture_exception, + prepare_generate_content_args, +) +from .streaming import ( + set_span_data_for_streaming_response, + accumulate_streaming_response, +) + + +class GoogleGenAIIntegration(Integration): + identifier = IDENTIFIER + origin = ORIGIN + + def __init__(self, include_prompts=True): + # type: (GoogleGenAIIntegration, bool) -> None + self.include_prompts = include_prompts + + @staticmethod + def setup_once(): + # type: () -> None + # Patch sync methods + Models.generate_content = _wrap_generate_content(Models.generate_content) + Models.generate_content_stream = _wrap_generate_content_stream( + Models.generate_content_stream + ) + + # Patch async methods + AsyncModels.generate_content = _wrap_async_generate_content( + AsyncModels.generate_content + ) + AsyncModels.generate_content_stream = _wrap_async_generate_content_stream( + AsyncModels.generate_content_stream + ) + + +def _wrap_generate_content_stream(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + @wraps(f) + def new_generate_content_stream(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(GoogleGenAIIntegration) + if integration is None: + return f(self, *args, **kwargs) + + _model, contents, model_name = prepare_generate_content_args(args, kwargs) + + span = get_start_span_function()( + op=OP.GEN_AI_INVOKE_AGENT, + name="invoke_agent", + origin=ORIGIN, + ) + span.__enter__() + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + set_span_data_for_request(span, integration, model_name, contents, kwargs) + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + chat_span = sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + name=f"chat {model_name}", + origin=ORIGIN, + ) + chat_span.__enter__() + chat_span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + chat_span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) + chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + set_span_data_for_request(chat_span, integration, model_name, contents, kwargs) + chat_span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + try: + stream = f(self, *args, **kwargs) + + # Create wrapper iterator to accumulate responses + def new_iterator(): + # type: () -> Iterator[Any] + chunks = [] # type: List[Any] + try: + for chunk in stream: + chunks.append(chunk) + yield chunk + except Exception as exc: + _capture_exception(exc) + chat_span.set_status(SPANSTATUS.ERROR) + raise + finally: + # Accumulate all chunks and set final response data on spans + if chunks: + accumulated_response = accumulate_streaming_response(chunks) + set_span_data_for_streaming_response( + chat_span, integration, accumulated_response + ) + set_span_data_for_streaming_response( + span, integration, accumulated_response + ) + chat_span.__exit__(None, None, None) + span.__exit__(None, None, None) + + return new_iterator() + + except Exception as exc: + _capture_exception(exc) + chat_span.__exit__(None, None, None) + span.__exit__(None, None, None) + raise + + return new_generate_content_stream + + +def _wrap_async_generate_content_stream(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + @wraps(f) + async def new_async_generate_content_stream(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(GoogleGenAIIntegration) + if integration is None: + return await f(self, *args, **kwargs) + + _model, contents, model_name = prepare_generate_content_args(args, kwargs) + + span = get_start_span_function()( + op=OP.GEN_AI_INVOKE_AGENT, + name="invoke_agent", + origin=ORIGIN, + ) + span.__enter__() + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + set_span_data_for_request(span, integration, model_name, contents, kwargs) + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + chat_span = sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + name=f"chat {model_name}", + origin=ORIGIN, + ) + chat_span.__enter__() + chat_span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + chat_span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) + chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + set_span_data_for_request(chat_span, integration, model_name, contents, kwargs) + chat_span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + try: + stream = await f(self, *args, **kwargs) + + # Create wrapper async iterator to accumulate responses + async def new_async_iterator(): + # type: () -> AsyncIterator[Any] + chunks = [] # type: List[Any] + try: + async for chunk in stream: + chunks.append(chunk) + yield chunk + except Exception as exc: + _capture_exception(exc) + chat_span.set_status(SPANSTATUS.ERROR) + raise + finally: + # Accumulate all chunks and set final response data on spans + if chunks: + accumulated_response = accumulate_streaming_response(chunks) + set_span_data_for_streaming_response( + chat_span, integration, accumulated_response + ) + set_span_data_for_streaming_response( + span, integration, accumulated_response + ) + chat_span.__exit__(None, None, None) + span.__exit__(None, None, None) + + return new_async_iterator() + + except Exception as exc: + _capture_exception(exc) + chat_span.__exit__(None, None, None) + span.__exit__(None, None, None) + raise + + return new_async_generate_content_stream + + +def _wrap_generate_content(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + @wraps(f) + def new_generate_content(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(GoogleGenAIIntegration) + if integration is None: + return f(self, *args, **kwargs) + + model, contents, model_name = prepare_generate_content_args(args, kwargs) + + with get_start_span_function()( + op=OP.GEN_AI_INVOKE_AGENT, + name="invoke_agent", + origin=ORIGIN, + ) as span: + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + set_span_data_for_request(span, integration, model_name, contents, kwargs) + + with sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + name=f"chat {model_name}", + origin=ORIGIN, + ) as chat_span: + chat_span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + chat_span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) + chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + set_span_data_for_request( + chat_span, integration, model_name, contents, kwargs + ) + + try: + response = f(self, *args, **kwargs) + except Exception as exc: + _capture_exception(exc) + chat_span.set_status(SPANSTATUS.ERROR) + raise + + set_span_data_for_response(chat_span, integration, response) + set_span_data_for_response(span, integration, response) + + return response + + return new_generate_content + + +def _wrap_async_generate_content(f): + # type: (Callable[..., Any]) -> Callable[..., Any] + @wraps(f) + async def new_async_generate_content(self, *args, **kwargs): + # type: (Any, Any, Any) -> Any + integration = sentry_sdk.get_client().get_integration(GoogleGenAIIntegration) + if integration is None: + return await f(self, *args, **kwargs) + + model, contents, model_name = prepare_generate_content_args(args, kwargs) + + with get_start_span_function()( + op=OP.GEN_AI_INVOKE_AGENT, + name="invoke_agent", + origin=ORIGIN, + ) as span: + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + set_span_data_for_request(span, integration, model_name, contents, kwargs) + + with sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + name=f"chat {model_name}", + origin=ORIGIN, + ) as chat_span: + chat_span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + chat_span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) + chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + set_span_data_for_request( + chat_span, integration, model_name, contents, kwargs + ) + try: + response = await f(self, *args, **kwargs) + except Exception as exc: + _capture_exception(exc) + chat_span.set_status(SPANSTATUS.ERROR) + raise + + set_span_data_for_response(chat_span, integration, response) + set_span_data_for_response(span, integration, response) + + return response + + return new_async_generate_content diff --git a/sentry_sdk/integrations/google_genai/consts.py b/sentry_sdk/integrations/google_genai/consts.py new file mode 100644 index 0000000000..5b53ebf0e2 --- /dev/null +++ b/sentry_sdk/integrations/google_genai/consts.py @@ -0,0 +1,16 @@ +GEN_AI_SYSTEM = "gcp.gemini" + +# Mapping of tool attributes to their descriptions +# These are all tools that are available in the Google GenAI API +TOOL_ATTRIBUTES_MAP = { + "google_search_retrieval": "Google Search retrieval tool", + "google_search": "Google Search tool", + "retrieval": "Retrieval tool", + "enterprise_web_search": "Enterprise web search tool", + "google_maps": "Google Maps tool", + "code_execution": "Code execution tool", + "computer_use": "Computer use tool", +} + +IDENTIFIER = "google_genai" +ORIGIN = f"auto.ai.{IDENTIFIER}" diff --git a/sentry_sdk/integrations/google_genai/streaming.py b/sentry_sdk/integrations/google_genai/streaming.py new file mode 100644 index 0000000000..03d09aadf6 --- /dev/null +++ b/sentry_sdk/integrations/google_genai/streaming.py @@ -0,0 +1,155 @@ +from typing import ( + TYPE_CHECKING, + Any, + List, + TypedDict, + Optional, +) + +from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import SPANDATA +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import ( + safe_serialize, +) +from .utils import ( + extract_tool_calls, + extract_finish_reasons, + extract_contents_text, + extract_usage_data, + UsageData, +) + +if TYPE_CHECKING: + from sentry_sdk.tracing import Span + from google.genai.types import GenerateContentResponse + + +class AccumulatedResponse(TypedDict): + id: Optional[str] + model: Optional[str] + text: str + finish_reasons: List[str] + tool_calls: List[dict[str, Any]] + usage_metadata: UsageData + + +def accumulate_streaming_response(chunks): + # type: (List[GenerateContentResponse]) -> AccumulatedResponse + """Accumulate streaming chunks into a single response-like object.""" + accumulated_text = [] + finish_reasons = [] + tool_calls = [] + total_input_tokens = 0 + total_output_tokens = 0 + total_tokens = 0 + total_cached_tokens = 0 + total_reasoning_tokens = 0 + response_id = None + model = None + + for chunk in chunks: + # Extract text and tool calls + if getattr(chunk, "candidates", None): + for candidate in getattr(chunk, "candidates", []): + if hasattr(candidate, "content") and getattr( + candidate.content, "parts", [] + ): + extracted_text = extract_contents_text(candidate.content) + if extracted_text: + accumulated_text.append(extracted_text) + + extracted_finish_reasons = extract_finish_reasons(chunk) + if extracted_finish_reasons: + finish_reasons.extend(extracted_finish_reasons) + + extracted_tool_calls = extract_tool_calls(chunk) + if extracted_tool_calls: + tool_calls.extend(extracted_tool_calls) + + # Accumulate token usage + extracted_usage_data = extract_usage_data(chunk) + total_input_tokens += extracted_usage_data["input_tokens"] + total_output_tokens += extracted_usage_data["output_tokens"] + total_cached_tokens += extracted_usage_data["input_tokens_cached"] + total_reasoning_tokens += extracted_usage_data["output_tokens_reasoning"] + total_tokens += extracted_usage_data["total_tokens"] + + accumulated_response = AccumulatedResponse( + text="".join(accumulated_text), + finish_reasons=finish_reasons, + tool_calls=tool_calls, + usage_metadata=UsageData( + input_tokens=total_input_tokens, + output_tokens=total_output_tokens, + input_tokens_cached=total_cached_tokens, + output_tokens_reasoning=total_reasoning_tokens, + total_tokens=total_tokens, + ), + id=response_id, + model=model, + ) + + return accumulated_response + + +def set_span_data_for_streaming_response(span, integration, accumulated_response): + # type: (Span, Any, AccumulatedResponse) -> None + """Set span data for accumulated streaming response.""" + if ( + should_send_default_pii() + and integration.include_prompts + and accumulated_response.get("text") + ): + span.set_data( + SPANDATA.GEN_AI_RESPONSE_TEXT, + safe_serialize([accumulated_response["text"]]), + ) + + if accumulated_response.get("finish_reasons"): + set_data_normalized( + span, + SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, + accumulated_response["finish_reasons"], + ) + + if accumulated_response.get("tool_calls"): + span.set_data( + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, + safe_serialize(accumulated_response["tool_calls"]), + ) + + if accumulated_response.get("id"): + span.set_data(SPANDATA.GEN_AI_RESPONSE_ID, accumulated_response["id"]) + if accumulated_response.get("model"): + span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, accumulated_response["model"]) + + if accumulated_response["usage_metadata"]["input_tokens"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, + accumulated_response["usage_metadata"]["input_tokens"], + ) + + if accumulated_response["usage_metadata"]["input_tokens_cached"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED, + accumulated_response["usage_metadata"]["input_tokens_cached"], + ) + + if accumulated_response["usage_metadata"]["output_tokens"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, + accumulated_response["usage_metadata"]["output_tokens"], + ) + + if accumulated_response["usage_metadata"]["output_tokens_reasoning"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING, + accumulated_response["usage_metadata"]["output_tokens_reasoning"], + ) + + if accumulated_response["usage_metadata"]["total_tokens"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, + accumulated_response["usage_metadata"]["total_tokens"], + ) diff --git a/sentry_sdk/integrations/google_genai/utils.py b/sentry_sdk/integrations/google_genai/utils.py new file mode 100644 index 0000000000..ff973b02d9 --- /dev/null +++ b/sentry_sdk/integrations/google_genai/utils.py @@ -0,0 +1,566 @@ +import copy +import inspect +from functools import wraps +from .consts import ORIGIN, TOOL_ATTRIBUTES_MAP, GEN_AI_SYSTEM +from typing import ( + cast, + TYPE_CHECKING, + Iterable, + Any, + Callable, + List, + Optional, + Union, + TypedDict, +) + +import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import ( + capture_internal_exceptions, + event_from_exception, + safe_serialize, +) +from google.genai.types import GenerateContentConfig + +if TYPE_CHECKING: + from sentry_sdk.tracing import Span + from google.genai.types import ( + GenerateContentResponse, + ContentListUnion, + Tool, + Model, + ) + + +class UsageData(TypedDict): + """Structure for token usage data.""" + + input_tokens: int + input_tokens_cached: int + output_tokens: int + output_tokens_reasoning: int + total_tokens: int + + +def extract_usage_data(response): + # type: (Union[GenerateContentResponse, dict[str, Any]]) -> UsageData + """Extract usage data from response into a structured format. + + Args: + response: The GenerateContentResponse object or dictionary containing usage metadata + + Returns: + UsageData: Dictionary with input_tokens, input_tokens_cached, + output_tokens, and output_tokens_reasoning fields + """ + usage_data = UsageData( + input_tokens=0, + input_tokens_cached=0, + output_tokens=0, + output_tokens_reasoning=0, + total_tokens=0, + ) + + # Handle dictionary response (from streaming) + if isinstance(response, dict): + usage = response.get("usage_metadata", {}) + if not usage: + return usage_data + + prompt_tokens = usage.get("prompt_token_count", 0) or 0 + tool_use_prompt_tokens = usage.get("tool_use_prompt_token_count", 0) or 0 + usage_data["input_tokens"] = prompt_tokens + tool_use_prompt_tokens + + cached_tokens = usage.get("cached_content_token_count", 0) or 0 + usage_data["input_tokens_cached"] = cached_tokens + + reasoning_tokens = usage.get("thoughts_token_count", 0) or 0 + usage_data["output_tokens_reasoning"] = reasoning_tokens + + candidates_tokens = usage.get("candidates_token_count", 0) or 0 + # python-genai reports output and reasoning tokens separately + # reasoning should be sub-category of output tokens + usage_data["output_tokens"] = candidates_tokens + reasoning_tokens + + total_tokens = usage.get("total_token_count", 0) or 0 + usage_data["total_tokens"] = total_tokens + + return usage_data + + if not hasattr(response, "usage_metadata"): + return usage_data + + usage = response.usage_metadata + + # Input tokens include both prompt and tool use prompt tokens + prompt_tokens = getattr(usage, "prompt_token_count", 0) or 0 + tool_use_prompt_tokens = getattr(usage, "tool_use_prompt_token_count", 0) or 0 + usage_data["input_tokens"] = prompt_tokens + tool_use_prompt_tokens + + # Cached input tokens + cached_tokens = getattr(usage, "cached_content_token_count", 0) or 0 + usage_data["input_tokens_cached"] = cached_tokens + + # Reasoning tokens + reasoning_tokens = getattr(usage, "thoughts_token_count", 0) or 0 + usage_data["output_tokens_reasoning"] = reasoning_tokens + + # output_tokens = candidates_tokens + reasoning_tokens + # google-genai reports output and reasoning tokens separately + candidates_tokens = getattr(usage, "candidates_token_count", 0) or 0 + usage_data["output_tokens"] = candidates_tokens + reasoning_tokens + + total_tokens = getattr(usage, "total_token_count", 0) or 0 + usage_data["total_tokens"] = total_tokens + + return usage_data + + +def _capture_exception(exc): + # type: (Any) -> None + """Capture exception with Google GenAI mechanism.""" + event, hint = event_from_exception( + exc, + client_options=sentry_sdk.get_client().options, + mechanism={"type": "google_genai", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + +def get_model_name(model): + # type: (Union[str, Model]) -> str + """Extract model name from model parameter.""" + if isinstance(model, str): + return model + # Handle case where model might be an object with a name attribute + if hasattr(model, "name"): + return str(model.name) + return str(model) + + +def extract_contents_text(contents): + # type: (ContentListUnion) -> Optional[str] + """Extract text from contents parameter which can have various formats.""" + if contents is None: + return None + + # Simple string case + if isinstance(contents, str): + return contents + + # List of contents or parts + if isinstance(contents, list): + texts = [] + for item in contents: + # Recursively extract text from each item + extracted = extract_contents_text(item) + if extracted: + texts.append(extracted) + return " ".join(texts) if texts else None + + # Dictionary case + if isinstance(contents, dict): + if "text" in contents: + return contents["text"] + # Try to extract from parts if present in dict + if "parts" in contents: + return extract_contents_text(contents["parts"]) + + # Content object with parts - recurse into parts + if getattr(contents, "parts", None): + return extract_contents_text(contents.parts) + + # Direct text attribute + if hasattr(contents, "text"): + return contents.text + + return None + + +def _format_tools_for_span(tools): + # type: (Iterable[Tool | Callable[..., Any]]) -> Optional[List[dict[str, Any]]] + """Format tools parameter for span data.""" + formatted_tools = [] + for tool in tools: + if callable(tool): + # Handle callable functions passed directly + formatted_tools.append( + { + "name": getattr(tool, "__name__", "unknown"), + "description": getattr(tool, "__doc__", None), + } + ) + elif ( + hasattr(tool, "function_declarations") + and tool.function_declarations is not None + ): + # Tool object with function declarations + for func_decl in tool.function_declarations: + formatted_tools.append( + { + "name": getattr(func_decl, "name", None), + "description": getattr(func_decl, "description", None), + } + ) + else: + # Check for predefined tool attributes - each of these tools + # is an attribute of the tool object, by default set to None + for attr_name, description in TOOL_ATTRIBUTES_MAP.items(): + if getattr(tool, attr_name, None): + formatted_tools.append( + { + "name": attr_name, + "description": description, + } + ) + break + + return formatted_tools if formatted_tools else None + + +def extract_tool_calls(response): + # type: (GenerateContentResponse) -> Optional[List[dict[str, Any]]] + """Extract tool/function calls from response candidates and automatic function calling history.""" + + tool_calls = [] + + # Extract from candidates, sometimes tool calls are nested under the content.parts object + if getattr(response, "candidates", []): + for candidate in response.candidates: + if not hasattr(candidate, "content") or not getattr( + candidate.content, "parts", [] + ): + continue + + for part in candidate.content.parts: + if getattr(part, "function_call", None): + function_call = part.function_call + tool_call = { + "name": getattr(function_call, "name", None), + "type": "function_call", + } + + # Extract arguments if available + if getattr(function_call, "args", None): + tool_call["arguments"] = safe_serialize(function_call.args) + + tool_calls.append(tool_call) + + # Extract from automatic_function_calling_history + # This is the history of tool calls made by the model + if getattr(response, "automatic_function_calling_history", None): + for content in response.automatic_function_calling_history: + if not getattr(content, "parts", None): + continue + + for part in getattr(content, "parts", []): + if getattr(part, "function_call", None): + function_call = part.function_call + tool_call = { + "name": getattr(function_call, "name", None), + "type": "function_call", + } + + # Extract arguments if available + if hasattr(function_call, "args"): + tool_call["arguments"] = safe_serialize(function_call.args) + + tool_calls.append(tool_call) + + return tool_calls if tool_calls else None + + +def _capture_tool_input(args, kwargs, tool): + # type: (tuple[Any, ...], dict[str, Any], Tool) -> dict[str, Any] + """Capture tool input from args and kwargs.""" + tool_input = kwargs.copy() if kwargs else {} + + # If we have positional args, try to map them to the function signature + if args: + try: + sig = inspect.signature(tool) + param_names = list(sig.parameters.keys()) + for i, arg in enumerate(args): + if i < len(param_names): + tool_input[param_names[i]] = arg + except Exception: + # Fallback if we can't get the signature + tool_input["args"] = args + + return tool_input + + +def _create_tool_span(tool_name, tool_doc): + # type: (str, Optional[str]) -> Span + """Create a span for tool execution.""" + span = sentry_sdk.start_span( + op=OP.GEN_AI_EXECUTE_TOOL, + name=f"execute_tool {tool_name}", + origin=ORIGIN, + ) + span.set_data(SPANDATA.GEN_AI_TOOL_NAME, tool_name) + span.set_data(SPANDATA.GEN_AI_TOOL_TYPE, "function") + if tool_doc: + span.set_data(SPANDATA.GEN_AI_TOOL_DESCRIPTION, tool_doc) + return span + + +def wrapped_tool(tool): + # type: (Tool | Callable[..., Any]) -> Tool | Callable[..., Any] + """Wrap a tool to emit execute_tool spans when called.""" + if not callable(tool): + # Not a callable function, return as-is (predefined tools) + return tool + + tool_name = getattr(tool, "__name__", "unknown") + tool_doc = tool.__doc__ + + if inspect.iscoroutinefunction(tool): + # Async function + @wraps(tool) + async def async_wrapped(*args, **kwargs): + # type: (Any, Any) -> Any + with _create_tool_span(tool_name, tool_doc) as span: + # Capture tool input + tool_input = _capture_tool_input(args, kwargs, tool) + with capture_internal_exceptions(): + span.set_data( + SPANDATA.GEN_AI_TOOL_INPUT, safe_serialize(tool_input) + ) + + try: + result = await tool(*args, **kwargs) + + # Capture tool output + with capture_internal_exceptions(): + span.set_data( + SPANDATA.GEN_AI_TOOL_OUTPUT, safe_serialize(result) + ) + + return result + except Exception as exc: + _capture_exception(exc) + raise + + return async_wrapped + else: + # Sync function + @wraps(tool) + def sync_wrapped(*args, **kwargs): + # type: (Any, Any) -> Any + with _create_tool_span(tool_name, tool_doc) as span: + # Capture tool input + tool_input = _capture_tool_input(args, kwargs, tool) + with capture_internal_exceptions(): + span.set_data( + SPANDATA.GEN_AI_TOOL_INPUT, safe_serialize(tool_input) + ) + + try: + result = tool(*args, **kwargs) + + # Capture tool output + with capture_internal_exceptions(): + span.set_data( + SPANDATA.GEN_AI_TOOL_OUTPUT, safe_serialize(result) + ) + + return result + except Exception as exc: + _capture_exception(exc) + raise + + return sync_wrapped + + +def wrapped_config_with_tools(config): + # type: (GenerateContentConfig) -> GenerateContentConfig + """Wrap tools in config to emit execute_tool spans. Tools are sometimes passed directly as + callable functions as a part of the config object.""" + + if not config or not getattr(config, "tools", None): + return config + + result = copy.copy(config) + result.tools = [wrapped_tool(tool) for tool in config.tools] + + return result + + +def _extract_response_text(response): + # type: (GenerateContentResponse) -> Optional[List[str]] + """Extract text from response candidates.""" + + if not response or not getattr(response, "candidates", []): + return None + + texts = [] + for candidate in response.candidates: + if not hasattr(candidate, "content") or not hasattr(candidate.content, "parts"): + continue + + for part in candidate.content.parts: + if getattr(part, "text", None): + texts.append(part.text) + + return texts if texts else None + + +def extract_finish_reasons(response): + # type: (GenerateContentResponse) -> Optional[List[str]] + """Extract finish reasons from response candidates.""" + if not response or not getattr(response, "candidates", []): + return None + + finish_reasons = [] + for candidate in response.candidates: + if getattr(candidate, "finish_reason", None): + # Convert enum value to string if necessary + reason = str(candidate.finish_reason) + # Remove enum prefix if present (e.g., "FinishReason.STOP" -> "STOP") + if "." in reason: + reason = reason.split(".")[-1] + finish_reasons.append(reason) + + return finish_reasons if finish_reasons else None + + +def set_span_data_for_request(span, integration, model, contents, kwargs): + # type: (Span, Any, str, ContentListUnion, dict[str, Any]) -> None + """Set span data for the request.""" + span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model) + + if kwargs.get("stream", False): + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + + config = kwargs.get("config") + + if config is None: + return + + config = cast(GenerateContentConfig, config) + + # Set input messages/prompts if PII is allowed + if should_send_default_pii() and integration.include_prompts: + messages = [] + + # Add system instruction if present + if hasattr(config, "system_instruction"): + system_instruction = config.system_instruction + if system_instruction: + system_text = extract_contents_text(system_instruction) + if system_text: + messages.append({"role": "system", "content": system_text}) + + # Add user message + contents_text = extract_contents_text(contents) + if contents_text: + messages.append({"role": "user", "content": contents_text}) + + if messages: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages, + unpack=False, + ) + + # Extract parameters directly from config (not nested under generation_config) + for param, span_key in [ + ("temperature", SPANDATA.GEN_AI_REQUEST_TEMPERATURE), + ("top_p", SPANDATA.GEN_AI_REQUEST_TOP_P), + ("top_k", SPANDATA.GEN_AI_REQUEST_TOP_K), + ("max_output_tokens", SPANDATA.GEN_AI_REQUEST_MAX_TOKENS), + ("presence_penalty", SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY), + ("frequency_penalty", SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY), + ("seed", SPANDATA.GEN_AI_REQUEST_SEED), + ]: + if hasattr(config, param): + value = getattr(config, param) + if value is not None: + span.set_data(span_key, value) + + # Set tools if available + if hasattr(config, "tools"): + tools = config.tools + if tools: + formatted_tools = _format_tools_for_span(tools) + if formatted_tools: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, + formatted_tools, + unpack=False, + ) + + +def set_span_data_for_response(span, integration, response): + # type: (Span, Any, GenerateContentResponse) -> None + """Set span data for the response.""" + if not response: + return + + if should_send_default_pii() and integration.include_prompts: + response_texts = _extract_response_text(response) + if response_texts: + # Format as JSON string array as per documentation + span.set_data(SPANDATA.GEN_AI_RESPONSE_TEXT, safe_serialize(response_texts)) + + tool_calls = extract_tool_calls(response) + if tool_calls: + # Tool calls should be JSON serialized + span.set_data(SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, safe_serialize(tool_calls)) + + finish_reasons = extract_finish_reasons(response) + if finish_reasons: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS, finish_reasons + ) + + if getattr(response, "response_id", None): + span.set_data(SPANDATA.GEN_AI_RESPONSE_ID, response.response_id) + + if getattr(response, "model_version", None): + span.set_data(SPANDATA.GEN_AI_RESPONSE_MODEL, response.model_version) + + usage_data = extract_usage_data(response) + + if usage_data["input_tokens"]: + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, usage_data["input_tokens"]) + + if usage_data["input_tokens_cached"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED, + usage_data["input_tokens_cached"], + ) + + if usage_data["output_tokens"]: + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, usage_data["output_tokens"]) + + if usage_data["output_tokens_reasoning"]: + span.set_data( + SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING, + usage_data["output_tokens_reasoning"], + ) + + if usage_data["total_tokens"]: + span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, usage_data["total_tokens"]) + + +def prepare_generate_content_args(args, kwargs): + # type: (tuple[Any, ...], dict[str, Any]) -> tuple[Any, Any, str] + """Extract and prepare common arguments for generate_content methods.""" + model = args[0] if args else kwargs.get("model", "unknown") + contents = args[1] if len(args) > 1 else kwargs.get("contents") + model_name = get_model_name(model) + + config = kwargs.get("config") + wrapped_config = wrapped_config_with_tools(config) + if wrapped_config is not config: + kwargs["config"] = wrapped_config + + return model, contents, model_name diff --git a/setup.py b/setup.py index 274e343be7..c6e391d27a 100644 --- a/setup.py +++ b/setup.py @@ -84,6 +84,7 @@ def get_file_text(file_name): "statsig": ["statsig>=0.55.3"], "tornado": ["tornado>=6"], "unleash": ["UnleashClient>=6.0.1"], + "google-genai": ["google-genai>=1.29.0"], }, entry_points={ "opentelemetry_propagator": [ diff --git a/tests/integrations/google_genai/__init__.py b/tests/integrations/google_genai/__init__.py new file mode 100644 index 0000000000..5143bf4536 --- /dev/null +++ b/tests/integrations/google_genai/__init__.py @@ -0,0 +1,4 @@ +import pytest + +pytest.importorskip("google") +pytest.importorskip("google.genai") diff --git a/tests/integrations/google_genai/test_google_genai.py b/tests/integrations/google_genai/test_google_genai.py new file mode 100644 index 0000000000..470be31944 --- /dev/null +++ b/tests/integrations/google_genai/test_google_genai.py @@ -0,0 +1,907 @@ +import json +import pytest +from unittest import mock + +from google import genai +from google.genai import types as genai_types + +from sentry_sdk import start_transaction +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.integrations.google_genai import GoogleGenAIIntegration + + +@pytest.fixture +def mock_genai_client(): + """Fixture that creates a real genai.Client with mocked HTTP responses.""" + client = genai.Client(api_key="test-api-key") + return client + + +def create_mock_http_response(response_body): + """ + Create a mock HTTP response that the API client's request() method would return. + + Args: + response_body: The JSON body as a string or dict + + Returns: + An HttpResponse object with headers and body + """ + if isinstance(response_body, dict): + response_body = json.dumps(response_body) + + return genai_types.HttpResponse( + headers={ + "content-type": "application/json; charset=UTF-8", + }, + body=response_body, + ) + + +def create_mock_streaming_responses(response_chunks): + """ + Create a generator that yields mock HTTP responses for streaming. + + Args: + response_chunks: List of dicts, each representing a chunk's JSON body + + Returns: + A generator that yields HttpResponse objects + """ + for chunk in response_chunks: + yield create_mock_http_response(chunk) + + +# Sample API response JSON (based on real API format from user) +EXAMPLE_API_RESPONSE_JSON = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "Hello! How can I help you today?"}], + }, + "finishReason": "STOP", + } + ], + "usageMetadata": { + "promptTokenCount": 10, + "candidatesTokenCount": 20, + "totalTokenCount": 30, + "cachedContentTokenCount": 5, + "thoughtsTokenCount": 3, + }, + "modelVersion": "gemini-1.5-flash", + "responseId": "response-id-123", +} + + +def create_test_config( + temperature=None, + top_p=None, + top_k=None, + max_output_tokens=None, + presence_penalty=None, + frequency_penalty=None, + seed=None, + system_instruction=None, + tools=None, +): + """Create a GenerateContentConfig.""" + config_dict = {} + + if temperature is not None: + config_dict["temperature"] = temperature + if top_p is not None: + config_dict["top_p"] = top_p + if top_k is not None: + config_dict["top_k"] = top_k + if max_output_tokens is not None: + config_dict["max_output_tokens"] = max_output_tokens + if presence_penalty is not None: + config_dict["presence_penalty"] = presence_penalty + if frequency_penalty is not None: + config_dict["frequency_penalty"] = frequency_penalty + if seed is not None: + config_dict["seed"] = seed + if system_instruction is not None: + # Convert string to Content for system instruction + if isinstance(system_instruction, str): + system_instruction = genai_types.Content( + parts=[genai_types.Part(text=system_instruction)], role="system" + ) + config_dict["system_instruction"] = system_instruction + if tools is not None: + config_dict["tools"] = tools + + return genai_types.GenerateContentConfig(**config_dict) + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [ + (True, True), + (True, False), + (False, True), + (False, False), + ], +) +def test_nonstreaming_generate_content( + sentry_init, capture_events, send_default_pii, include_prompts, mock_genai_client +): + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + # Mock the HTTP response at the _api_client.request() level + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, + "request", + return_value=mock_http_response, + ): + with start_transaction(name="google_genai"): + config = create_test_config(temperature=0.7, max_output_tokens=100) + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Tell me a joke", config=config + ) + assert len(events) == 1 + (event,) = events + + assert event["type"] == "transaction" + assert event["transaction"] == "google_genai" + + # Should have 2 spans: invoke_agent and chat + assert len(event["spans"]) == 2 + invoke_span, chat_span = event["spans"] + + # Check invoke_agent span + assert invoke_span["op"] == OP.GEN_AI_INVOKE_AGENT + assert invoke_span["description"] == "invoke_agent" + assert invoke_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == "gemini-1.5-flash" + assert invoke_span["data"][SPANDATA.GEN_AI_SYSTEM] == "gcp.gemini" + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "gemini-1.5-flash" + assert invoke_span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "invoke_agent" + + # Check chat span + assert chat_span["op"] == OP.GEN_AI_CHAT + assert chat_span["description"] == "chat gemini-1.5-flash" + assert chat_span["data"][SPANDATA.GEN_AI_OPERATION_NAME] == "chat" + assert chat_span["data"][SPANDATA.GEN_AI_SYSTEM] == "gcp.gemini" + assert chat_span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "gemini-1.5-flash" + + if send_default_pii and include_prompts: + # Messages are serialized as JSON strings + messages = json.loads(invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES]) + assert messages == [{"role": "user", "content": "Tell me a joke"}] + + # Response text is stored as a JSON array + response_text = chat_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + # Parse the JSON array + response_texts = json.loads(response_text) + assert response_texts == ["Hello! How can I help you today?"] + else: + assert SPANDATA.GEN_AI_REQUEST_MESSAGES not in invoke_span["data"] + assert SPANDATA.GEN_AI_RESPONSE_TEXT not in chat_span["data"] + + # Check token usage + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 10 + # Output tokens now include reasoning tokens: candidates_token_count (20) + thoughts_token_count (3) = 23 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 23 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 30 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED] == 5 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING] == 3 + + # Check configuration parameters + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_TEMPERATURE] == 0.7 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MAX_TOKENS] == 100 + + +def test_generate_content_with_system_instruction( + sentry_init, capture_events, mock_genai_client +): + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config( + system_instruction="You are a helpful assistant", + temperature=0.5, + ) + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="What is 2+2?", config=config + ) + + (event,) = events + invoke_span = event["spans"][0] + + # Check that system instruction is included in messages + # (PII is enabled and include_prompts is True in this test) + messages_str = invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + # Parse the JSON string to verify content + messages = json.loads(messages_str) + assert len(messages) == 2 + assert messages[0] == {"role": "system", "content": "You are a helpful assistant"} + assert messages[1] == {"role": "user", "content": "What is 2+2?"} + + +def test_generate_content_with_tools(sentry_init, capture_events, mock_genai_client): + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Create a mock tool function + def get_weather(location: str) -> str: + """Get the weather for a location""" + return f"The weather in {location} is sunny" + + # Create a tool with function declarations using real types + function_declaration = genai_types.FunctionDeclaration( + name="get_weather_tool", + description="Get weather information (tool object)", + parameters=genai_types.Schema( + type=genai_types.Type.OBJECT, + properties={ + "location": genai_types.Schema( + type=genai_types.Type.STRING, + description="The location to get weather for", + ) + }, + required=["location"], + ), + ) + + mock_tool = genai_types.Tool(function_declarations=[function_declaration]) + + # API response for tool usage + tool_response_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "I'll check the weather."}], + }, + "finishReason": "STOP", + } + ], + "usageMetadata": { + "promptTokenCount": 15, + "candidatesTokenCount": 10, + "totalTokenCount": 25, + }, + } + + mock_http_response = create_mock_http_response(tool_response_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config(tools=[get_weather, mock_tool]) + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="What's the weather?", config=config + ) + + (event,) = events + invoke_span = event["spans"][0] + + # Check that tools are recorded (data is serialized as a string) + tools_data_str = invoke_span["data"][SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS] + # Parse the JSON string to verify content + tools_data = json.loads(tools_data_str) + assert len(tools_data) == 2 + + # The order of tools may not be guaranteed, so sort by name and description for comparison + sorted_tools = sorted( + tools_data, key=lambda t: (t.get("name", ""), t.get("description", "")) + ) + + # The function tool + assert sorted_tools[0]["name"] == "get_weather" + assert sorted_tools[0]["description"] == "Get the weather for a location" + + # The FunctionDeclaration tool + assert sorted_tools[1]["name"] == "get_weather_tool" + assert sorted_tools[1]["description"] == "Get weather information (tool object)" + + +def test_tool_execution(sentry_init, capture_events): + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + # Create a mock tool function + def get_weather(location: str) -> str: + """Get the weather for a location""" + return f"The weather in {location} is sunny" + + # Create wrapped version of the tool + from sentry_sdk.integrations.google_genai.utils import wrapped_tool + + wrapped_weather = wrapped_tool(get_weather) + + # Execute the wrapped tool + with start_transaction(name="test_tool"): + result = wrapped_weather("San Francisco") + + assert result == "The weather in San Francisco is sunny" + + (event,) = events + assert len(event["spans"]) == 1 + tool_span = event["spans"][0] + + assert tool_span["op"] == OP.GEN_AI_EXECUTE_TOOL + assert tool_span["description"] == "execute_tool get_weather" + assert tool_span["data"][SPANDATA.GEN_AI_TOOL_NAME] == "get_weather" + assert tool_span["data"][SPANDATA.GEN_AI_TOOL_TYPE] == "function" + assert ( + tool_span["data"][SPANDATA.GEN_AI_TOOL_DESCRIPTION] + == "Get the weather for a location" + ) + + +def test_error_handling(sentry_init, capture_events, mock_genai_client): + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Mock an error at the HTTP level + with mock.patch.object( + mock_genai_client._api_client, "request", side_effect=Exception("API Error") + ): + with start_transaction(name="google_genai"): + with pytest.raises(Exception, match="API Error"): + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", + contents="This will fail", + config=create_test_config(), + ) + + # Should have both transaction and error events + assert len(events) == 2 + error_event, transaction_event = events + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "Exception" + assert error_event["exception"]["values"][0]["value"] == "API Error" + assert error_event["exception"]["values"][0]["mechanism"]["type"] == "google_genai" + + +def test_streaming_generate_content(sentry_init, capture_events, mock_genai_client): + """Test streaming with generate_content_stream, verifying chunk accumulation.""" + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + # Create streaming chunks - simulating a multi-chunk response + # Chunk 1: First part of text with partial usage metadata + chunk1_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "Hello! "}], + }, + # No finishReason in intermediate chunks + } + ], + "usageMetadata": { + "promptTokenCount": 10, + "candidatesTokenCount": 2, + "totalTokenCount": 12, # Not set in intermediate chunks + }, + "responseId": "response-id-stream-123", + "modelVersion": "gemini-1.5-flash", + } + + # Chunk 2: Second part of text with more usage metadata + chunk2_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "How can I "}], + }, + } + ], + "usageMetadata": { + "promptTokenCount": 10, + "candidatesTokenCount": 3, + "totalTokenCount": 13, + }, + } + + # Chunk 3: Final part with finish reason and complete usage metadata + chunk3_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "help you today?"}], + }, + "finishReason": "STOP", + } + ], + "usageMetadata": { + "promptTokenCount": 10, + "candidatesTokenCount": 7, + "totalTokenCount": 25, + "cachedContentTokenCount": 5, + "thoughtsTokenCount": 3, + }, + } + + # Create streaming mock responses + stream_chunks = [chunk1_json, chunk2_json, chunk3_json] + mock_stream = create_mock_streaming_responses(stream_chunks) + + with mock.patch.object( + mock_genai_client._api_client, "request_streamed", return_value=mock_stream + ): + with start_transaction(name="google_genai"): + config = create_test_config() + stream = mock_genai_client.models.generate_content_stream( + model="gemini-1.5-flash", contents="Stream me a response", config=config + ) + + # Consume the stream (this is what users do with the integration wrapper) + collected_chunks = list(stream) + + # Verify we got all chunks + assert len(collected_chunks) == 3 + assert collected_chunks[0].candidates[0].content.parts[0].text == "Hello! " + assert collected_chunks[1].candidates[0].content.parts[0].text == "How can I " + assert collected_chunks[2].candidates[0].content.parts[0].text == "help you today?" + + (event,) = events + + # There should be 2 spans: invoke_agent and chat + assert len(event["spans"]) == 2 + invoke_span = event["spans"][0] + chat_span = event["spans"][1] + + # Check that streaming flag is set on both spans + assert invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True + assert chat_span["data"][SPANDATA.GEN_AI_RESPONSE_STREAMING] is True + + # Verify accumulated response text (all chunks combined) + expected_full_text = "Hello! How can I help you today?" + # Response text is stored as a JSON string + chat_response_text = json.loads(chat_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT]) + invoke_response_text = json.loads( + invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + ) + assert chat_response_text == [expected_full_text] + assert invoke_response_text == [expected_full_text] + + # Verify finish reasons (only the final chunk has a finish reason) + # When there's a single finish reason, it's stored as a plain string (not JSON) + assert SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS in chat_span["data"] + assert SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS in invoke_span["data"] + assert chat_span["data"][SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS] == "STOP" + assert invoke_span["data"][SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS] == "STOP" + + # Verify token counts - should reflect accumulated values + # Input tokens: max of all chunks = 10 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 30 + assert invoke_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS] == 30 + + # Output tokens: candidates (2 + 3 + 7 = 12) + reasoning (3) = 15 + # Note: output_tokens includes both candidates and reasoning tokens + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 15 + assert invoke_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS] == 15 + + # Total tokens: from the last chunk + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 50 + assert invoke_span["data"][SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS] == 50 + + # Cached tokens: max of all chunks = 5 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED] == 5 + assert invoke_span["data"][SPANDATA.GEN_AI_USAGE_INPUT_TOKENS_CACHED] == 5 + + # Reasoning tokens: sum of thoughts_token_count = 3 + assert chat_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING] == 3 + assert invoke_span["data"][SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS_REASONING] == 3 + + # Verify model name + assert chat_span["data"][SPANDATA.GEN_AI_REQUEST_MODEL] == "gemini-1.5-flash" + assert invoke_span["data"][SPANDATA.GEN_AI_AGENT_NAME] == "gemini-1.5-flash" + + +def test_span_origin(sentry_init, capture_events, mock_genai_client): + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config() + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Test origin", config=config + ) + + (event,) = events + + assert event["contexts"]["trace"]["origin"] == "manual" + for span in event["spans"]: + assert span["origin"] == "auto.ai.google_genai" + + +def test_response_without_usage_metadata( + sentry_init, capture_events, mock_genai_client +): + """Test handling of responses without usage metadata""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Response without usage metadata + response_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "No usage data"}], + }, + "finishReason": "STOP", + } + ], + } + + mock_http_response = create_mock_http_response(response_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config() + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Test", config=config + ) + + (event,) = events + chat_span = event["spans"][1] + + # Usage data should not be present + assert SPANDATA.GEN_AI_USAGE_INPUT_TOKENS not in chat_span["data"] + assert SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS not in chat_span["data"] + assert SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS not in chat_span["data"] + + +def test_multiple_candidates(sentry_init, capture_events, mock_genai_client): + """Test handling of multiple response candidates""" + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + # Response with multiple candidates + multi_candidate_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "Response 1"}], + }, + "finishReason": "STOP", + }, + { + "content": { + "role": "model", + "parts": [{"text": "Response 2"}], + }, + "finishReason": "MAX_TOKENS", + }, + ], + "usageMetadata": { + "promptTokenCount": 5, + "candidatesTokenCount": 15, + "totalTokenCount": 20, + }, + } + + mock_http_response = create_mock_http_response(multi_candidate_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config() + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Generate multiple", config=config + ) + + (event,) = events + chat_span = event["spans"][1] + + # Should capture all responses + # Response text is stored as a JSON string when there are multiple responses + response_text = chat_span["data"][SPANDATA.GEN_AI_RESPONSE_TEXT] + if isinstance(response_text, str) and response_text.startswith("["): + # It's a JSON array + response_list = json.loads(response_text) + assert response_list == ["Response 1", "Response 2"] + else: + # It's concatenated + assert response_text == "Response 1\nResponse 2" + + # Finish reasons are serialized as JSON + finish_reasons = json.loads( + chat_span["data"][SPANDATA.GEN_AI_RESPONSE_FINISH_REASONS] + ) + assert finish_reasons == ["STOP", "MAX_TOKENS"] + + +def test_all_configuration_parameters(sentry_init, capture_events, mock_genai_client): + """Test that all configuration parameters are properly recorded""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + config = create_test_config( + temperature=0.8, + top_p=0.95, + top_k=40, + max_output_tokens=2048, + presence_penalty=0.1, + frequency_penalty=0.2, + seed=12345, + ) + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Test all params", config=config + ) + + (event,) = events + invoke_span = event["spans"][0] + + # Check all parameters are recorded + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_TEMPERATURE] == 0.8 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_TOP_P] == 0.95 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_TOP_K] == 40 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MAX_TOKENS] == 2048 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY] == 0.1 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY] == 0.2 + assert invoke_span["data"][SPANDATA.GEN_AI_REQUEST_SEED] == 12345 + + +def test_empty_response(sentry_init, capture_events, mock_genai_client): + """Test handling of minimal response with no content""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Minimal response with empty candidates array + minimal_response_json = {"candidates": []} + mock_http_response = create_mock_http_response(minimal_response_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + response = mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Test", config=create_test_config() + ) + + # Response will have an empty candidates list + assert response is not None + assert len(response.candidates) == 0 + + (event,) = events + # Should still create spans even with empty candidates + assert len(event["spans"]) == 2 + + +def test_response_with_different_id_fields( + sentry_init, capture_events, mock_genai_client +): + """Test handling of different response ID field names""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Response with response_id and model_version + response_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [{"text": "Test"}], + }, + "finishReason": "STOP", + } + ], + "responseId": "resp-456", + "modelVersion": "gemini-1.5-flash-001", + } + + mock_http_response = create_mock_http_response(response_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents="Test", config=create_test_config() + ) + + (event,) = events + chat_span = event["spans"][1] + + assert chat_span["data"][SPANDATA.GEN_AI_RESPONSE_ID] == "resp-456" + assert chat_span["data"][SPANDATA.GEN_AI_RESPONSE_MODEL] == "gemini-1.5-flash-001" + + +def test_tool_with_async_function(sentry_init, capture_events): + """Test that async tool functions are properly wrapped""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + capture_events() + + # Create an async tool function + async def async_tool(param: str) -> str: + """An async tool""" + return f"Async result: {param}" + + # Import is skipped in sync tests, but we can test the wrapping logic + from sentry_sdk.integrations.google_genai.utils import wrapped_tool + + # The wrapper should handle async functions + wrapped_async_tool = wrapped_tool(async_tool) + assert wrapped_async_tool != async_tool # Should be wrapped + assert hasattr(wrapped_async_tool, "__wrapped__") # Should preserve original + + +def test_contents_as_none(sentry_init, capture_events, mock_genai_client): + """Test handling when contents parameter is None""" + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", contents=None, config=create_test_config() + ) + + (event,) = events + invoke_span = event["spans"][0] + + # Should handle None contents gracefully + messages = invoke_span["data"].get(SPANDATA.GEN_AI_REQUEST_MESSAGES, []) + # Should only have system message if any, not user message + assert all(msg["role"] != "user" or msg["content"] is not None for msg in messages) + + +def test_tool_calls_extraction(sentry_init, capture_events, mock_genai_client): + """Test extraction of tool/function calls from response""" + sentry_init( + integrations=[GoogleGenAIIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + # Response with function calls + function_call_response_json = { + "candidates": [ + { + "content": { + "role": "model", + "parts": [ + {"text": "I'll help you with that."}, + { + "functionCall": { + "name": "get_weather", + "args": { + "location": "San Francisco", + "unit": "celsius", + }, + } + }, + { + "functionCall": { + "name": "get_time", + "args": {"timezone": "PST"}, + } + }, + ], + }, + "finishReason": "STOP", + } + ], + "usageMetadata": { + "promptTokenCount": 20, + "candidatesTokenCount": 30, + "totalTokenCount": 50, + }, + } + + mock_http_response = create_mock_http_response(function_call_response_json) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", + contents="What's the weather and time?", + config=create_test_config(), + ) + + (event,) = events + chat_span = event["spans"][1] # The chat span + + # Check that tool calls are extracted and stored + assert SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS in chat_span["data"] + + # Parse the JSON string to verify content + tool_calls = json.loads(chat_span["data"][SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS]) + + assert len(tool_calls) == 2 + + # First tool call + assert tool_calls[0]["name"] == "get_weather" + assert tool_calls[0]["type"] == "function_call" + # Arguments are serialized as JSON strings + assert json.loads(tool_calls[0]["arguments"]) == { + "location": "San Francisco", + "unit": "celsius", + } + + # Second tool call + assert tool_calls[1]["name"] == "get_time" + assert tool_calls[1]["type"] == "function_call" + # Arguments are serialized as JSON strings + assert json.loads(tool_calls[1]["arguments"]) == {"timezone": "PST"} diff --git a/tox.ini b/tox.ini index 4bfc90cee9..2490c6ffb5 100644 --- a/tox.ini +++ b/tox.ini @@ -78,6 +78,11 @@ envlist = {py3.9,py3.12,py3.13}-langgraph-v0.6.8 {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 + {py3.9,py3.12,py3.13}-google_genai-v1.29.0 + {py3.9,py3.12,py3.13}-google_genai-v1.33.0 + {py3.9,py3.12,py3.13}-google_genai-v1.37.0 + {py3.9,py3.12,py3.13}-google_genai-v1.42.0 + {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 @@ -87,14 +92,14 @@ envlist = {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 - {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc2 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc4 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.46 + {py3.9,py3.12,py3.13}-boto3-v1.40.48 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -110,14 +115,14 @@ envlist = {py3.6}-pymongo-v3.5.1 {py3.6,py3.10,py3.11}-pymongo-v3.13.0 - {py3.9,py3.12,py3.13}-pymongo-v4.15.2 + {py3.9,py3.12,py3.13}-pymongo-v4.15.3 {py3.6}-redis-v2.10.6 {py3.6,py3.7,py3.8}-redis-v3.5.3 {py3.7,py3.10,py3.11}-redis-v4.6.0 {py3.8,py3.11,py3.12}-redis-v5.3.1 {py3.9,py3.12,py3.13}-redis-v6.4.0 - {py3.9,py3.12,py3.13}-redis-v7.0.0b2 + {py3.9,py3.12,py3.13}-redis-v7.0.0b3 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 @@ -153,7 +158,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.9,py3.12,py3.13}-strawberry-v0.283.1 + {py3.9,py3.12,py3.13}-strawberry-v0.283.2 # ~~~ Network ~~~ @@ -222,7 +227,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.118.0 + {py3.8,py3.12,py3.13}-fastapi-v0.118.2 # ~~~ Web 2 ~~~ @@ -381,6 +386,12 @@ deps = langgraph-v0.6.8: langgraph==0.6.8 langgraph-v1.0.0a4: langgraph==1.0.0a4 + google_genai-v1.29.0: google-genai==1.29.0 + google_genai-v1.33.0: google-genai==1.33.0 + google_genai-v1.37.0: google-genai==1.37.0 + google_genai-v1.42.0: google-genai==1.42.0 + google_genai: pytest-asyncio + openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 @@ -391,7 +402,7 @@ deps = huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 huggingface_hub-v0.35.3: huggingface_hub==0.35.3 - huggingface_hub-v1.0.0rc2: huggingface_hub==1.0.0rc2 + huggingface_hub-v1.0.0rc4: huggingface_hub==1.0.0rc4 huggingface_hub: responses huggingface_hub: pytest-httpx @@ -400,7 +411,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.46: boto3==1.40.46 + boto3-v1.40.48: boto3==1.40.48 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -419,7 +430,7 @@ deps = pymongo-v3.5.1: pymongo==3.5.1 pymongo-v3.13.0: pymongo==3.13.0 - pymongo-v4.15.2: pymongo==4.15.2 + pymongo-v4.15.3: pymongo==4.15.3 pymongo: mockupdb redis-v2.10.6: redis==2.10.6 @@ -427,7 +438,7 @@ deps = redis-v4.6.0: redis==4.6.0 redis-v5.3.1: redis==5.3.1 redis-v6.4.0: redis==6.4.0 - redis-v7.0.0b2: redis==7.0.0b2 + redis-v7.0.0b3: redis==7.0.0b3 redis: fakeredis!=1.7.4 redis: pytest<8.0.0 redis-v4.6.0: fakeredis<2.31.0 @@ -477,7 +488,7 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.283.1: strawberry-graphql[fastapi,flask]==0.283.1 + strawberry-v0.283.2: strawberry-graphql[fastapi,flask]==0.283.2 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 @@ -604,7 +615,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.92.0: fastapi==0.92.0 fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.118.0: fastapi==0.118.0 + fastapi-v0.118.2: fastapi==0.118.2 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -747,6 +758,7 @@ setenv = falcon: TESTPATH=tests/integrations/falcon fastapi: TESTPATH=tests/integrations/fastapi flask: TESTPATH=tests/integrations/flask + google_genai: TESTPATH=tests/integrations/google_genai gql: TESTPATH=tests/integrations/gql graphene: TESTPATH=tests/integrations/graphene grpc: TESTPATH=tests/integrations/grpc From 97d675655fd44a444623004afab7903026f41bef Mon Sep 17 00:00:00 2001 From: svartalf Date: Fri, 10 Oct 2025 13:06:00 +0200 Subject: [PATCH 537/591] fix(Ray): Retain the original function name when patching Ray tasks (#4858) ### Description Without "@functools.wraps" added, Ray exposes Prometheus metrics with all tasks named "new_func" #### Issues * Follow up to [!4430](https://github.com/getsentry/sentry-python/pull/4430) comments #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- sentry_sdk/integrations/ray.py | 24 ++++++++++++++++++++---- tests/integrations/ray/test_ray.py | 3 +++ 2 files changed, 23 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/ray.py b/sentry_sdk/integrations/ray.py index 8d6cdc1201..08e78b7585 100644 --- a/sentry_sdk/integrations/ray.py +++ b/sentry_sdk/integrations/ray.py @@ -1,4 +1,5 @@ import inspect +import functools import sys import sentry_sdk @@ -17,7 +18,6 @@ import ray # type: ignore[import-not-found] except ImportError: raise DidNotEnable("Ray not installed.") -import functools from typing import TYPE_CHECKING @@ -54,12 +54,13 @@ def new_remote(f=None, *args, **kwargs): def wrapper(user_f): # type: (Callable[..., Any]) -> Any - def new_func(*f_args, _tracing=None, **f_kwargs): + @functools.wraps(user_f) + def new_func(*f_args, _sentry_tracing=None, **f_kwargs): # type: (Any, Optional[dict[str, Any]], Any) -> Any _check_sentry_initialized() transaction = sentry_sdk.continue_trace( - _tracing or {}, + _sentry_tracing or {}, op=OP.QUEUE_TASK_RAY, name=qualname_from_function(user_f), origin=RayIntegration.origin, @@ -78,6 +79,19 @@ def new_func(*f_args, _tracing=None, **f_kwargs): return result + # Patching new_func signature to add the _sentry_tracing parameter to it + # Ray later inspects the signature and finds the unexpected parameter otherwise + signature = inspect.signature(new_func) + params = list(signature.parameters.values()) + params.append( + inspect.Parameter( + "_sentry_tracing", + kind=inspect.Parameter.KEYWORD_ONLY, + default=None, + ) + ) + new_func.__signature__ = signature.replace(parameters=params) # type: ignore[attr-defined] + if f: rv = old_remote(new_func) else: @@ -99,7 +113,9 @@ def _remote_method_with_header_propagation(*args, **kwargs): for k, v in sentry_sdk.get_current_scope().iter_trace_propagation_headers() } try: - result = old_remote_method(*args, **kwargs, _tracing=tracing) + result = old_remote_method( + *args, **kwargs, _sentry_tracing=tracing + ) span.set_status(SPANSTATUS.OK) except Exception: span.set_status(SPANSTATUS.INTERNAL_ERROR) diff --git a/tests/integrations/ray/test_ray.py b/tests/integrations/ray/test_ray.py index f4e67df038..6aaced391e 100644 --- a/tests/integrations/ray/test_ray.py +++ b/tests/integrations/ray/test_ray.py @@ -100,6 +100,9 @@ def example_task(): else: example_task = ray.remote(example_task) + # Function name shouldn't be overwritten by Sentry wrapper + assert example_task._function_name == "tests.integrations.ray.test_ray.example_task" + with sentry_sdk.start_transaction(op="task", name="ray test transaction"): worker_envelopes = ray.get(example_task.remote()) From f8b9069f03c4b6adc42b7b9a820709140b8970bd Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 10 Oct 2025 13:42:34 +0200 Subject: [PATCH 538/591] tests: Update tox (#4913) ### Description Updating tox + reorg the AI group alphabetically. New openai release doesn't work on 3.8, explicitly testing on 3.9+ from there Doing this now to unblock https://github.com/getsentry/sentry-python/pull/4906#issuecomment-3389460982 #### Issues #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-ai.yml | 24 +++--- scripts/populate_tox/config.py | 10 ++- scripts/populate_tox/releases.jsonl | 15 ++-- .../split_tox_gh_actions.py | 6 +- tox.ini | 78 +++++++++---------- 5 files changed, 72 insertions(+), 61 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 65cc636cd9..1b9a341f17 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -58,6 +58,14 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-cohere" + - name: Test google_genai + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-google_genai" + - name: Test huggingface_hub + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-huggingface_hub" - name: Test langchain-base run: | set -x # print commands that are executed @@ -66,6 +74,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-langchain-notiktoken" + - name: Test langgraph + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-langgraph" - name: Test litellm run: | set -x # print commands that are executed @@ -78,22 +90,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-openai-notiktoken" - - name: Test langgraph - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-langgraph" - - name: Test google_genai - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-google_genai" - name: Test openai_agents run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_agents" - - name: Test huggingface_hub - run: | - set -x # print commands that are executed - ./scripts/runtox.sh "py${{ matrix.python-version }}-huggingface_hub" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 85988ac3cf..1f23b3fb08 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -240,7 +240,10 @@ "*": ["pytest-asyncio", "tiktoken"], "<1.55": ["httpx<0.28"], }, - "python": ">=3.8", + "python": { + ">0.0,<2.3": ">=3.8", + ">=2.3": ">=3.9", + }, }, "openai-notiktoken": { "package": "openai", @@ -249,7 +252,10 @@ "*": ["pytest-asyncio"], "<1.55": ["httpx<0.28"], }, - "python": ">=3.8", + "python": { + ">0.0,<2.3": ">=3.8", + ">=2.3": ">=3.9", + }, }, "openai_agents": { "package": "openai-agents", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 5ee83b65bc..ac5fe1de14 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.48", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -66,7 +66,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} @@ -99,11 +99,11 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc4", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.8", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.10", "yanked": false}} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} @@ -114,8 +114,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.100.2", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.107.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.2.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.57.4", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.86.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.1.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.3.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index abfa1b63cc..9dea95842b 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -72,15 +72,15 @@ "AI": [ "anthropic", "cohere", + "google_genai", + "huggingface_hub", "langchain-base", "langchain-notiktoken", + "langgraph", "litellm", "openai-base", "openai-notiktoken", - "langgraph", - "google_genai", "openai_agents", - "huggingface_hub", ], "Cloud": [ "aws_lambda", diff --git a/tox.ini b/tox.ini index 2490c6ffb5..5fb05f01bc 100644 --- a/tox.ini +++ b/tox.ini @@ -57,6 +57,17 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.13.12 {py3.9,py3.11,py3.12}-cohere-v5.18.0 + {py3.9,py3.12,py3.13}-google_genai-v1.29.0 + {py3.9,py3.12,py3.13}-google_genai-v1.33.0 + {py3.9,py3.12,py3.13}-google_genai-v1.37.0 + {py3.9,py3.12,py3.13}-google_genai-v1.42.0 + + {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc5 + {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.11,py3.12}-langchain-base-v0.2.17 {py3.9,py3.12,py3.13}-langchain-base-v0.3.27 @@ -65,41 +76,30 @@ envlist = {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.2.17 {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 + {py3.9,py3.12,py3.13}-langgraph-v0.6.10 + {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 + {py3.9,py3.12,py3.13}-litellm-v1.77.7 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.8,py3.12,py3.13}-openai-base-v2.2.0 + {py3.9,py3.12,py3.13}-openai-base-v2.3.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.8,py3.12,py3.13}-openai-notiktoken-v2.2.0 - - {py3.9,py3.12,py3.13}-langgraph-v0.6.8 - {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 - - {py3.9,py3.12,py3.13}-google_genai-v1.29.0 - {py3.9,py3.12,py3.13}-google_genai-v1.33.0 - {py3.9,py3.12,py3.13}-google_genai-v1.37.0 - {py3.9,py3.12,py3.13}-google_genai-v1.42.0 + {py3.9,py3.12,py3.13}-openai-notiktoken-v2.3.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 {py3.10,py3.12,py3.13}-openai_agents-v0.3.3 - {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 - {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc4 - # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.48 + {py3.9,py3.12,py3.13}-boto3-v1.40.49 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -227,7 +227,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.118.2 + {py3.8,py3.12,py3.13}-fastapi-v0.118.3 # ~~~ Web 2 ~~~ @@ -353,6 +353,20 @@ deps = cohere-v5.13.12: cohere==5.13.12 cohere-v5.18.0: cohere==5.18.0 + google_genai-v1.29.0: google-genai==1.29.0 + google_genai-v1.33.0: google-genai==1.33.0 + google_genai-v1.37.0: google-genai==1.37.0 + google_genai-v1.42.0: google-genai==1.42.0 + google_genai: pytest-asyncio + + huggingface_hub-v0.24.7: huggingface_hub==0.24.7 + huggingface_hub-v0.28.1: huggingface_hub==0.28.1 + huggingface_hub-v0.32.6: huggingface_hub==0.32.6 + huggingface_hub-v0.35.3: huggingface_hub==0.35.3 + huggingface_hub-v1.0.0rc5: huggingface_hub==1.0.0rc5 + huggingface_hub: responses + huggingface_hub: pytest-httpx + langchain-base-v0.1.20: langchain==0.1.20 langchain-base-v0.2.17: langchain==0.2.17 langchain-base-v0.3.27: langchain==0.3.27 @@ -368,50 +382,36 @@ deps = langchain-notiktoken: langchain-openai langchain-notiktoken-v0.3.27: langchain-community + langgraph-v0.6.10: langgraph==0.6.10 + langgraph-v1.0.0a4: langgraph==1.0.0a4 + litellm-v1.77.7: litellm==1.77.7 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.2.0: openai==2.2.0 + openai-base-v2.3.0: openai==2.3.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.2.0: openai==2.2.0 + openai-notiktoken-v2.3.0: openai==2.3.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 - langgraph-v0.6.8: langgraph==0.6.8 - langgraph-v1.0.0a4: langgraph==1.0.0a4 - - google_genai-v1.29.0: google-genai==1.29.0 - google_genai-v1.33.0: google-genai==1.33.0 - google_genai-v1.37.0: google-genai==1.37.0 - google_genai-v1.42.0: google-genai==1.42.0 - google_genai: pytest-asyncio - openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 openai_agents-v0.3.3: openai-agents==0.3.3 openai_agents: pytest-asyncio - huggingface_hub-v0.24.7: huggingface_hub==0.24.7 - huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.32.6: huggingface_hub==0.32.6 - huggingface_hub-v0.35.3: huggingface_hub==0.35.3 - huggingface_hub-v1.0.0rc4: huggingface_hub==1.0.0rc4 - huggingface_hub: responses - huggingface_hub: pytest-httpx - # ~~~ Cloud ~~~ boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.48: boto3==1.40.48 + boto3-v1.40.49: boto3==1.40.49 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -615,7 +615,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.92.0: fastapi==0.92.0 fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.118.2: fastapi==0.118.2 + fastapi-v0.118.3: fastapi==0.118.3 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart From cab17a4df648e4dac84e196b3e235b694c7d1367 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Fri, 10 Oct 2025 13:49:13 +0200 Subject: [PATCH 539/591] feat: Add source information for slow outgoing HTTP requests (#4902) Add code source attributes to outgoing HTTP requests as described in https://github.com/getsentry/sentry-docs/pull/15161. The attributes are only added if the time to receive a response to an HTTP request exceeds a configurable threshold value. Factors out functionality from SQL query source and tests that it works in the HTTP request setting. Closes https://github.com/getsentry/sentry-python/issues/4881 --- sentry_sdk/consts.py | 9 + sentry_sdk/integrations/aiohttp.py | 5 +- sentry_sdk/integrations/httpx.py | 21 +- sentry_sdk/integrations/stdlib.py | 9 +- sentry_sdk/tracing_utils.py | 88 +++-- tests/integrations/aiohttp/__init__.py | 6 + .../aiohttp/aiohttp_helpers/__init__.py | 0 .../aiohttp/aiohttp_helpers/helpers.py | 2 + tests/integrations/aiohttp/test_aiohttp.py | 352 +++++++++++++++++- tests/integrations/httpx/__init__.py | 6 + .../httpx/httpx_helpers/__init__.py | 0 .../httpx/httpx_helpers/helpers.py | 6 + tests/integrations/httpx/test_httpx.py | 310 +++++++++++++++ tests/integrations/stdlib/__init__.py | 6 + .../stdlib/httplib_helpers/__init__.py | 0 .../stdlib/httplib_helpers/helpers.py | 3 + tests/integrations/stdlib/test_httplib.py | 230 ++++++++++++ 17 files changed, 1021 insertions(+), 32 deletions(-) create mode 100644 tests/integrations/aiohttp/aiohttp_helpers/__init__.py create mode 100644 tests/integrations/aiohttp/aiohttp_helpers/helpers.py create mode 100644 tests/integrations/httpx/httpx_helpers/__init__.py create mode 100644 tests/integrations/httpx/httpx_helpers/helpers.py create mode 100644 tests/integrations/stdlib/__init__.py create mode 100644 tests/integrations/stdlib/httplib_helpers/__init__.py create mode 100644 tests/integrations/stdlib/httplib_helpers/helpers.py diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 2158276f9b..c1e587cbeb 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -909,6 +909,8 @@ def __init__( error_sampler=None, # type: Optional[Callable[[Event, Hint], Union[float, bool]]] enable_db_query_source=True, # type: bool db_query_source_threshold_ms=100, # type: int + enable_http_request_source=False, # type: bool + http_request_source_threshold_ms=100, # type: int spotlight=None, # type: Optional[Union[bool, str]] cert_file=None, # type: Optional[str] key_file=None, # type: Optional[str] @@ -1264,6 +1266,13 @@ def __init__( The query location will be added to the query for queries slower than the specified threshold. + :param enable_http_request_source: When enabled, the source location will be added to outgoing HTTP requests. + + :param http_request_source_threshold_ms: The threshold in milliseconds for adding the source location to an + outgoing HTTP request. + + The request location will be added to the request for requests slower than the specified threshold. + :param custom_repr: A custom `repr `_ function to run while serializing an object. diff --git a/sentry_sdk/integrations/aiohttp.py b/sentry_sdk/integrations/aiohttp.py index ad3202bf2c..0a417f8dc4 100644 --- a/sentry_sdk/integrations/aiohttp.py +++ b/sentry_sdk/integrations/aiohttp.py @@ -22,7 +22,7 @@ SOURCE_FOR_STYLE, TransactionSource, ) -from sentry_sdk.tracing_utils import should_propagate_trace +from sentry_sdk.tracing_utils import should_propagate_trace, add_http_request_source from sentry_sdk.utils import ( capture_internal_exceptions, ensure_integration_enabled, @@ -279,6 +279,9 @@ async def on_request_end(session, trace_config_ctx, params): span.set_data("reason", params.response.reason) span.finish() + with capture_internal_exceptions(): + add_http_request_source(span) + trace_config = TraceConfig() trace_config.on_request_start.append(on_request_start) diff --git a/sentry_sdk/integrations/httpx.py b/sentry_sdk/integrations/httpx.py index 2ddd44489f..2ada95aad0 100644 --- a/sentry_sdk/integrations/httpx.py +++ b/sentry_sdk/integrations/httpx.py @@ -1,8 +1,13 @@ import sentry_sdk +from sentry_sdk import start_span from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import Integration, DidNotEnable from sentry_sdk.tracing import BAGGAGE_HEADER_NAME -from sentry_sdk.tracing_utils import Baggage, should_propagate_trace +from sentry_sdk.tracing_utils import ( + Baggage, + should_propagate_trace, + add_http_request_source, +) from sentry_sdk.utils import ( SENSITIVE_DATA_SUBSTITUTE, capture_internal_exceptions, @@ -52,7 +57,7 @@ def send(self, request, **kwargs): with capture_internal_exceptions(): parsed_url = parse_url(https://github.com/str(request.url), sanitize=False) - with sentry_sdk.start_span( + with start_span( op=OP.HTTP_CLIENT, name="%s %s" % ( @@ -88,7 +93,10 @@ def send(self, request, **kwargs): span.set_http_status(rv.status_code) span.set_data("reason", rv.reason_phrase) - return rv + with capture_internal_exceptions(): + add_http_request_source(span) + + return rv Client.send = send @@ -106,7 +114,7 @@ async def send(self, request, **kwargs): with capture_internal_exceptions(): parsed_url = parse_url(https://github.com/str(request.url), sanitize=False) - with sentry_sdk.start_span( + with start_span( op=OP.HTTP_CLIENT, name="%s %s" % ( @@ -144,7 +152,10 @@ async def send(self, request, **kwargs): span.set_http_status(rv.status_code) span.set_data("reason", rv.reason_phrase) - return rv + with capture_internal_exceptions(): + add_http_request_source(span) + + return rv AsyncClient.send = send diff --git a/sentry_sdk/integrations/stdlib.py b/sentry_sdk/integrations/stdlib.py index d388c5bca6..3db97e5685 100644 --- a/sentry_sdk/integrations/stdlib.py +++ b/sentry_sdk/integrations/stdlib.py @@ -8,7 +8,11 @@ from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import Integration from sentry_sdk.scope import add_global_event_processor -from sentry_sdk.tracing_utils import EnvironHeaders, should_propagate_trace +from sentry_sdk.tracing_utils import ( + EnvironHeaders, + should_propagate_trace, + add_http_request_source, +) from sentry_sdk.utils import ( SENSITIVE_DATA_SUBSTITUTE, capture_internal_exceptions, @@ -135,6 +139,9 @@ def getresponse(self, *args, **kwargs): finally: span.finish() + with capture_internal_exceptions(): + add_http_request_source(span) + return rv HTTPConnection.putrequest = putrequest # type: ignore[method-assign] diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index b81d647c6d..587133ad67 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -218,33 +218,11 @@ def _should_be_included( ) -def add_query_source(span): - # type: (sentry_sdk.tracing.Span) -> None +def add_source(span, project_root, in_app_include, in_app_exclude): + # type: (sentry_sdk.tracing.Span, Optional[str], Optional[list[str]], Optional[list[str]]) -> None """ Adds OTel compatible source code information to the span """ - client = sentry_sdk.get_client() - if not client.is_active(): - return - - if span.timestamp is None or span.start_timestamp is None: - return - - should_add_query_source = client.options.get("enable_db_query_source", True) - if not should_add_query_source: - return - - duration = span.timestamp - span.start_timestamp - threshold = client.options.get("db_query_source_threshold_ms", 0) - slow_query = duration / timedelta(milliseconds=1) > threshold - - if not slow_query: - return - - project_root = client.options["project_root"] - in_app_include = client.options.get("in_app_include") - in_app_exclude = client.options.get("in_app_exclude") - # Find the correct frame frame = sys._getframe() # type: Union[FrameType, None] while frame is not None: @@ -309,6 +287,68 @@ def add_query_source(span): span.set_data(SPANDATA.CODE_FUNCTION, frame.f_code.co_name) +def add_query_source(span): + # type: (sentry_sdk.tracing.Span) -> None + """ + Adds OTel compatible source code information to a database query span + """ + client = sentry_sdk.get_client() + if not client.is_active(): + return + + if span.timestamp is None or span.start_timestamp is None: + return + + should_add_query_source = client.options.get("enable_db_query_source", True) + if not should_add_query_source: + return + + duration = span.timestamp - span.start_timestamp + threshold = client.options.get("db_query_source_threshold_ms", 0) + slow_query = duration / timedelta(milliseconds=1) > threshold + + if not slow_query: + return + + add_source( + span=span, + project_root=client.options["project_root"], + in_app_include=client.options.get("in_app_include"), + in_app_exclude=client.options.get("in_app_exclude"), + ) + + +def add_http_request_source(span): + # type: (sentry_sdk.tracing.Span) -> None + """ + Adds OTel compatible source code information to a span for an outgoing HTTP request + """ + client = sentry_sdk.get_client() + if not client.is_active(): + return + + if span.timestamp is None or span.start_timestamp is None: + return + + should_add_request_source = client.options.get("enable_http_request_source", False) + if not should_add_request_source: + return + + duration = span.timestamp - span.start_timestamp + threshold = client.options.get("http_request_source_threshold_ms", 0) + slow_query = duration / timedelta(milliseconds=1) > threshold + + if not slow_query: + return + + add_source( + span=span, + project_root=client.options["project_root"], + in_app_include=client.options.get("in_app_include"), + in_app_exclude=client.options.get("in_app_exclude"), + ) + + def extract_sentrytrace_data(header): # type: (Optional[str]) -> Optional[Dict[str, Union[str, bool, None]]] """ diff --git a/tests/integrations/aiohttp/__init__.py b/tests/integrations/aiohttp/__init__.py index 0e1409fda0..a585c11e34 100644 --- a/tests/integrations/aiohttp/__init__.py +++ b/tests/integrations/aiohttp/__init__.py @@ -1,3 +1,9 @@ +import os +import sys import pytest pytest.importorskip("aiohttp") + +# Load `aiohttp_helpers` into the module search path to test request source path names relative to module. See +# `test_request_source_with_module_in_search_path` +sys.path.insert(0, os.path.join(os.path.dirname(__file__))) diff --git a/tests/integrations/aiohttp/aiohttp_helpers/__init__.py b/tests/integrations/aiohttp/aiohttp_helpers/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/aiohttp/aiohttp_helpers/helpers.py b/tests/integrations/aiohttp/aiohttp_helpers/helpers.py new file mode 100644 index 0000000000..86a6fa39e3 --- /dev/null +++ b/tests/integrations/aiohttp/aiohttp_helpers/helpers.py @@ -0,0 +1,2 @@ +async def get_request_with_client(client, url): + await client.get(url) diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index 267ce08fdd..811bf7efca 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -1,3 +1,5 @@ +import os +import datetime import asyncio import json @@ -18,7 +20,8 @@ ) from sentry_sdk import capture_message, start_transaction -from sentry_sdk.integrations.aiohttp import AioHttpIntegration +from sentry_sdk.integrations.aiohttp import AioHttpIntegration, create_trace_config +from sentry_sdk.consts import SPANDATA from tests.conftest import ApproxDict @@ -633,6 +636,353 @@ async def handler(request): ) +@pytest.mark.asyncio +@pytest.mark.parametrize("enable_http_request_source", [None, False]) +async def test_request_source_disabled( + sentry_init, + aiohttp_raw_server, + aiohttp_client, + capture_events, + enable_http_request_source, +): + sentry_options = { + "integrations": [AioHttpIntegration()], + "traces_sample_rate": 1.0, + "http_request_source_threshold_ms": 0, + } + + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source + + sentry_init(**sentry_options) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + async def hello(request): + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", hello) + + events = capture_events() + + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +@pytest.mark.asyncio +async def test_request_source_enabled( + sentry_init, + aiohttp_raw_server, + aiohttp_client, + capture_events, +): + sentry_options = { + "integrations": [AioHttpIntegration()], + "traces_sample_rate": 1.0, + "enable_http_request_source": True, + "http_request_source_threshold_ms": 0, + } + + sentry_init(**sentry_options) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + async def hello(request): + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", hello) + + events = capture_events() + + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + +@pytest.mark.asyncio +async def test_request_source( + sentry_init, aiohttp_raw_server, aiohttp_client, capture_events +): + sentry_init( + integrations=[AioHttpIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + async def handler_with_outgoing_request(request): + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", handler_with_outgoing_request) + + events = capture_events() + + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert ( + data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.aiohttp.test_aiohttp" + ) + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/aiohttp/test_aiohttp.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "handler_with_outgoing_request" + + +@pytest.mark.asyncio +async def test_request_source_with_module_in_search_path( + sentry_init, aiohttp_raw_server, aiohttp_client, capture_events +): + """ + Test that request source is relative to the path of the module it ran in + """ + sentry_init( + integrations=[AioHttpIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + from aiohttp_helpers.helpers import get_request_with_client + + async def handler_with_outgoing_request(request): + span_client = await aiohttp_client(raw_server) + await get_request_with_client(span_client, "/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", handler_with_outgoing_request) + + events = capture_events() + + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "aiohttp_helpers.helpers" + assert data.get(SPANDATA.CODE_FILEPATH) == "aiohttp_helpers/helpers.py" + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "get_request_with_client" + + +@pytest.mark.asyncio +async def test_no_request_source_if_duration_too_short( + sentry_init, aiohttp_raw_server, aiohttp_client, capture_events +): + sentry_init( + integrations=[AioHttpIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + async def handler_with_outgoing_request(request): + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", handler_with_outgoing_request) + + events = capture_events() + + def fake_create_trace_context(*args, **kwargs): + trace_context = create_trace_config() + + async def overwrite_timestamps(session, trace_config_ctx, params): + span = trace_config_ctx.span + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=99999) + + trace_context.on_request_end.insert(0, overwrite_timestamps) + + return trace_context + + with mock.patch( + "sentry_sdk.integrations.aiohttp.create_trace_config", + fake_create_trace_context, + ): + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +@pytest.mark.asyncio +async def test_request_source_if_duration_over_threshold( + sentry_init, aiohttp_raw_server, aiohttp_client, capture_events +): + sentry_init( + integrations=[AioHttpIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + # server for making span request + async def handler(request): + return web.Response(text="OK") + + raw_server = await aiohttp_raw_server(handler) + + async def handler_with_outgoing_request(request): + span_client = await aiohttp_client(raw_server) + await span_client.get("/") + return web.Response(text="hello") + + app = web.Application() + app.router.add_get(r"/", handler_with_outgoing_request) + + events = capture_events() + + def fake_create_trace_context(*args, **kwargs): + trace_context = create_trace_config() + + async def overwrite_timestamps(session, trace_config_ctx, params): + span = trace_config_ctx.span + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=100001) + + trace_context.on_request_end.insert(0, overwrite_timestamps) + + return trace_context + + with mock.patch( + "sentry_sdk.integrations.aiohttp.create_trace_config", + fake_create_trace_context, + ): + client = await aiohttp_client(app) + await client.get("/") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert ( + data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.aiohttp.test_aiohttp" + ) + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/aiohttp/test_aiohttp.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "handler_with_outgoing_request" + + @pytest.mark.asyncio async def test_span_origin( sentry_init, diff --git a/tests/integrations/httpx/__init__.py b/tests/integrations/httpx/__init__.py index 1afd90ea3a..e524321b8b 100644 --- a/tests/integrations/httpx/__init__.py +++ b/tests/integrations/httpx/__init__.py @@ -1,3 +1,9 @@ +import os +import sys import pytest pytest.importorskip("httpx") + +# Load `httpx_helpers` into the module search path to test request source path names relative to module. See +# `test_request_source_with_module_in_search_path` +sys.path.insert(0, os.path.join(os.path.dirname(__file__))) diff --git a/tests/integrations/httpx/httpx_helpers/__init__.py b/tests/integrations/httpx/httpx_helpers/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/httpx/httpx_helpers/helpers.py b/tests/integrations/httpx/httpx_helpers/helpers.py new file mode 100644 index 0000000000..f1d4f3c98b --- /dev/null +++ b/tests/integrations/httpx/httpx_helpers/helpers.py @@ -0,0 +1,6 @@ +def get_request_with_client(client, url): + client.get(url) + + +async def async_get_request_with_client(client, url): + await client.get(url) diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index 4fd5275fb7..1f30fdf945 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -1,8 +1,11 @@ +import os +import datetime import asyncio from unittest import mock import httpx import pytest +from contextlib import contextmanager import sentry_sdk from sentry_sdk import capture_message, start_transaction @@ -393,6 +396,313 @@ def test_omit_url_data_if_parsing_fails(sentry_init, capture_events, httpx_mock) assert SPANDATA.HTTP_QUERY not in event["breadcrumbs"]["values"][0]["data"] +@pytest.mark.parametrize("enable_http_request_source", [None, False]) +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_request_source_disabled( + sentry_init, capture_events, enable_http_request_source, httpx_client, httpx_mock +): + httpx_mock.add_response() + sentry_options = { + "integrations": [HttpxIntegration()], + "traces_sample_rate": 1.0, + "http_request_source_threshold_ms": 0, + } + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source + + sentry_init(**sentry_options) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + if asyncio.iscoroutinefunction(httpx_client.get): + asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) + else: + httpx_client.get(url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_request_source_enabled(sentry_init, capture_events, httpx_client, httpx_mock): + httpx_mock.add_response() + sentry_options = { + "integrations": [HttpxIntegration()], + "traces_sample_rate": 1.0, + "enable_http_request_source": True, + "http_request_source_threshold_ms": 0, + } + sentry_init(**sentry_options) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + if asyncio.iscoroutinefunction(httpx_client.get): + asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) + else: + httpx_client.get(url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_request_source(sentry_init, capture_events, httpx_client, httpx_mock): + httpx_mock.add_response() + + sentry_init( + integrations=[HttpxIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + if asyncio.iscoroutinefunction(httpx_client.get): + asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) + else: + httpx_client.get(url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.httpx.test_httpx" + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/httpx/test_httpx.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "test_request_source" + + +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_request_source_with_module_in_search_path( + sentry_init, capture_events, httpx_client, httpx_mock +): + """ + Test that request source is relative to the path of the module it ran in + """ + httpx_mock.add_response() + sentry_init( + integrations=[HttpxIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + if asyncio.iscoroutinefunction(httpx_client.get): + from httpx_helpers.helpers import async_get_request_with_client + + asyncio.get_event_loop().run_until_complete( + async_get_request_with_client(httpx_client, url) + ) + else: + from httpx_helpers.helpers import get_request_with_client + + get_request_with_client(httpx_client, url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "httpx_helpers.helpers" + assert data.get(SPANDATA.CODE_FILEPATH) == "httpx_helpers/helpers.py" + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + if asyncio.iscoroutinefunction(httpx_client.get): + assert data.get(SPANDATA.CODE_FUNCTION) == "async_get_request_with_client" + else: + assert data.get(SPANDATA.CODE_FUNCTION) == "get_request_with_client" + + +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_no_request_source_if_duration_too_short( + sentry_init, capture_events, httpx_client, httpx_mock +): + httpx_mock.add_response() + + sentry_init( + integrations=[HttpxIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + + @contextmanager + def fake_start_span(*args, **kwargs): + with sentry_sdk.start_span(*args, **kwargs) as span: + pass + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=99999) + yield span + + with mock.patch( + "sentry_sdk.integrations.httpx.start_span", + fake_start_span, + ): + if asyncio.iscoroutinefunction(httpx_client.get): + asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) + else: + httpx_client.get(url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +@pytest.mark.parametrize( + "httpx_client", + (httpx.Client(), httpx.AsyncClient()), +) +def test_request_source_if_duration_over_threshold( + sentry_init, capture_events, httpx_client, httpx_mock +): + httpx_mock.add_response() + + sentry_init( + integrations=[HttpxIntegration()], + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + events = capture_events() + + url = "http://example.com/" + + with start_transaction(name="test_transaction"): + + @contextmanager + def fake_start_span(*args, **kwargs): + with sentry_sdk.start_span(*args, **kwargs) as span: + pass + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=100001) + yield span + + with mock.patch( + "sentry_sdk.integrations.httpx.start_span", + fake_start_span, + ): + if asyncio.iscoroutinefunction(httpx_client.get): + asyncio.get_event_loop().run_until_complete(httpx_client.get(url)) + else: + httpx_client.get(url) + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.httpx.test_httpx" + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/httpx/test_httpx.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert ( + data.get(SPANDATA.CODE_FUNCTION) + == "test_request_source_if_duration_over_threshold" + ) + + @pytest.mark.parametrize( "httpx_client", (httpx.Client(), httpx.AsyncClient()), diff --git a/tests/integrations/stdlib/__init__.py b/tests/integrations/stdlib/__init__.py new file mode 100644 index 0000000000..472e0151b2 --- /dev/null +++ b/tests/integrations/stdlib/__init__.py @@ -0,0 +1,6 @@ +import os +import sys + +# Load `httplib_helpers` into the module search path to test request source path names relative to module. See +# `test_request_source_with_module_in_search_path` +sys.path.insert(0, os.path.join(os.path.dirname(__file__))) diff --git a/tests/integrations/stdlib/httplib_helpers/__init__.py b/tests/integrations/stdlib/httplib_helpers/__init__.py new file mode 100644 index 0000000000..e69de29bb2 diff --git a/tests/integrations/stdlib/httplib_helpers/helpers.py b/tests/integrations/stdlib/httplib_helpers/helpers.py new file mode 100644 index 0000000000..875052e7b5 --- /dev/null +++ b/tests/integrations/stdlib/httplib_helpers/helpers.py @@ -0,0 +1,3 @@ +def get_request_with_connection(connection, url): + connection.request("GET", url) + connection.getresponse() diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index b8d46d0558..9bd53d6ad1 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -1,3 +1,5 @@ +import os +import datetime from http.client import HTTPConnection, HTTPSConnection from socket import SocketIO from urllib.error import HTTPError @@ -374,6 +376,234 @@ def test_option_trace_propagation_targets( assert "baggage" not in request_headers +@pytest.mark.parametrize("enable_http_request_source", [None, False]) +def test_request_source_disabled( + sentry_init, capture_events, enable_http_request_source +): + sentry_options = { + "traces_sample_rate": 1.0, + "http_request_source_threshold_ms": 0, + } + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source + + sentry_init(**sentry_options) + + events = capture_events() + + with start_transaction(name="foo"): + conn = HTTPConnection("localhost", port=PORT) + conn.request("GET", "/foo") + conn.getresponse() + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +def test_request_source_enabled(sentry_init, capture_events): + sentry_options = { + "traces_sample_rate": 1.0, + "enable_http_request_source": True, + "http_request_source_threshold_ms": 0, + } + sentry_init(**sentry_options) + + events = capture_events() + + with start_transaction(name="foo"): + conn = HTTPConnection("localhost", port=PORT) + conn.request("GET", "/foo") + conn.getresponse() + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + +def test_request_source(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + events = capture_events() + + with start_transaction(name="foo"): + conn = HTTPConnection("localhost", port=PORT) + conn.request("GET", "/foo") + conn.getresponse() + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.stdlib.test_httplib" + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/stdlib/test_httplib.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "test_request_source" + + +def test_request_source_with_module_in_search_path(sentry_init, capture_events): + """ + Test that request source is relative to the path of the module it ran in + """ + sentry_init( + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=0, + ) + + events = capture_events() + + with start_transaction(name="foo"): + from httplib_helpers.helpers import get_request_with_connection + + conn = HTTPConnection("localhost", port=PORT) + get_request_with_connection(conn, "/foo") + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "httplib_helpers.helpers" + assert data.get(SPANDATA.CODE_FILEPATH) == "httplib_helpers/helpers.py" + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert data.get(SPANDATA.CODE_FUNCTION) == "get_request_with_connection" + + +def test_no_request_source_if_duration_too_short(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + already_patched_putrequest = HTTPConnection.putrequest + + class HttpConnectionWithPatchedSpan(HTTPConnection): + def putrequest(self, *args, **kwargs) -> None: + already_patched_putrequest(self, *args, **kwargs) + span = self._sentrysdk_span # type: ignore + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=99999) + + events = capture_events() + + with start_transaction(name="foo"): + conn = HttpConnectionWithPatchedSpan("localhost", port=PORT) + conn.request("GET", "/foo") + conn.getresponse() + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO not in data + assert SPANDATA.CODE_NAMESPACE not in data + assert SPANDATA.CODE_FILEPATH not in data + assert SPANDATA.CODE_FUNCTION not in data + + +def test_request_source_if_duration_over_threshold(sentry_init, capture_events): + sentry_init( + traces_sample_rate=1.0, + enable_http_request_source=True, + http_request_source_threshold_ms=100, + ) + + already_patched_putrequest = HTTPConnection.putrequest + + class HttpConnectionWithPatchedSpan(HTTPConnection): + def putrequest(self, *args, **kwargs) -> None: + already_patched_putrequest(self, *args, **kwargs) + span = self._sentrysdk_span # type: ignore + span.start_timestamp = datetime.datetime(2024, 1, 1, microsecond=0) + span.timestamp = datetime.datetime(2024, 1, 1, microsecond=100001) + + events = capture_events() + + with start_transaction(name="foo"): + conn = HttpConnectionWithPatchedSpan("localhost", port=PORT) + conn.request("GET", "/foo") + conn.getresponse() + + (event,) = events + + span = event["spans"][-1] + assert span["description"].startswith("GET") + + data = span.get("data", {}) + + assert SPANDATA.CODE_LINENO in data + assert SPANDATA.CODE_NAMESPACE in data + assert SPANDATA.CODE_FILEPATH in data + assert SPANDATA.CODE_FUNCTION in data + + assert type(data.get(SPANDATA.CODE_LINENO)) == int + assert data.get(SPANDATA.CODE_LINENO) > 0 + assert data.get(SPANDATA.CODE_NAMESPACE) == "tests.integrations.stdlib.test_httplib" + assert data.get(SPANDATA.CODE_FILEPATH).endswith( + "tests/integrations/stdlib/test_httplib.py" + ) + + is_relative_path = data.get(SPANDATA.CODE_FILEPATH)[0] != os.sep + assert is_relative_path + + assert ( + data.get(SPANDATA.CODE_FUNCTION) + == "test_request_source_if_duration_over_threshold" + ) + + def test_span_origin(sentry_init, capture_events): sentry_init(traces_sample_rate=1.0, debug=True) events = capture_events() From 643d87e0b3fcbe8d6c69e1a84a3a9c1dd1090dcc Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 13 Oct 2025 09:26:12 +0200 Subject: [PATCH 540/591] build(deps): bump github/codeql-action from 3 to 4 (#4916) Bumps [github/codeql-action](https://github.com/github/codeql-action) from 3 to 4.
Release notes

Sourced from github/codeql-action's releases.

v3.30.8

CodeQL Action Changelog

See the releases page for the relevant changes to the CodeQL CLI and language packs.

3.30.8 - 10 Oct 2025

No user facing changes.

See the full CHANGELOG.md for more information.

v3.30.7

CodeQL Action Changelog

See the releases page for the relevant changes to the CodeQL CLI and language packs.

3.30.7 - 06 Oct 2025

No user facing changes.

See the full CHANGELOG.md for more information.

v3.30.6

CodeQL Action Changelog

See the releases page for the relevant changes to the CodeQL CLI and language packs.

3.30.6 - 02 Oct 2025

  • Update default CodeQL bundle version to 2.23.2. #3168

See the full CHANGELOG.md for more information.

v3.30.5

CodeQL Action Changelog

See the releases page for the relevant changes to the CodeQL CLI and language packs.

3.30.5 - 26 Sep 2025

  • We fixed a bug that was introduced in 3.30.4 with upload-sarif which resulted in files without a .sarif extension not getting uploaded. #3160

See the full CHANGELOG.md for more information.

v3.30.4

CodeQL Action Changelog

See the releases page for the relevant changes to the CodeQL CLI and language packs.

3.30.4 - 25 Sep 2025

... (truncated)

Changelog

Sourced from github/codeql-action's changelog.

3.29.4 - 23 Jul 2025

No user facing changes.

3.29.3 - 21 Jul 2025

No user facing changes.

3.29.2 - 30 Jun 2025

  • Experimental: When the quality-queries input for the init action is provided with an argument, separate .quality.sarif files are produced and uploaded for each language with the results of the specified queries. Do not use this in production as it is part of an internal experiment and subject to change at any time. #2935

3.29.1 - 27 Jun 2025

  • Fix bug in PR analysis where user-provided include query filter fails to exclude non-included queries. #2938
  • Update default CodeQL bundle version to 2.22.1. #2950

3.29.0 - 11 Jun 2025

  • Update default CodeQL bundle version to 2.22.0. #2925
  • Bump minimum CodeQL bundle version to 2.16.6. #2912

3.28.21 - 28 July 2025

No user facing changes.

3.28.20 - 21 July 2025

3.28.19 - 03 Jun 2025

  • The CodeQL Action no longer includes its own copy of the extractor for the actions language, which is currently in public preview. The actions extractor has been included in the CodeQL CLI since v2.20.6. If your workflow has enabled the actions language and you have pinned your tools: property to a specific version of the CodeQL CLI earlier than v2.20.6, you will need to update to at least CodeQL v2.20.6 or disable actions analysis.
  • Update default CodeQL bundle version to 2.21.4. #2910

3.28.18 - 16 May 2025

  • Update default CodeQL bundle version to 2.21.3. #2893
  • Skip validating SARIF produced by CodeQL for improved performance. #2894
  • The number of threads and amount of RAM used by CodeQL can now be set via the CODEQL_THREADS and CODEQL_RAM runner environment variables. If set, these environment variables override the threads and ram inputs respectively. #2891

3.28.17 - 02 May 2025

  • Update default CodeQL bundle version to 2.21.2. #2872

3.28.16 - 23 Apr 2025

... (truncated)

Commits
  • a841c54 Scratch uploadSpecifiedFiles tests, make uploadPayload tests instead
  • aeb12f6 Merge branch 'main' into redsun82/skip-sarif-upload-tests
  • 6fd4ceb Merge pull request #3189 from github/henrymercer/download-codeql-rate-limit
  • 196a3e5 Merge pull request #3188 from github/mbg/telemetry/partial-config
  • 98abb87 Add configuration error for rate limited CodeQL download
  • bdd2cdf Also include language in error status report for start-proxy, if available
  • fb14878 Include languages in start-proxy telemetry
  • 2ff418f Parse language before calling getCredentials
  • See full diff in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=github/codeql-action&package-manager=github_actions&previous-version=3&new-version=4)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/codeql-analysis.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index 74664add46..de0b8217da 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -52,7 +52,7 @@ jobs: # Initializes the CodeQL tools for scanning. - name: Initialize CodeQL - uses: github/codeql-action/init@v3 + uses: github/codeql-action/init@v4 with: languages: ${{ matrix.language }} # If you wish to specify custom queries, you can do so here or in a config file. @@ -63,7 +63,7 @@ jobs: # Autobuild attempts to build any compiled languages (C/C++, C#, or Java). # If this step fails, then you should remove it and run the build manually (see below) - name: Autobuild - uses: github/codeql-action/autobuild@v3 + uses: github/codeql-action/autobuild@v4 # ℹ️ Command-line programs to run using the OS shell. # 📚 https://docs.github.com/en/actions/reference/workflow-syntax-for-github-actions @@ -77,4 +77,4 @@ jobs: # make release - name: Perform CodeQL Analysis - uses: github/codeql-action/analyze@v3 + uses: github/codeql-action/analyze@v4 From f99a17bfa5dbd7538d208e80b94aca7c1f80e5c5 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 13 Oct 2025 07:36:17 +0000 Subject: [PATCH 541/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/13)=20(#4917)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/releases.jsonl | 17 ++++++++------- tox.ini | 34 +++++++++++++++-------------- 2 files changed, 27 insertions(+), 24 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index ac5fe1de14..2ff66f2b18 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -16,7 +16,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "", "version": "1.2.19", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "1.3.24", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7", "version": "1.4.54", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.43", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.44", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.49", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.50", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -66,13 +66,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.118.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.119.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.33.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.37.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.42.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.34.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.39.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.43.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} @@ -108,6 +108,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} @@ -152,7 +153,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}} -{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.49.2", "yanked": false}} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.50.0", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}} @@ -200,7 +201,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.3", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} diff --git a/tox.ini b/tox.ini index 5fb05f01bc..f72f05e25a 100644 --- a/tox.ini +++ b/tox.ini @@ -58,9 +58,9 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.18.0 {py3.9,py3.12,py3.13}-google_genai-v1.29.0 - {py3.9,py3.12,py3.13}-google_genai-v1.33.0 - {py3.9,py3.12,py3.13}-google_genai-v1.37.0 - {py3.9,py3.12,py3.13}-google_genai-v1.42.0 + {py3.9,py3.12,py3.13}-google_genai-v1.34.0 + {py3.9,py3.12,py3.13}-google_genai-v1.39.1 + {py3.9,py3.12,py3.13}-google_genai-v1.43.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 @@ -80,6 +80,7 @@ envlist = {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 {py3.9,py3.12,py3.13}-litellm-v1.77.7 + {py3.9,py3.12,py3.13}-litellm-v1.78.0 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 @@ -99,7 +100,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.49 + {py3.9,py3.12,py3.13}-boto3-v1.40.50 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -129,7 +130,7 @@ envlist = {py3.6,py3.8,py3.9}-sqlalchemy-v1.3.24 {py3.6,py3.11,py3.12}-sqlalchemy-v1.4.54 - {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.43 + {py3.7,py3.12,py3.13}-sqlalchemy-v2.0.44 # ~~~ Flags ~~~ @@ -158,7 +159,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.9,py3.12,py3.13}-strawberry-v0.283.2 + {py3.9,py3.12,py3.13}-strawberry-v0.283.3 # ~~~ Network ~~~ @@ -195,7 +196,7 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.3 {py3.9,py3.10}-ray-v2.7.2 - {py3.9,py3.12,py3.13}-ray-v2.49.2 + {py3.9,py3.12,py3.13}-ray-v2.50.0 {py3.6}-rq-v0.8.2 {py3.6,py3.7}-rq-v0.13.0 @@ -227,7 +228,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.118.3 + {py3.8,py3.12,py3.13}-fastapi-v0.119.0 # ~~~ Web 2 ~~~ @@ -354,9 +355,9 @@ deps = cohere-v5.18.0: cohere==5.18.0 google_genai-v1.29.0: google-genai==1.29.0 - google_genai-v1.33.0: google-genai==1.33.0 - google_genai-v1.37.0: google-genai==1.37.0 - google_genai-v1.42.0: google-genai==1.42.0 + google_genai-v1.34.0: google-genai==1.34.0 + google_genai-v1.39.1: google-genai==1.39.1 + google_genai-v1.43.0: google-genai==1.43.0 google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 @@ -386,6 +387,7 @@ deps = langgraph-v1.0.0a4: langgraph==1.0.0a4 litellm-v1.77.7: litellm==1.77.7 + litellm-v1.78.0: litellm==1.78.0 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 @@ -411,7 +413,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.49: boto3==1.40.49 + boto3-v1.40.50: boto3==1.40.50 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -450,7 +452,7 @@ deps = sqlalchemy-v1.3.24: sqlalchemy==1.3.24 sqlalchemy-v1.4.54: sqlalchemy==1.4.54 - sqlalchemy-v2.0.43: sqlalchemy==2.0.43 + sqlalchemy-v2.0.44: sqlalchemy==2.0.44 # ~~~ Flags ~~~ @@ -488,7 +490,7 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.283.2: strawberry-graphql[fastapi,flask]==0.283.2 + strawberry-v0.283.3: strawberry-graphql[fastapi,flask]==0.283.3 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 @@ -543,7 +545,7 @@ deps = huey-v2.5.3: huey==2.5.3 ray-v2.7.2: ray==2.7.2 - ray-v2.49.2: ray==2.49.2 + ray-v2.50.0: ray==2.50.0 rq-v0.8.2: rq==0.8.2 rq-v0.13.0: rq==0.13.0 @@ -615,7 +617,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.92.0: fastapi==0.92.0 fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.118.3: fastapi==0.118.3 + fastapi-v0.119.0: fastapi==0.119.0 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart From 4179fde73a887d24af42b4c249be402ac12c1c15 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Wed, 15 Oct 2025 07:20:35 +0000 Subject: [PATCH 542/591] release: 2.42.0 --- CHANGELOG.md | 11 +++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 14 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8f59545d6b..7366683ef4 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,16 @@ # Changelog +## 2.42.0 + +### Various fixes & improvements + +- ci: 🤖 Update test matrix with new releases (10/13) (#4917) by @github-actions +- build(deps): bump github/codeql-action from 3 to 4 (#4916) by @dependabot +- feat: Add source information for slow outgoing HTTP requests (#4902) by @alexander-alderman-webb +- tests: Update tox (#4913) by @sentrivana +- fix(Ray): Retain the original function name when patching Ray tasks (#4858) by @svartalf +- feat(ai): Add `python-genai` integration (#4891) by @vgrozdanic + ## 2.41.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index b3522a913e..2d54f45170 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.41.0" +release = "2.42.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index c1e587cbeb..2a3c9411be 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1348,4 +1348,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.41.0" +VERSION = "2.42.0" diff --git a/setup.py b/setup.py index c6e391d27a..37c9cf54a6 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.41.0", + version="2.42.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From c79da3f3a8f77ef050641bea91ef6fc14e1ec176 Mon Sep 17 00:00:00 2001 From: Alexander Alderman Webb Date: Wed, 15 Oct 2025 09:26:18 +0200 Subject: [PATCH 543/591] Remove bot commits from changelog --- CHANGELOG.md | 2 -- 1 file changed, 2 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 7366683ef4..7a333567b5 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,8 +4,6 @@ ### Various fixes & improvements -- ci: 🤖 Update test matrix with new releases (10/13) (#4917) by @github-actions -- build(deps): bump github/codeql-action from 3 to 4 (#4916) by @dependabot - feat: Add source information for slow outgoing HTTP requests (#4902) by @alexander-alderman-webb - tests: Update tox (#4913) by @sentrivana - fix(Ray): Retain the original function name when patching Ray tasks (#4858) by @svartalf From cc9ba3d31055a92ab5d3efa8cbaffcbe326c6bb4 Mon Sep 17 00:00:00 2001 From: Alexander Alderman Webb Date: Wed, 15 Oct 2025 09:34:33 +0200 Subject: [PATCH 544/591] Add snippet for enabling google genai to changelog --- CHANGELOG.md | 18 ++++++++++++++++++ 1 file changed, 18 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 7a333567b5..4d1119ddde 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -8,6 +8,24 @@ - tests: Update tox (#4913) by @sentrivana - fix(Ray): Retain the original function name when patching Ray tasks (#4858) by @svartalf - feat(ai): Add `python-genai` integration (#4891) by @vgrozdanic + Enable the new Google GenAI integration with the code snippet below, and you can use the Sentry AI dashboards to observe your AI calls: + + ```python + import sentry_sdk + from sentry_sdk.integrations.google_genai import GoogleGenAIIntegration + sentry_sdk.init( + dsn="", + # Set traces_sample_rate to 1.0 to capture 100% + # of transactions for tracing. + traces_sample_rate=1.0, + # Add data like inputs and responses; + # see https://docs.sentry.io/platforms/python/data-management/data-collected/ for more info + send_default_pii=True, + integrations=[ + GoogleGenAIIntegration(), + ], + ) + ``` ## 2.41.0 From 749e40915abb79597fe298c8190d7981bd30347d Mon Sep 17 00:00:00 2001 From: Saurabh Misra Date: Wed, 15 Oct 2025 00:50:56 -0700 Subject: [PATCH 545/591] =?UTF-8?q?=E2=9A=A1=EF=B8=8F=20Speed=20up=20funct?= =?UTF-8?q?ion=20`=5Fget=5Fdb=5Fspan=5Fdescription`=20(#4924)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Hi all, I am building Codeflash.ai which is an automated performance optimizer for Python codebases. I tried optimizing sentry and found a bunch of great optimizations that I would like to contribute. Would love to collaborate with your team to get them reviewed and merged. Let me know what's the best way to get in touch. #### 📄 44% (0.44x) speedup for ***`_get_db_span_description` in `sentry_sdk/integrations/redis/modules/queries.py`*** ⏱️ Runtime : **`586 microseconds`** **→** **`408 microseconds`** (best of `269` runs) #### 📝 Explanation and details The optimization achieves a **43% speedup** by eliminating redundant function calls inside the loop in `_get_safe_command()`. **Key optimizations applied:** 1. **Cached `should_send_default_pii()` call**: The original code called this function inside the loop for every non-key argument (up to 146 times in profiling). The optimized version calls it once before the loop and stores the result in `send_default_pii`, reducing expensive function calls from O(n) to O(1). 2. **Pre-computed `name.lower()`**: The original code computed `name.lower()` inside the loop for every argument (204 times in profiling). The optimized version computes it once before the loop and reuses the `name_low` variable. **Performance impact from profiling:** - The `should_send_default_pii()` calls dropped from 1.40ms (65.2% of total time) to 625μs (45.9% of total time) - The `name.lower()` calls were eliminated from the loop entirely, removing 99ms of redundant computation - Overall `_get_safe_command` execution time improved from 2.14ms to 1.36ms (36% faster) **Test case patterns where this optimization excels:** - **Multiple arguments**: Commands with many arguments see dramatic improvements (up to 262% faster for large arg lists) - **Large-scale operations**: Tests with 1000+ arguments show 171-223% speedups - **Frequent Redis commands**: Any command processing multiple values benefits significantly The optimization is most effective when processing Redis commands with multiple arguments, which is common in batch operations and complex data manipulations. ✅ **Correctness verification report:** | Test | Status | | --------------------------- | ----------------- | | ⚙️ Existing Unit Tests | 🔘 **None Found** | | 🌀 Generated Regression Tests | ✅ **48 Passed** | | ⏪ Replay Tests | 🔘 **None Found** | | 🔎 Concolic Coverage Tests | 🔘 **None Found** | |📊 Tests Coverage | 100.0% |
🌀 Generated Regression Tests and Runtime ```python import pytest from sentry_sdk.integrations.redis.modules.queries import \ _get_db_span_description _MAX_NUM_ARGS = 10 # Dummy RedisIntegration class for testing class RedisIntegration: def __init__(self, max_data_size=None): self.max_data_size = max_data_size # Dummy should_send_default_pii function for testing _send_pii = False from sentry_sdk.integrations.redis.modules.queries import \ _get_db_span_description # --- Basic Test Cases --- def test_basic_no_args(): """Test command with no arguments.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "PING", ()); desc = codeflash_output # 2.55μs -> 7.76μs (67.2% slower) def test_basic_single_arg_pii_false(): """Test command with one argument, PII off.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("mykey",)); desc = codeflash_output # 3.62μs -> 7.86μs (54.0% slower) def test_basic_single_arg_pii_true(): """Test command with one argument, PII on.""" global _send_pii _send_pii = True integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("mykey",)); desc = codeflash_output # 3.28μs -> 7.40μs (55.7% slower) def test_basic_multiple_args_pii_false(): """Test command with multiple args, PII off.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("mykey", "value1", "value2")); desc = codeflash_output # 12.6μs -> 8.24μs (52.8% faster) def test_basic_multiple_args_pii_true(): """Test command with multiple args, PII on.""" global _send_pii _send_pii = True integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("mykey", "value1", "value2")); desc = codeflash_output # 9.92μs -> 8.47μs (17.0% faster) def test_basic_sensitive_command(): """Test sensitive command: should always filter after command name.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "SET", ("mykey", "secret")); desc = codeflash_output # 7.96μs -> 7.56μs (5.33% faster) def test_basic_sensitive_command_case_insensitive(): """Test sensitive command with different casing.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "set", ("mykey", "secret")); desc = codeflash_output # 7.77μs -> 7.84μs (0.881% slower) def test_basic_max_num_args(): """Test that args beyond _MAX_NUM_ARGS are ignored.""" integration = RedisIntegration() args = tuple(f"arg{i}" for i in range(_MAX_NUM_ARGS + 2)) codeflash_output = _get_db_span_description(integration, "GET", args); desc = codeflash_output # 28.0μs -> 9.43μs (197% faster) # Only up to _MAX_NUM_ARGS+1 args are processed (the first arg is key) expected = "GET 'arg0'" + " [Filtered]" * _MAX_NUM_ARGS # --- Edge Test Cases --- def test_edge_empty_command_name(): """Test with empty command name.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "", ("key",)); desc = codeflash_output # 3.22μs -> 7.46μs (56.9% slower) def test_edge_empty_args(): """Test with empty args tuple.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "DEL", ()); desc = codeflash_output # 2.09μs -> 6.73μs (69.0% slower) def test_edge_none_arg(): """Test with None argument.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", (None,)); desc = codeflash_output # 3.37μs -> 7.57μs (55.5% slower) def test_edge_mixed_types_args(): """Test with mixed argument types.""" integration = RedisIntegration() args = ("key", 123, 45.6, True, None, ["a", "b"], {"x": 1}) codeflash_output = _get_db_span_description(integration, "GET", args); desc = codeflash_output # 19.9μs -> 8.46μs (136% faster) def test_edge_sensitive_command_with_pii_true(): """Sensitive commands should always filter, even if PII is on.""" global _send_pii _send_pii = True integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "AUTH", ("user", "pass")); desc = codeflash_output # 3.40μs -> 7.50μs (54.7% slower) def test_edge_max_data_size_truncation(): """Test truncation when description exceeds max_data_size.""" integration = RedisIntegration(max_data_size=15) codeflash_output = _get_db_span_description(integration, "GET", ("verylongkeyname", "value")); desc = codeflash_output # 9.20μs -> 8.72μs (5.57% faster) # "GET 'verylongkeyname' [Filtered]" is longer than 15 # Truncate to 15-len("...") = 12, then add "..." expected = "GET 'verylo..." def test_edge_max_data_size_exact_length(): """Test truncation when description is exactly max_data_size.""" integration = RedisIntegration(max_data_size=23) codeflash_output = _get_db_span_description(integration, "GET", ("shortkey",)); desc = codeflash_output # 3.33μs -> 7.63μs (56.4% slower) def test_edge_max_data_size_less_than_ellipsis(): """Test when max_data_size is less than length of ellipsis.""" integration = RedisIntegration(max_data_size=2) codeflash_output = _get_db_span_description(integration, "GET", ("key",)); desc = codeflash_output # 4.07μs -> 8.65μs (52.9% slower) def test_edge_args_are_empty_strings(): """Test when args are empty strings.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("", "")); desc = codeflash_output # 8.52μs -> 7.74μs (10.1% faster) def test_edge_command_name_is_space(): """Test when command name is a space.""" integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, " ", ("key",)); desc = codeflash_output # 3.09μs -> 7.34μs (57.9% slower) # --- Large Scale Test Cases --- def test_large_many_args_pii_false(): """Test with a large number of arguments, PII off.""" integration = RedisIntegration() args = tuple(f"arg{i}" for i in range(1000)) codeflash_output = _get_db_span_description(integration, "GET", args); desc = codeflash_output # 32.3μs -> 10.3μs (213% faster) # Only first arg shown, rest are filtered, up to _MAX_NUM_ARGS expected = "GET 'arg0'" + " [Filtered]" * min(len(args)-1, _MAX_NUM_ARGS) def test_large_many_args_pii_true(): """Test with a large number of arguments, PII on.""" global _send_pii _send_pii = True integration = RedisIntegration() args = tuple(f"arg{i}" for i in range(1000)) # Only up to _MAX_NUM_ARGS are processed expected = "GET " + " ".join([repr(f"arg{i}") for i in range(_MAX_NUM_ARGS+1)]) codeflash_output = _get_db_span_description(integration, "GET", args); desc = codeflash_output # 28.1μs -> 9.55μs (194% faster) def test_large_long_command_name_and_args(): """Test with very long command name and args.""" integration = RedisIntegration() cmd = "LONGCOMMAND" * 10 args = tuple("X"*100 for _ in range(_MAX_NUM_ARGS+1)) expected = cmd + " " + " ".join([repr("X"*100) if i == 0 else "[Filtered]" for i in range(_MAX_NUM_ARGS+1)]) codeflash_output = _get_db_span_description(integration, cmd, args); desc = codeflash_output # 34.2μs -> 9.45μs (262% faster) def test_large_truncation(): """Test truncation with very large description.""" integration = RedisIntegration(max_data_size=50) args = tuple("X"*20 for _ in range(_MAX_NUM_ARGS+1)) codeflash_output = _get_db_span_description(integration, "GET", args); desc = codeflash_output # 28.3μs -> 10.0μs (182% faster) def test_large_sensitive_command(): """Test large sensitive command, all args filtered.""" integration = RedisIntegration() args = tuple(f"secret{i}" for i in range(1000)) codeflash_output = _get_db_span_description(integration, "SET", args); desc = codeflash_output # 28.0μs -> 10.1μs (178% faster) # Only up to _MAX_NUM_ARGS+1 args are processed, all filtered expected = "SET" + " [Filtered]" * (_MAX_NUM_ARGS+1) # codeflash_output is used to check that the output of the original code is the same as that of the optimized code. #------------------------------------------------ import pytest # used for our unit tests from sentry_sdk.integrations.redis.modules.queries import \ _get_db_span_description _MAX_NUM_ARGS = 10 # Minimal RedisIntegration stub for testing class RedisIntegration: def __init__(self, max_data_size=None): self.max_data_size = max_data_size # Minimal Scope and client stub for should_send_default_pii class ClientStub: def __init__(self, send_pii): self._send_pii = send_pii def should_send_default_pii(self): return self._send_pii class Scope: _client = ClientStub(send_pii=False) @classmethod def get_client(cls): return cls._client def should_send_default_pii(): return Scope.get_client().should_send_default_pii() from sentry_sdk.integrations.redis.modules.queries import \ _get_db_span_description # --- Begin: Unit Tests --- # 1. Basic Test Cases def test_basic_single_arg_no_pii(): # Test a simple command with one argument, PII disabled Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("mykey",)); result = codeflash_output # 3.46μs -> 7.84μs (55.9% slower) def test_basic_multiple_args_no_pii(): # Test a command with multiple arguments, PII disabled Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "SET", ("mykey", "myvalue")); result = codeflash_output # 8.35μs -> 8.05μs (3.70% faster) def test_basic_multiple_args_with_pii(): # Test a command with multiple arguments, PII enabled Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "SET", ("mykey", "myvalue")); result = codeflash_output # 7.97μs -> 7.63μs (4.39% faster) def test_basic_sensitive_command(): # Test a sensitive command, should always be filtered Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "AUTH", ("user", "password")); result = codeflash_output # 3.40μs -> 7.46μs (54.4% slower) def test_basic_no_args(): # Test a command with no arguments Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "PING", ()); result = codeflash_output # 2.16μs -> 6.63μs (67.4% slower) # 2. Edge Test Cases def test_edge_max_num_args(): # Test with more than _MAX_NUM_ARGS arguments, should truncate at _MAX_NUM_ARGS Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = tuple(f"arg{i}" for i in range(_MAX_NUM_ARGS + 2)) codeflash_output = _get_db_span_description(integration, "SET", args); result = codeflash_output # 32.4μs -> 9.05μs (258% faster) # Only up to _MAX_NUM_ARGS should be included expected = "SET " + " ".join( [repr(args[0])] + [repr(arg) for arg in args[1:_MAX_NUM_ARGS+1]] ) def test_edge_empty_string_key(): # Test with an empty string as key Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", ("",)); result = codeflash_output # 3.42μs -> 7.51μs (54.5% slower) def test_edge_none_key(): # Test with None as key Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", (None,)); result = codeflash_output # 3.25μs -> 7.42μs (56.2% slower) def test_edge_non_string_key(): # Test with integer as key Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", (12345,)); result = codeflash_output # 3.24μs -> 7.62μs (57.5% slower) def test_edge_sensitive_command_case_insensitive(): # Test sensitive command with mixed case Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "AuTh", ("user", "password")); result = codeflash_output # 3.57μs -> 7.72μs (53.8% slower) def test_edge_truncation_exact(): # Test truncation where description is exactly max_data_size Scope._client = ClientStub(send_pii=True) integration = RedisIntegration(max_data_size=13) codeflash_output = _get_db_span_description(integration, "GET", ("mykey",)); result = codeflash_output # 3.61μs -> 8.05μs (55.1% slower) def test_edge_truncation_needed(): # Test truncation where description exceeds max_data_size Scope._client = ClientStub(send_pii=True) integration = RedisIntegration(max_data_size=10) codeflash_output = _get_db_span_description(integration, "GET", ("mykey",)); result = codeflash_output # 4.32μs -> 7.96μs (45.8% slower) def test_edge_truncation_with_filtered(): # Truncation with filtered data Scope._client = ClientStub(send_pii=False) integration = RedisIntegration(max_data_size=10) codeflash_output = _get_db_span_description(integration, "SET", ("mykey", "myvalue")); result = codeflash_output # 10.3μs -> 8.92μs (15.7% faster) def test_edge_args_are_bytes(): # Test arguments are bytes Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "GET", (b"mykey",)); result = codeflash_output # 3.42μs -> 7.54μs (54.7% slower) def test_edge_args_are_mixed_types(): # Test arguments are mixed types Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = ("key", 123, None, b"bytes") codeflash_output = _get_db_span_description(integration, "SET", args); result = codeflash_output # 13.7μs -> 8.31μs (65.1% faster) expected = "SET 'key' 123 None b'bytes'" def test_edge_args_are_empty_tuple(): # Test arguments is empty tuple Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "PING", ()); result = codeflash_output # 2.14μs -> 6.67μs (67.9% slower) def test_edge_args_are_list(): # Test arguments as a list (should still work as sequence) Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() codeflash_output = _get_db_span_description(integration, "SET", ["key", "val"]); result = codeflash_output # 8.54μs -> 7.96μs (7.30% faster) def test_edge_args_are_dict(): # Test arguments as a dict (should treat as sequence of keys) Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = {"a": 1, "b": 2} codeflash_output = _get_db_span_description(integration, "SET", args); result = codeflash_output # 7.87μs -> 7.86μs (0.102% faster) def test_edge_args_are_long_string(): # Test argument is a very long string (truncation) Scope._client = ClientStub(send_pii=True) integration = RedisIntegration(max_data_size=20) long_str = "x" * 100 codeflash_output = _get_db_span_description(integration, "SET", (long_str,)); result = codeflash_output # 4.46μs -> 8.43μs (47.1% slower) # 3. Large Scale Test Cases def test_large_many_args_no_pii(): # Test with large number of arguments, PII disabled Scope._client = ClientStub(send_pii=False) integration = RedisIntegration() args = tuple(f"key{i}" for i in range(999)) codeflash_output = _get_db_span_description(integration, "MGET", args); result = codeflash_output # 28.6μs -> 10.6μs (171% faster) # Only first is shown, rest are filtered (up to _MAX_NUM_ARGS) expected = "MGET 'key0'" + " [Filtered]" * _MAX_NUM_ARGS def test_large_many_args_with_pii(): # Test with large number of arguments, PII enabled Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = tuple(f"key{i}" for i in range(999)) codeflash_output = _get_db_span_description(integration, "MGET", args); result = codeflash_output # 30.9μs -> 9.87μs (213% faster) # Only up to _MAX_NUM_ARGS are shown expected = "MGET " + " ".join([repr(arg) for arg in args[:_MAX_NUM_ARGS+1]]) def test_large_truncation(): # Test truncation with large description Scope._client = ClientStub(send_pii=True) integration = RedisIntegration(max_data_size=50) args = tuple("x" * 10 for _ in range(20)) codeflash_output = _get_db_span_description(integration, "MGET", args); result = codeflash_output # 31.0μs -> 10.4μs (198% faster) def test_large_sensitive_command(): # Test large sensitive command, should always be filtered Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = tuple("x" * 10 for _ in range(20)) codeflash_output = _get_db_span_description(integration, "AUTH", args); result = codeflash_output # 5.42μs -> 9.30μs (41.8% slower) def test_large_args_are_large_numbers(): # Test with large integer arguments Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = tuple(10**6 + i for i in range(_MAX_NUM_ARGS + 1)) codeflash_output = _get_db_span_description(integration, "MGET", args); result = codeflash_output # 27.6μs -> 9.38μs (194% faster) expected = "MGET " + " ".join([repr(arg) for arg in args[:_MAX_NUM_ARGS+1]]) def test_large_args_are_large_bytes(): # Test with large bytes arguments Scope._client = ClientStub(send_pii=True) integration = RedisIntegration() args = tuple(b"x" * 100 for _ in range(_MAX_NUM_ARGS + 1)) codeflash_output = _get_db_span_description(integration, "MGET", args); result = codeflash_output # 30.2μs -> 9.35μs (223% faster) expected = "MGET " + " ".join([repr(arg) for arg in args[:_MAX_NUM_ARGS+1]]) # codeflash_output is used to check that the output of the original code is the same as that of the optimized code. ```
To edit these changes `git checkout codeflash/optimize-_get_db_span_description-mg9vzvxu` and push. [![Codeflash](https://img.shields.io/badge/Optimized%20with-Codeflash-yellow?style=flat&color=%23ffc428&logo=data:image/svg+xml;base64,PHN2ZyB3aWR0aD0iNDgwIiBoZWlnaHQ9ImF1dG8iIHZpZXdCb3g9IjAgMCA0ODAgMjgwIiBmaWxsPSJub25lIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPgo8cGF0aCBmaWxsLXJ1bGU9ImV2ZW5vZGQiIGNsaXAtcnVsZT0iZXZlbm9kZCIgZD0iTTI4Ni43IDAuMzc4NDE4SDIwMS43NTFMNTAuOTAxIDE0OC45MTFIMTM1Ljg1MUwwLjk2MDkzOCAyODEuOTk5SDk1LjQzNTJMMjgyLjMyNCA4OS45NjE2SDE5Ni4zNDVMMjg2LjcgMC4zNzg0MThaIiBmaWxsPSIjRkZDMDQzIi8+CjxwYXRoIGZpbGwtcnVsZT0iZXZlbm9kZCIgY2xpcC1ydWxlPSJldmVub2RkIiBkPSJNMzExLjYwNyAwLjM3ODkwNkwyNTguNTc4IDU0Ljk1MjZIMzc5LjU2N0w0MzIuMzM5IDAuMzc4OTA2SDMxMS42MDdaIiBmaWxsPSIjMEIwQTBBIi8+CjxwYXRoIGZpbGwtcnVsZT0iZXZlbm9kZCIgY2xpcC1ydWxlPSJldmVub2RkIiBkPSJNMzA5LjU0NyA4OS45NjAxTDI1Ni41MTggMTQ0LjI3NkgzNzcuNTA2TDQzMC4wMjEgODkuNzAyNkgzMDkuNTQ3Vjg5Ljk2MDFaIiBmaWxsPSIjMEIwQTBBIi8+CjxwYXRoIGZpbGwtcnVsZT0iZXZlbm9kZCIgY2xpcC1ydWxlPSJldmVub2RkIiBkPSJNMjQyLjg3MyAxNjQuNjZMMTg5Ljg0NCAyMTkuMjM0SDMxMC44MzNMMzYzLjM0NyAxNjQuNjZIMjQyLjg3M1oiIGZpbGw9IiMwQjBBMEEiLz4KPC9zdmc+Cg==)](https://codeflash.ai) Co-authored-by: codeflash-ai[bot] <148906541+codeflash-ai[bot]@users.noreply.github.com> --- sentry_sdk/integrations/redis/utils.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/redis/utils.py b/sentry_sdk/integrations/redis/utils.py index cf230f6648..7bb73f3372 100644 --- a/sentry_sdk/integrations/redis/utils.py +++ b/sentry_sdk/integrations/redis/utils.py @@ -20,12 +20,13 @@ def _get_safe_command(name, args): # type: (str, Sequence[Any]) -> str command_parts = [name] + name_low = name.lower() + send_default_pii = should_send_default_pii() + for i, arg in enumerate(args): if i > _MAX_NUM_ARGS: break - name_low = name.lower() - if name_low in _COMMANDS_INCLUDING_SENSITIVE_DATA: command_parts.append(SENSITIVE_DATA_SUBSTITUTE) continue @@ -33,9 +34,8 @@ def _get_safe_command(name, args): arg_is_the_key = i == 0 if arg_is_the_key: command_parts.append(repr(arg)) - else: - if should_send_default_pii(): + if send_default_pii: command_parts.append(repr(arg)) else: command_parts.append(SENSITIVE_DATA_SUBSTITUTE) From a311e3b4d4f75e696c388352158c9a38a8718e21 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Wed, 15 Oct 2025 13:42:07 +0200 Subject: [PATCH 546/591] Generalize NOT_GIVEN check with omit for openai (#4926) ### Description openai uses `Omit` now instead of `NotGiven` https://github.com/openai/openai-python/commit/82602884b61ef2f407f4c5f4fcae7d07243897be #### Issues * resolves: #4923 * resolves: PY-1885 --- sentry_sdk/integrations/openai.py | 28 +++++++++++++++++++----- tests/integrations/openai/test_openai.py | 7 +++++- 2 files changed, 29 insertions(+), 6 deletions(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index e9bd2efa23..19d7717b3c 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -1,4 +1,5 @@ from functools import wraps +from collections.abc import Iterable import sentry_sdk from sentry_sdk import consts @@ -17,14 +18,19 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Any, Iterable, List, Optional, Callable, AsyncIterator, Iterator + from typing import Any, List, Optional, Callable, AsyncIterator, Iterator from sentry_sdk.tracing import Span try: try: - from openai import NOT_GIVEN + from openai import NotGiven except ImportError: - NOT_GIVEN = None + NotGiven = None + + try: + from openai import Omit + except ImportError: + Omit = None from openai.resources.chat.completions import Completions, AsyncCompletions from openai.resources import Embeddings, AsyncEmbeddings @@ -204,12 +210,12 @@ def _set_input_data(span, kwargs, operation, integration): for key, attribute in kwargs_keys_to_attributes.items(): value = kwargs.get(key) - if value is not NOT_GIVEN and value is not None: + if value is not None and _is_given(value): set_data_normalized(span, attribute, value) # Input attributes: Tools tools = kwargs.get("tools") - if tools is not NOT_GIVEN and tools is not None and len(tools) > 0: + if tools is not None and _is_given(tools) and len(tools) > 0: set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) ) @@ -689,3 +695,15 @@ async def _sentry_patched_responses_async(*args, **kwargs): return await _execute_async(f, *args, **kwargs) return _sentry_patched_responses_async + + +def _is_given(obj): + # type: (Any) -> bool + """ + Check for givenness safely across different openai versions. + """ + if NotGiven is not None and isinstance(obj, NotGiven): + return False + if Omit is not None and isinstance(obj, Omit): + return False + return True diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 06e0a09fcf..276a1b4886 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -7,6 +7,11 @@ except ImportError: NOT_GIVEN = None +try: + from openai import omit +except ImportError: + omit = None + from openai import AsyncOpenAI, OpenAI, AsyncStream, Stream, OpenAIError from openai.types import CompletionUsage, CreateEmbeddingResponse, Embedding from openai.types.chat import ChatCompletion, ChatCompletionMessage, ChatCompletionChunk @@ -1424,7 +1429,7 @@ async def test_streaming_responses_api_async( ) @pytest.mark.parametrize( "tools", - [[], None, NOT_GIVEN], + [[], None, NOT_GIVEN, omit], ) def test_empty_tools_in_chat_completion(sentry_init, capture_events, tools): sentry_init( From 23411e57cfa78ce4b949d24c458a709ce8b902d7 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 15 Oct 2025 15:04:34 +0200 Subject: [PATCH 547/591] fix(litellm): Classify embeddings correctly (#4918) Check the `call_type` value to distinguish embeddings from chats. The `client` decorator sets `call_type` by introspecting the function name and wraps all of the top-level `litellm` functions. If users import from `litellm.llms`, embedding calls still may appear as chats, but the input callback we provide does not have enough information in that case. Closes https://github.com/getsentry/sentry-python/issues/4908 --- sentry_sdk/integrations/litellm.py | 8 ++++++-- tests/integrations/litellm/test_litellm.py | 5 +++-- 2 files changed, 9 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/litellm.py b/sentry_sdk/integrations/litellm.py index 2582c2bc05..1f047b1c1d 100644 --- a/sentry_sdk/integrations/litellm.py +++ b/sentry_sdk/integrations/litellm.py @@ -48,8 +48,11 @@ def _input_callback(kwargs): model = full_model provider = "unknown" - messages = kwargs.get("messages", []) - operation = "chat" if messages else "embeddings" + call_type = kwargs.get("call_type", None) + if call_type == "embedding": + operation = "embeddings" + else: + operation = "chat" # Start a new span/transaction span = get_start_span_function()( @@ -71,6 +74,7 @@ def _input_callback(kwargs): set_data_normalized(span, SPANDATA.GEN_AI_OPERATION_NAME, operation) # Record messages if allowed + messages = kwargs.get("messages", []) if messages and should_send_default_pii() and integration.include_prompts: set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False diff --git a/tests/integrations/litellm/test_litellm.py b/tests/integrations/litellm/test_litellm.py index b600c32905..19ae206c85 100644 --- a/tests/integrations/litellm/test_litellm.py +++ b/tests/integrations/litellm/test_litellm.py @@ -208,14 +208,15 @@ def test_embeddings_create(sentry_init, capture_events): ) events = capture_events() + messages = [{"role": "user", "content": "Some text to test embeddings"}] mock_response = MockEmbeddingResponse() with start_transaction(name="litellm test"): - # For embeddings, messages would be empty kwargs = { "model": "text-embedding-ada-002", "input": "Hello!", - "messages": [], # Empty for embeddings + "messages": messages, + "call_type": "embedding", } _input_callback(kwargs) From 43258029347740c585cab8850d05452d5e2fb4bd Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Wed, 15 Oct 2025 15:17:07 +0200 Subject: [PATCH 548/591] Handle ValueError in scope resets (#4928) ### Description when async generators throw a `GeneratorExit` we end up with ``` ValueError: at 0x7f04ceb17340> was created in a different Context ``` so just catch that and rely on GC to cleanup the contextvar since we can't be smarter than that anyway for this case. #### Issues * resolves: #4925 * resolves: PY-1886 --- sentry_sdk/scope.py | 12 ++++++------ tests/test_scope.py | 14 ++++++++------ 2 files changed, 14 insertions(+), 12 deletions(-) diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index c871e6a467..f9caf7e1d6 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -1679,7 +1679,7 @@ def new_scope(): try: # restore original scope _current_scope.reset(token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) @@ -1717,7 +1717,7 @@ def use_scope(scope): try: # restore original scope _current_scope.reset(token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) @@ -1761,12 +1761,12 @@ def isolation_scope(): # restore original scopes try: _current_scope.reset(current_token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) try: _isolation_scope.reset(isolation_token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) @@ -1808,12 +1808,12 @@ def use_isolation_scope(isolation_scope): # restore original scopes try: _current_scope.reset(current_token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) try: _isolation_scope.reset(isolation_token) - except LookupError: + except (LookupError, ValueError): capture_internal_exception(sys.exc_info()) diff --git a/tests/test_scope.py b/tests/test_scope.py index e645d84234..68c93f3036 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -908,6 +908,7 @@ def test_last_event_id_cleared(sentry_init): @pytest.mark.tests_internal_exceptions +@pytest.mark.parametrize("error_cls", [LookupError, ValueError]) @pytest.mark.parametrize( "scope_manager", [ @@ -915,10 +916,10 @@ def test_last_event_id_cleared(sentry_init): use_scope, ], ) -def test_handle_lookup_error_on_token_reset_current_scope(scope_manager): +def test_handle_error_on_token_reset_current_scope(error_cls, scope_manager): with mock.patch("sentry_sdk.scope.capture_internal_exception") as mock_capture: with mock.patch("sentry_sdk.scope._current_scope") as mock_token_var: - mock_token_var.reset.side_effect = LookupError() + mock_token_var.reset.side_effect = error_cls() mock_token = mock.Mock() mock_token_var.set.return_value = mock_token @@ -932,13 +933,14 @@ def test_handle_lookup_error_on_token_reset_current_scope(scope_manager): pass except Exception: - pytest.fail("Context manager should handle LookupError gracefully") + pytest.fail(f"Context manager should handle {error_cls} gracefully") mock_capture.assert_called_once() mock_token_var.reset.assert_called_once_with(mock_token) @pytest.mark.tests_internal_exceptions +@pytest.mark.parametrize("error_cls", [LookupError, ValueError]) @pytest.mark.parametrize( "scope_manager", [ @@ -946,13 +948,13 @@ def test_handle_lookup_error_on_token_reset_current_scope(scope_manager): use_isolation_scope, ], ) -def test_handle_lookup_error_on_token_reset_isolation_scope(scope_manager): +def test_handle_error_on_token_reset_isolation_scope(error_cls, scope_manager): with mock.patch("sentry_sdk.scope.capture_internal_exception") as mock_capture: with mock.patch("sentry_sdk.scope._current_scope") as mock_current_scope: with mock.patch( "sentry_sdk.scope._isolation_scope" ) as mock_isolation_scope: - mock_isolation_scope.reset.side_effect = LookupError() + mock_isolation_scope.reset.side_effect = error_cls() mock_current_token = mock.Mock() mock_current_scope.set.return_value = mock_current_token @@ -965,7 +967,7 @@ def test_handle_lookup_error_on_token_reset_isolation_scope(scope_manager): pass except Exception: - pytest.fail("Context manager should handle LookupError gracefully") + pytest.fail(f"Context manager should handle {error_cls} gracefully") mock_capture.assert_called_once() mock_current_scope.reset.assert_called_once_with(mock_current_token) From d21fabd6fd22f38025df3258938b961c87c18a13 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 16 Oct 2025 12:48:40 +0000 Subject: [PATCH 549/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/16)=20(#4945)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> --- scripts/populate_tox/releases.jsonl | 16 +++++++-------- tox.ini | 32 ++++++++++++++--------------- 2 files changed, 24 insertions(+), 24 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 2ff66f2b18..537b3a64e8 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -26,7 +26,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.34.2", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.52.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.69.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.70.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} @@ -46,7 +46,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.50", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.53", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -55,10 +55,10 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.13.12", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.18.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.10.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.15.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.19.0", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.9.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": "", "version": "1.4.1", "yanked": false}} @@ -72,7 +72,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.34.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.39.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.43.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.45.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} @@ -99,7 +99,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc5", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc6", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} @@ -108,7 +108,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.0", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} diff --git a/tox.ini b/tox.ini index f72f05e25a..668e5888b4 100644 --- a/tox.ini +++ b/tox.ini @@ -50,23 +50,23 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.34.2 {py3.8,py3.11,py3.12}-anthropic-v0.52.2 - {py3.8,py3.12,py3.13}-anthropic-v0.69.0 + {py3.8,py3.12,py3.13}-anthropic-v0.70.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 - {py3.9,py3.11,py3.12}-cohere-v5.9.4 - {py3.9,py3.11,py3.12}-cohere-v5.13.12 - {py3.9,py3.11,py3.12}-cohere-v5.18.0 + {py3.9,py3.11,py3.12}-cohere-v5.10.0 + {py3.9,py3.11,py3.12}-cohere-v5.15.0 + {py3.9,py3.11,py3.12}-cohere-v5.19.0 {py3.9,py3.12,py3.13}-google_genai-v1.29.0 {py3.9,py3.12,py3.13}-google_genai-v1.34.0 {py3.9,py3.12,py3.13}-google_genai-v1.39.1 - {py3.9,py3.12,py3.13}-google_genai-v1.43.0 + {py3.9,py3.12,py3.13}-google_genai-v1.45.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 - {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc5 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc6 {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.11,py3.12}-langchain-base-v0.2.17 @@ -80,7 +80,7 @@ envlist = {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 {py3.9,py3.12,py3.13}-litellm-v1.77.7 - {py3.9,py3.12,py3.13}-litellm-v1.78.0 + {py3.9,py3.12,py3.13}-litellm-v1.78.2 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 @@ -100,7 +100,7 @@ envlist = {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.50 + {py3.9,py3.12,py3.13}-boto3-v1.40.53 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -344,27 +344,27 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.34.2: anthropic==0.34.2 anthropic-v0.52.2: anthropic==0.52.2 - anthropic-v0.69.0: anthropic==0.69.0 + anthropic-v0.70.0: anthropic==0.70.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.34.2: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 - cohere-v5.9.4: cohere==5.9.4 - cohere-v5.13.12: cohere==5.13.12 - cohere-v5.18.0: cohere==5.18.0 + cohere-v5.10.0: cohere==5.10.0 + cohere-v5.15.0: cohere==5.15.0 + cohere-v5.19.0: cohere==5.19.0 google_genai-v1.29.0: google-genai==1.29.0 google_genai-v1.34.0: google-genai==1.34.0 google_genai-v1.39.1: google-genai==1.39.1 - google_genai-v1.43.0: google-genai==1.43.0 + google_genai-v1.45.0: google-genai==1.45.0 google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 huggingface_hub-v0.35.3: huggingface_hub==0.35.3 - huggingface_hub-v1.0.0rc5: huggingface_hub==1.0.0rc5 + huggingface_hub-v1.0.0rc6: huggingface_hub==1.0.0rc6 huggingface_hub: responses huggingface_hub: pytest-httpx @@ -387,7 +387,7 @@ deps = langgraph-v1.0.0a4: langgraph==1.0.0a4 litellm-v1.77.7: litellm==1.77.7 - litellm-v1.78.0: litellm==1.78.0 + litellm-v1.78.2: litellm==1.78.2 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 @@ -413,7 +413,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.50: boto3==1.40.50 + boto3-v1.40.53: boto3==1.40.53 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 From ca4df942041810c397d44028e5fec8e6c570b101 Mon Sep 17 00:00:00 2001 From: Jason Cameron Date: Thu, 16 Oct 2025 09:21:20 -0400 Subject: [PATCH 550/591] fix(openai): Use non-deprecated Pydantic method to extract response text (#4942) Switch to Pydantic v2's `model_dump()` instead of `dict()` for serialization. The change avoids deprecation warnings during OpenAI response parsing that created issues in Sentry. --- sentry_sdk/integrations/openai.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 19d7717b3c..315d54f750 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -237,7 +237,7 @@ def _set_output_data(span, response, kwargs, integration, finish_span=True): if hasattr(response, "choices"): if should_send_default_pii() and integration.include_prompts: - response_text = [choice.message.dict() for choice in response.choices] + response_text = [choice.message.model_dump() for choice in response.choices] if len(response_text) > 0: set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, response_text) From 814cd5a0bc8700478526796da53fd9217223d042 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 16 Oct 2025 15:26:43 +0200 Subject: [PATCH 551/591] fix(ai): introduce message truncation for openai (#4946) --- sentry_sdk/ai/utils.py | 53 +++- sentry_sdk/client.py | 19 ++ sentry_sdk/integrations/openai.py | 18 +- sentry_sdk/scope.py | 5 + tests/integrations/openai/test_openai.py | 63 ++++- tests/test_ai_monitoring.py | 300 +++++++++++++++++++++++ 6 files changed, 445 insertions(+), 13 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index 0c0b937006..1fb291bdac 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -1,14 +1,18 @@ import json - +from collections import deque from typing import TYPE_CHECKING +from sys import getsizeof if TYPE_CHECKING: - from typing import Any, Callable + from typing import Any, Callable, Dict, List, Optional, Tuple + from sentry_sdk.tracing import Span import sentry_sdk from sentry_sdk.utils import logger +MAX_GEN_AI_MESSAGE_BYTES = 20_000 # 20KB + class GEN_AI_ALLOWED_MESSAGE_ROLES: SYSTEM = "system" @@ -95,3 +99,48 @@ def get_start_span_function(): current_span is not None and current_span.containing_transaction is not None ) return sentry_sdk.start_span if transaction_exists else sentry_sdk.start_transaction + + +def _find_truncation_index(messages, max_bytes): + # type: (List[Dict[str, Any]], int) -> int + """ + Find the index of the first message that would exceed the max bytes limit. + Compute the individual message sizes, and return the index of the first message from the back + of the list that would exceed the max bytes limit. + """ + running_sum = 0 + for idx in range(len(messages) - 1, -1, -1): + size = len(json.dumps(messages[idx], separators=(",", ":")).encode("utf-8")) + running_sum += size + if running_sum > max_bytes: + return idx + 1 + + return 0 + + +def truncate_messages_by_size(messages, max_bytes=MAX_GEN_AI_MESSAGE_BYTES): + # type: (List[Dict[str, Any]], int) -> Tuple[List[Dict[str, Any]], int] + serialized_json = json.dumps(messages, separators=(",", ":")) + current_size = len(serialized_json.encode("utf-8")) + + if current_size <= max_bytes: + return messages, 0 + + truncation_index = _find_truncation_index(messages, max_bytes) + return messages[truncation_index:], truncation_index + + +def truncate_and_annotate_messages( + messages, span, scope, max_bytes=MAX_GEN_AI_MESSAGE_BYTES +): + # type: (Optional[List[Dict[str, Any]]], Any, Any, int) -> Optional[List[Dict[str, Any]]] + if not messages: + return None + + truncated_messages, removed_count = truncate_messages_by_size(messages, max_bytes) + if removed_count > 0: + scope._gen_ai_messages_truncated[span.span_id] = len(messages) - len( + truncated_messages + ) + + return truncated_messages diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index d17f922642..ffd899b545 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -598,6 +598,24 @@ def _prepare_event( if event_scrubber: event_scrubber.scrub_event(event) + if scope is not None and scope._gen_ai_messages_truncated: + spans = event.get("spans", []) # type: List[Dict[str, Any]] | AnnotatedValue + if isinstance(spans, list): + for span in spans: + span_id = span.get("span_id", None) + span_data = span.get("data", {}) + if ( + span_id + and span_id in scope._gen_ai_messages_truncated + and SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data + ): + span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] = AnnotatedValue( + span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES], + { + "len": scope._gen_ai_messages_truncated[span_id] + + len(span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES]) + }, + ) if previous_total_spans is not None: event["spans"] = AnnotatedValue( event.get("spans", []), {"len": previous_total_spans} @@ -606,6 +624,7 @@ def _prepare_event( event["breadcrumbs"] = AnnotatedValue( event.get("breadcrumbs", []), {"len": previous_total_breadcrumbs} ) + # Postprocess the event here so that annotated types do # generally not surface in before_send if event is not None: diff --git a/sentry_sdk/integrations/openai.py b/sentry_sdk/integrations/openai.py index 315d54f750..bb93341f35 100644 --- a/sentry_sdk/integrations/openai.py +++ b/sentry_sdk/integrations/openai.py @@ -1,10 +1,13 @@ from functools import wraps -from collections.abc import Iterable import sentry_sdk from sentry_sdk import consts from sentry_sdk.ai.monitoring import record_token_usage -from sentry_sdk.ai.utils import set_data_normalized, normalize_message_roles +from sentry_sdk.ai.utils import ( + set_data_normalized, + normalize_message_roles, + truncate_and_annotate_messages, +) from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -18,7 +21,7 @@ from typing import TYPE_CHECKING if TYPE_CHECKING: - from typing import Any, List, Optional, Callable, AsyncIterator, Iterator + from typing import Any, Iterable, List, Optional, Callable, AsyncIterator, Iterator from sentry_sdk.tracing import Span try: @@ -189,9 +192,12 @@ def _set_input_data(span, kwargs, operation, integration): and integration.include_prompts ): normalized_messages = normalize_message_roles(messages) - set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, normalized_messages, unpack=False - ) + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages(normalized_messages, span, scope) + if messages_data is not None: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages_data, unpack=False + ) # Input attributes: Common set_data_normalized(span, SPANDATA.GEN_AI_SYSTEM, "openai") diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index f9caf7e1d6..5815a65440 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -188,6 +188,7 @@ class Scope: "_extras", "_breadcrumbs", "_n_breadcrumbs_truncated", + "_gen_ai_messages_truncated", "_event_processors", "_error_processors", "_should_capture", @@ -213,6 +214,7 @@ def __init__(self, ty=None, client=None): self._name = None # type: Optional[str] self._propagation_context = None # type: Optional[PropagationContext] self._n_breadcrumbs_truncated = 0 # type: int + self._gen_ai_messages_truncated = {} # type: Dict[str, int] self.client = NonRecordingClient() # type: sentry_sdk.client.BaseClient @@ -247,6 +249,7 @@ def __copy__(self): rv._breadcrumbs = copy(self._breadcrumbs) rv._n_breadcrumbs_truncated = self._n_breadcrumbs_truncated + rv._gen_ai_messages_truncated = self._gen_ai_messages_truncated.copy() rv._event_processors = self._event_processors.copy() rv._error_processors = self._error_processors.copy() rv._propagation_context = self._propagation_context @@ -1583,6 +1586,8 @@ def update_from_scope(self, scope): self._n_breadcrumbs_truncated = ( self._n_breadcrumbs_truncated + scope._n_breadcrumbs_truncated ) + if scope._gen_ai_messages_truncated: + self._gen_ai_messages_truncated.update(scope._gen_ai_messages_truncated) if scope._span: self._span = scope._span if scope._attachments: diff --git a/tests/integrations/openai/test_openai.py b/tests/integrations/openai/test_openai.py index 276a1b4886..ccef4f336e 100644 --- a/tests/integrations/openai/test_openai.py +++ b/tests/integrations/openai/test_openai.py @@ -1,3 +1,4 @@ +import json import pytest from sentry_sdk.utils import package_version @@ -6,7 +7,6 @@ from openai import NOT_GIVEN except ImportError: NOT_GIVEN = None - try: from openai import omit except ImportError: @@ -44,6 +44,9 @@ OpenAIIntegration, _calculate_token_usage, ) +from sentry_sdk.ai.utils import MAX_GEN_AI_MESSAGE_BYTES +from sentry_sdk._types import AnnotatedValue +from sentry_sdk.serializer import serialize from unittest import mock # python 3.3 and above @@ -1456,6 +1459,7 @@ def test_empty_tools_in_chat_completion(sentry_init, capture_events, tools): def test_openai_message_role_mapping(sentry_init, capture_events): """Test that OpenAI integration properly maps message roles like 'ai' to 'assistant'""" + sentry_init( integrations=[OpenAIIntegration(include_prompts=True)], traces_sample_rate=1.0, @@ -1465,7 +1469,6 @@ def test_openai_message_role_mapping(sentry_init, capture_events): client = OpenAI(api_key="z") client.chat.completions._post = mock.Mock(return_value=EXAMPLE_CHAT_COMPLETION) - # Test messages with mixed roles including "ai" that should be mapped to "assistant" test_messages = [ {"role": "system", "content": "You are helpful."}, @@ -1476,11 +1479,9 @@ def test_openai_message_role_mapping(sentry_init, capture_events): with start_transaction(name="openai tx"): client.chat.completions.create(model="test-model", messages=test_messages) - + # Verify that the span was created correctly (event,) = events span = event["spans"][0] - - # Verify that the span was created correctly assert span["op"] == "gen_ai.chat" assert SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"] @@ -1505,3 +1506,55 @@ def test_openai_message_role_mapping(sentry_init, capture_events): # Verify no "ai" roles remain roles = [msg["role"] for msg in stored_messages] assert "ai" not in roles + + +def test_openai_message_truncation(sentry_init, capture_events): + """Test that large messages are truncated properly in OpenAI integration.""" + sentry_init( + integrations=[OpenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = OpenAI(api_key="z") + client.chat.completions._post = mock.Mock(return_value=EXAMPLE_CHAT_COMPLETION) + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + large_messages = [ + {"role": "system", "content": "You are a helpful assistant."}, + {"role": "user", "content": large_content}, + {"role": "assistant", "content": large_content}, + {"role": "user", "content": large_content}, + ] + + with start_transaction(name="openai tx"): + client.chat.completions.create( + model="some-model", + messages=large_messages, + ) + + (event,) = events + span = event["spans"][0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in span["data"] + + messages_data = span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) <= len(large_messages) + + if "_meta" in event and len(parsed_messages) < len(large_messages): + meta_path = event["_meta"] + if ( + "spans" in meta_path + and "0" in meta_path["spans"] + and "data" in meta_path["spans"]["0"] + ): + span_meta = meta_path["spans"]["0"]["data"] + if SPANDATA.GEN_AI_REQUEST_MESSAGES in span_meta: + messages_meta = span_meta[SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert "len" in messages_meta.get("", {}) diff --git a/tests/test_ai_monitoring.py b/tests/test_ai_monitoring.py index ee757f82cd..be66860384 100644 --- a/tests/test_ai_monitoring.py +++ b/tests/test_ai_monitoring.py @@ -1,7 +1,19 @@ +import json + import pytest import sentry_sdk +from sentry_sdk._types import AnnotatedValue from sentry_sdk.ai.monitoring import ai_track +from sentry_sdk.ai.utils import ( + MAX_GEN_AI_MESSAGE_BYTES, + set_data_normalized, + truncate_and_annotate_messages, + truncate_messages_by_size, + _find_truncation_index, +) +from sentry_sdk.serializer import serialize +from sentry_sdk.utils import safe_serialize def test_ai_track(sentry_init, capture_events): @@ -160,3 +172,291 @@ async def async_tool(**kwargs): assert span["description"] == "my async tool" assert span["op"] == "custom.async.operation" + + +@pytest.fixture +def sample_messages(): + """Sample messages similar to what gen_ai integrations would use""" + return [ + {"role": "system", "content": "You are a helpful assistant."}, + { + "role": "user", + "content": "What is the difference between a list and a tuple in Python?", + }, + { + "role": "assistant", + "content": "Lists are mutable and use [], tuples are immutable and use ().", + }, + {"role": "user", "content": "Can you give me some examples?"}, + { + "role": "assistant", + "content": "Sure! Here are examples:\n\n```python\n# List\nmy_list = [1, 2, 3]\nmy_list.append(4)\n\n# Tuple\nmy_tuple = (1, 2, 3)\n# my_tuple.append(4) would error\n```", + }, + ] + + +@pytest.fixture +def large_messages(): + """Messages that will definitely exceed size limits""" + large_content = "This is a very long message. " * 100 + return [ + {"role": "system", "content": large_content}, + {"role": "user", "content": large_content}, + {"role": "assistant", "content": large_content}, + {"role": "user", "content": large_content}, + ] + + +class TestTruncateMessagesBySize: + def test_no_truncation_needed(self, sample_messages): + """Test that messages under the limit are not truncated""" + result, removed_count = truncate_messages_by_size( + sample_messages, max_bytes=MAX_GEN_AI_MESSAGE_BYTES + ) + assert len(result) == len(sample_messages) + assert result == sample_messages + assert removed_count == 0 + + def test_truncation_removes_oldest_first(self, large_messages): + """Test that oldest messages are removed first during truncation""" + small_limit = 3000 + result, removed_count = truncate_messages_by_size( + large_messages, max_bytes=small_limit + ) + assert len(result) < len(large_messages) + + if result: + assert result[-1] == large_messages[-1] + assert removed_count == len(large_messages) - len(result) + + def test_empty_messages_list(self): + """Test handling of empty messages list""" + result, removed_count = truncate_messages_by_size( + [], max_bytes=MAX_GEN_AI_MESSAGE_BYTES // 500 + ) + assert result == [] + assert removed_count == 0 + + def test_find_truncation_index( + self, + ): + """Test that the truncation index is found correctly""" + # when represented in JSON, these are each 7 bytes long + messages = ["A" * 5, "B" * 5, "C" * 5, "D" * 5, "E" * 5] + truncation_index = _find_truncation_index(messages, 20) + assert truncation_index == 3 + assert messages[truncation_index:] == ["D" * 5, "E" * 5] + + messages = ["A" * 5, "B" * 5, "C" * 5, "D" * 5, "E" * 5] + truncation_index = _find_truncation_index(messages, 40) + assert truncation_index == 0 + assert messages[truncation_index:] == [ + "A" * 5, + "B" * 5, + "C" * 5, + "D" * 5, + "E" * 5, + ] + + def test_progressive_truncation(self, large_messages): + """Test that truncation works progressively with different limits""" + limits = [ + MAX_GEN_AI_MESSAGE_BYTES // 5, + MAX_GEN_AI_MESSAGE_BYTES // 10, + MAX_GEN_AI_MESSAGE_BYTES // 25, + MAX_GEN_AI_MESSAGE_BYTES // 100, + MAX_GEN_AI_MESSAGE_BYTES // 500, + ] + prev_count = len(large_messages) + + for limit in limits: + result = truncate_messages_by_size(large_messages, max_bytes=limit) + current_count = len(result) + + assert current_count <= prev_count + assert current_count >= 1 + prev_count = current_count + + +class TestTruncateAndAnnotateMessages: + def test_no_truncation_returns_list(self, sample_messages): + class MockSpan: + def __init__(self): + self.span_id = "test_span_id" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + span = MockSpan() + scope = MockScope() + result = truncate_and_annotate_messages(sample_messages, span, scope) + + assert isinstance(result, list) + assert not isinstance(result, AnnotatedValue) + assert len(result) == len(sample_messages) + assert result == sample_messages + assert span.span_id not in scope._gen_ai_messages_truncated + + def test_truncation_sets_metadata_on_scope(self, large_messages): + class MockSpan: + def __init__(self): + self.span_id = "test_span_id" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + small_limit = 1000 + span = MockSpan() + scope = MockScope() + original_count = len(large_messages) + result = truncate_and_annotate_messages( + large_messages, span, scope, max_bytes=small_limit + ) + + assert isinstance(result, list) + assert not isinstance(result, AnnotatedValue) + assert len(result) < len(large_messages) + n_removed = original_count - len(result) + assert scope._gen_ai_messages_truncated[span.span_id] == n_removed + + def test_scope_tracks_removed_messages(self, large_messages): + class MockSpan: + def __init__(self): + self.span_id = "test_span_id" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + small_limit = 1000 + original_count = len(large_messages) + span = MockSpan() + scope = MockScope() + + result = truncate_and_annotate_messages( + large_messages, span, scope, max_bytes=small_limit + ) + + n_removed = original_count - len(result) + assert scope._gen_ai_messages_truncated[span.span_id] == n_removed + assert len(result) + n_removed == original_count + + def test_empty_messages_returns_none(self): + class MockSpan: + def __init__(self): + self.span_id = "test_span_id" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + span = MockSpan() + scope = MockScope() + result = truncate_and_annotate_messages([], span, scope) + assert result is None + + result = truncate_and_annotate_messages(None, span, scope) + assert result is None + + def test_truncated_messages_newest_first(self, large_messages): + class MockSpan: + def __init__(self): + self.span_id = "test_span_id" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + small_limit = 3000 + span = MockSpan() + scope = MockScope() + result = truncate_and_annotate_messages( + large_messages, span, scope, max_bytes=small_limit + ) + + assert isinstance(result, list) + assert result[0] == large_messages[-len(result)] + + +class TestClientAnnotation: + def test_client_wraps_truncated_messages_in_annotated_value(self, large_messages): + """Test that client.py properly wraps truncated messages in AnnotatedValue using scope data""" + from sentry_sdk._types import AnnotatedValue + from sentry_sdk.consts import SPANDATA + + class MockSpan: + def __init__(self): + self.span_id = "test_span_123" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_messages_truncated = {} + + small_limit = 3000 + span = MockSpan() + scope = MockScope() + original_count = len(large_messages) + + # Simulate what integrations do + truncated_messages = truncate_and_annotate_messages( + large_messages, span, scope, max_bytes=small_limit + ) + span.set_data(SPANDATA.GEN_AI_REQUEST_MESSAGES, truncated_messages) + + # Verify metadata was set on scope + assert span.span_id in scope._gen_ai_messages_truncated + assert scope._gen_ai_messages_truncated[span.span_id] > 0 + + # Simulate what client.py does + event = {"spans": [{"span_id": span.span_id, "data": span.data.copy()}]} + + # Mimic client.py logic - using scope to get the removed count + for event_span in event["spans"]: + span_id = event_span.get("span_id") + span_data = event_span.get("data", {}) + if ( + span_id + and span_id in scope._gen_ai_messages_truncated + and SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data + ): + messages = span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] + n_removed = scope._gen_ai_messages_truncated[span_id] + n_remaining = len(messages) if isinstance(messages, list) else 0 + original_count_calculated = n_removed + n_remaining + + span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] = AnnotatedValue( + safe_serialize(messages), + {"len": original_count_calculated}, + ) + + # Verify the annotation happened + messages_value = event["spans"][0]["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_value, AnnotatedValue) + assert messages_value.metadata["len"] == original_count + assert isinstance(messages_value.value, str) From b11c2f2c0e1b36b7c6128eabe994f8e7257a50d0 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Fri, 17 Oct 2025 11:16:35 +0200 Subject: [PATCH 552/591] fix(ai): correct size calculation, rename internal property for message truncation & add test (#4949) --- sentry_sdk/ai/utils.py | 4 +- sentry_sdk/client.py | 9 +-- sentry_sdk/scope.py | 12 ++-- tests/test_ai_monitoring.py | 113 +++++++++++++++++++++++++++--------- 4 files changed, 95 insertions(+), 43 deletions(-) diff --git a/sentry_sdk/ai/utils.py b/sentry_sdk/ai/utils.py index 1fb291bdac..06c9a23604 100644 --- a/sentry_sdk/ai/utils.py +++ b/sentry_sdk/ai/utils.py @@ -139,8 +139,6 @@ def truncate_and_annotate_messages( truncated_messages, removed_count = truncate_messages_by_size(messages, max_bytes) if removed_count > 0: - scope._gen_ai_messages_truncated[span.span_id] = len(messages) - len( - truncated_messages - ) + scope._gen_ai_original_message_count[span.span_id] = len(messages) return truncated_messages diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index ffd899b545..b4a3e8bb6b 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -598,7 +598,7 @@ def _prepare_event( if event_scrubber: event_scrubber.scrub_event(event) - if scope is not None and scope._gen_ai_messages_truncated: + if scope is not None and scope._gen_ai_original_message_count: spans = event.get("spans", []) # type: List[Dict[str, Any]] | AnnotatedValue if isinstance(spans, list): for span in spans: @@ -606,15 +606,12 @@ def _prepare_event( span_data = span.get("data", {}) if ( span_id - and span_id in scope._gen_ai_messages_truncated + and span_id in scope._gen_ai_original_message_count and SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data ): span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] = AnnotatedValue( span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES], - { - "len": scope._gen_ai_messages_truncated[span_id] - + len(span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES]) - }, + {"len": scope._gen_ai_original_message_count[span_id]}, ) if previous_total_spans is not None: event["spans"] = AnnotatedValue( diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index 5815a65440..ecb8f370c5 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -188,7 +188,7 @@ class Scope: "_extras", "_breadcrumbs", "_n_breadcrumbs_truncated", - "_gen_ai_messages_truncated", + "_gen_ai_original_message_count", "_event_processors", "_error_processors", "_should_capture", @@ -214,7 +214,7 @@ def __init__(self, ty=None, client=None): self._name = None # type: Optional[str] self._propagation_context = None # type: Optional[PropagationContext] self._n_breadcrumbs_truncated = 0 # type: int - self._gen_ai_messages_truncated = {} # type: Dict[str, int] + self._gen_ai_original_message_count = {} # type: Dict[str, int] self.client = NonRecordingClient() # type: sentry_sdk.client.BaseClient @@ -249,7 +249,7 @@ def __copy__(self): rv._breadcrumbs = copy(self._breadcrumbs) rv._n_breadcrumbs_truncated = self._n_breadcrumbs_truncated - rv._gen_ai_messages_truncated = self._gen_ai_messages_truncated.copy() + rv._gen_ai_original_message_count = self._gen_ai_original_message_count.copy() rv._event_processors = self._event_processors.copy() rv._error_processors = self._error_processors.copy() rv._propagation_context = self._propagation_context @@ -1586,8 +1586,10 @@ def update_from_scope(self, scope): self._n_breadcrumbs_truncated = ( self._n_breadcrumbs_truncated + scope._n_breadcrumbs_truncated ) - if scope._gen_ai_messages_truncated: - self._gen_ai_messages_truncated.update(scope._gen_ai_messages_truncated) + if scope._gen_ai_original_message_count: + self._gen_ai_original_message_count.update( + scope._gen_ai_original_message_count + ) if scope._span: self._span = scope._span if scope._attachments: diff --git a/tests/test_ai_monitoring.py b/tests/test_ai_monitoring.py index be66860384..5ff136f810 100644 --- a/tests/test_ai_monitoring.py +++ b/tests/test_ai_monitoring.py @@ -1,4 +1,5 @@ import json +import uuid import pytest @@ -210,32 +211,32 @@ def large_messages(): class TestTruncateMessagesBySize: def test_no_truncation_needed(self, sample_messages): """Test that messages under the limit are not truncated""" - result, removed_count = truncate_messages_by_size( + result, truncation_index = truncate_messages_by_size( sample_messages, max_bytes=MAX_GEN_AI_MESSAGE_BYTES ) assert len(result) == len(sample_messages) assert result == sample_messages - assert removed_count == 0 + assert truncation_index == 0 def test_truncation_removes_oldest_first(self, large_messages): """Test that oldest messages are removed first during truncation""" small_limit = 3000 - result, removed_count = truncate_messages_by_size( + result, truncation_index = truncate_messages_by_size( large_messages, max_bytes=small_limit ) assert len(result) < len(large_messages) if result: assert result[-1] == large_messages[-1] - assert removed_count == len(large_messages) - len(result) + assert truncation_index == len(large_messages) - len(result) def test_empty_messages_list(self): """Test handling of empty messages list""" - result, removed_count = truncate_messages_by_size( + result, truncation_index = truncate_messages_by_size( [], max_bytes=MAX_GEN_AI_MESSAGE_BYTES // 500 ) assert result == [] - assert removed_count == 0 + assert truncation_index == 0 def test_find_truncation_index( self, @@ -290,7 +291,7 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} span = MockSpan() scope = MockScope() @@ -300,7 +301,7 @@ def __init__(self): assert not isinstance(result, AnnotatedValue) assert len(result) == len(sample_messages) assert result == sample_messages - assert span.span_id not in scope._gen_ai_messages_truncated + assert span.span_id not in scope._gen_ai_original_message_count def test_truncation_sets_metadata_on_scope(self, large_messages): class MockSpan: @@ -313,9 +314,9 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} - small_limit = 1000 + small_limit = 3000 span = MockSpan() scope = MockScope() original_count = len(large_messages) @@ -326,10 +327,9 @@ def __init__(self): assert isinstance(result, list) assert not isinstance(result, AnnotatedValue) assert len(result) < len(large_messages) - n_removed = original_count - len(result) - assert scope._gen_ai_messages_truncated[span.span_id] == n_removed + assert scope._gen_ai_original_message_count[span.span_id] == original_count - def test_scope_tracks_removed_messages(self, large_messages): + def test_scope_tracks_original_message_count(self, large_messages): class MockSpan: def __init__(self): self.span_id = "test_span_id" @@ -340,9 +340,9 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} - small_limit = 1000 + small_limit = 3000 original_count = len(large_messages) span = MockSpan() scope = MockScope() @@ -351,9 +351,8 @@ def __init__(self): large_messages, span, scope, max_bytes=small_limit ) - n_removed = original_count - len(result) - assert scope._gen_ai_messages_truncated[span.span_id] == n_removed - assert len(result) + n_removed == original_count + assert scope._gen_ai_original_message_count[span.span_id] == original_count + assert len(result) == 1 def test_empty_messages_returns_none(self): class MockSpan: @@ -366,7 +365,7 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} span = MockSpan() scope = MockScope() @@ -387,7 +386,7 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} small_limit = 3000 span = MockSpan() @@ -416,7 +415,7 @@ def set_data(self, key, value): class MockScope: def __init__(self): - self._gen_ai_messages_truncated = {} + self._gen_ai_original_message_count = {} small_limit = 3000 span = MockSpan() @@ -430,29 +429,27 @@ def __init__(self): span.set_data(SPANDATA.GEN_AI_REQUEST_MESSAGES, truncated_messages) # Verify metadata was set on scope - assert span.span_id in scope._gen_ai_messages_truncated - assert scope._gen_ai_messages_truncated[span.span_id] > 0 + assert span.span_id in scope._gen_ai_original_message_count + assert scope._gen_ai_original_message_count[span.span_id] > 0 # Simulate what client.py does event = {"spans": [{"span_id": span.span_id, "data": span.data.copy()}]} - # Mimic client.py logic - using scope to get the removed count + # Mimic client.py logic - using scope to get the original length for event_span in event["spans"]: span_id = event_span.get("span_id") span_data = event_span.get("data", {}) if ( span_id - and span_id in scope._gen_ai_messages_truncated + and span_id in scope._gen_ai_original_message_count and SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data ): messages = span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] - n_removed = scope._gen_ai_messages_truncated[span_id] - n_remaining = len(messages) if isinstance(messages, list) else 0 - original_count_calculated = n_removed + n_remaining + n_original_count = scope._gen_ai_original_message_count[span_id] span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] = AnnotatedValue( safe_serialize(messages), - {"len": original_count_calculated}, + {"len": n_original_count}, ) # Verify the annotation happened @@ -460,3 +457,61 @@ def __init__(self): assert isinstance(messages_value, AnnotatedValue) assert messages_value.metadata["len"] == original_count assert isinstance(messages_value.value, str) + + def test_annotated_value_shows_correct_original_length(self, large_messages): + """Test that the annotated value correctly shows the original message count before truncation""" + from sentry_sdk.consts import SPANDATA + + class MockSpan: + def __init__(self): + self.span_id = "test_span_456" + self.data = {} + + def set_data(self, key, value): + self.data[key] = value + + class MockScope: + def __init__(self): + self._gen_ai_original_message_count = {} + + small_limit = 3000 + span = MockSpan() + scope = MockScope() + original_message_count = len(large_messages) + + truncated_messages = truncate_and_annotate_messages( + large_messages, span, scope, max_bytes=small_limit + ) + + assert len(truncated_messages) < original_message_count + + assert span.span_id in scope._gen_ai_original_message_count + stored_original_length = scope._gen_ai_original_message_count[span.span_id] + assert stored_original_length == original_message_count + + event = { + "spans": [ + { + "span_id": span.span_id, + "data": {SPANDATA.GEN_AI_REQUEST_MESSAGES: truncated_messages}, + } + ] + } + + for event_span in event["spans"]: + span_id = event_span.get("span_id") + span_data = event_span.get("data", {}) + if ( + span_id + and span_id in scope._gen_ai_original_message_count + and SPANDATA.GEN_AI_REQUEST_MESSAGES in span_data + ): + span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES] = AnnotatedValue( + span_data[SPANDATA.GEN_AI_REQUEST_MESSAGES], + {"len": scope._gen_ai_original_message_count[span_id]}, + ) + + messages_value = event["spans"][0]["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_value, AnnotatedValue) + assert messages_value.metadata["len"] == stored_original_length + assert len(messages_value.value) == len(truncated_messages) From 843c062903ae683bb2438f00c261bad4d215decd Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Fri, 17 Oct 2025 14:14:40 +0200 Subject: [PATCH 553/591] fix(ai): add message truncation in langchain (#4950) --- sentry_sdk/integrations/langchain.py | 62 ++++++++++----- .../integrations/langchain/test_langchain.py | 77 ++++++++++++++++++- 2 files changed, 116 insertions(+), 23 deletions(-) diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index 724d908665..a8ff499831 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -9,6 +9,7 @@ normalize_message_roles, set_data_normalized, get_start_span_function, + truncate_and_annotate_messages, ) from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration @@ -221,12 +222,17 @@ def on_llm_start( } for prompt in prompts ] - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): # type: (SentryLangchainCallback, Dict[str, Any], List[List[BaseMessage]], UUID, Any) -> Any @@ -278,13 +284,17 @@ def on_chat_model_start(self, serialized, messages, *, run_id, **kwargs): self._normalize_langchain_message(message) ) normalized_messages = normalize_message_roles(normalized_messages) - - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) def on_chat_model_end(self, response, *, run_id, **kwargs): # type: (SentryLangchainCallback, LLMResult, UUID, Any) -> Any @@ -758,12 +768,17 @@ def new_invoke(self, *args, **kwargs): and integration.include_prompts ): normalized_messages = normalize_message_roles([input]) - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) output = result.get("output") if ( @@ -813,12 +828,17 @@ def new_stream(self, *args, **kwargs): and integration.include_prompts ): normalized_messages = normalize_message_roles([input]) - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) # Run the agent result = f(self, *args, **kwargs) diff --git a/tests/integrations/langchain/test_langchain.py b/tests/integrations/langchain/test_langchain.py index 661208432f..1a6c4885fb 100644 --- a/tests/integrations/langchain/test_langchain.py +++ b/tests/integrations/langchain/test_langchain.py @@ -1,3 +1,4 @@ +import json from typing import List, Optional, Any, Iterator from unittest import mock from unittest.mock import Mock, patch @@ -884,8 +885,6 @@ def test_langchain_message_role_mapping(sentry_init, capture_events): # Parse the message data (might be JSON string) if isinstance(messages_data, str): - import json - try: messages = json.loads(messages_data) except json.JSONDecodeError: @@ -958,3 +957,77 @@ def test_langchain_message_role_normalization_units(): assert normalized[3]["role"] == "system" # system unchanged assert "role" not in normalized[4] # Message without role unchanged assert normalized[5] == "string message" # String message unchanged + + +def test_langchain_message_truncation(sentry_init, capture_events): + """Test that large messages are truncated properly in Langchain integration.""" + from langchain_core.outputs import LLMResult, Generation + + sentry_init( + integrations=[LangchainIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + callback = SentryLangchainCallback(max_span_map_size=100, include_prompts=True) + + run_id = "12345678-1234-1234-1234-123456789012" + serialized = {"_type": "openai-chat", "model_name": "gpt-3.5-turbo"} + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + prompts = [ + "small message 1", + large_content, + large_content, + "small message 4", + "small message 5", + ] + + with start_transaction(): + callback.on_llm_start( + serialized=serialized, + prompts=prompts, + run_id=run_id, + invocation_params={ + "temperature": 0.7, + "max_tokens": 100, + "model": "gpt-3.5-turbo", + }, + ) + + response = LLMResult( + generations=[[Generation(text="The response")]], + llm_output={ + "token_usage": { + "total_tokens": 25, + "prompt_tokens": 10, + "completion_tokens": 15, + } + }, + ) + callback.on_llm_end(response=response, run_id=run_id) + + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + llm_spans = [ + span for span in tx.get("spans", []) if span.get("op") == "gen_ai.pipeline" + ] + assert len(llm_spans) > 0 + + llm_span = llm_spans[0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in llm_span["data"] + + messages_data = llm_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) == 2 + assert "small message 4" in str(parsed_messages[0]) + assert "small message 5" in str(parsed_messages[1]) + assert tx["_meta"]["spans"]["0"]["data"]["gen_ai.request.messages"][""]["len"] == 5 From cc432a6f301c24612ee9d4d38003afe6e8589a65 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Fri, 17 Oct 2025 14:30:30 +0200 Subject: [PATCH 554/591] fix: Default breadcrumbs value for events without breadcrumbs (#4952) Change the default value when annotating truncated breadcrumbs to a dictionary with the expected keys instead of an empty list. Closes https://github.com/getsentry/sentry-python/issues/4951 --- sentry_sdk/client.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index b4a3e8bb6b..91096c6b4f 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -619,7 +619,8 @@ def _prepare_event( ) if previous_total_breadcrumbs is not None: event["breadcrumbs"] = AnnotatedValue( - event.get("breadcrumbs", []), {"len": previous_total_breadcrumbs} + event.get("breadcrumbs", {"values": []}), + {"len": previous_total_breadcrumbs}, ) # Postprocess the event here so that annotated types do From 23ec3984bdc7e048b2a8a82e8e2864065f841283 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Fri, 17 Oct 2025 15:24:20 +0200 Subject: [PATCH 555/591] fix(ai): add message truncation to langgraph (#4954) --- sentry_sdk/integrations/langgraph.py | 36 ++++++++---- .../integrations/langgraph/test_langgraph.py | 57 +++++++++++++++++++ 2 files changed, 82 insertions(+), 11 deletions(-) diff --git a/sentry_sdk/integrations/langgraph.py b/sentry_sdk/integrations/langgraph.py index 11aa1facf4..5bb0e0fd08 100644 --- a/sentry_sdk/integrations/langgraph.py +++ b/sentry_sdk/integrations/langgraph.py @@ -2,7 +2,11 @@ from typing import Any, Callable, List, Optional import sentry_sdk -from sentry_sdk.ai.utils import set_data_normalized, normalize_message_roles +from sentry_sdk.ai.utils import ( + set_data_normalized, + normalize_message_roles, + truncate_and_annotate_messages, +) from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -181,12 +185,17 @@ def new_invoke(self, *args, **kwargs): input_messages = _parse_langgraph_messages(args[0]) if input_messages: normalized_input_messages = normalize_message_roles(input_messages) - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_input_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_input_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) result = f(self, *args, **kwargs) @@ -232,12 +241,17 @@ async def new_ainvoke(self, *args, **kwargs): input_messages = _parse_langgraph_messages(args[0]) if input_messages: normalized_input_messages = normalize_message_roles(input_messages) - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - normalized_input_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_input_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) result = await f(self, *args, **kwargs) diff --git a/tests/integrations/langgraph/test_langgraph.py b/tests/integrations/langgraph/test_langgraph.py index 6ec6d9a96d..7cb86a5b03 100644 --- a/tests/integrations/langgraph/test_langgraph.py +++ b/tests/integrations/langgraph/test_langgraph.py @@ -696,3 +696,60 @@ def __init__(self, content, message_type="human"): # Verify no "ai" roles remain roles = [msg["role"] for msg in stored_messages if "role" in msg] assert "ai" not in roles + + +def test_langgraph_message_truncation(sentry_init, capture_events): + """Test that large messages are truncated properly in Langgraph integration.""" + import json + + sentry_init( + integrations=[LanggraphIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + test_state = { + "messages": [ + MockMessage("small message 1", name="user"), + MockMessage(large_content, name="assistant"), + MockMessage(large_content, name="user"), + MockMessage("small message 4", name="assistant"), + MockMessage("small message 5", name="user"), + ] + } + + pregel = MockPregelInstance("test_graph") + + def original_invoke(self, *args, **kwargs): + return {"messages": args[0].get("messages", [])} + + with start_transaction(): + wrapped_invoke = _wrap_pregel_invoke(original_invoke) + result = wrapped_invoke(pregel, test_state) + + assert result is not None + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + invoke_spans = [ + span for span in tx.get("spans", []) if span.get("op") == OP.GEN_AI_INVOKE_AGENT + ] + assert len(invoke_spans) > 0 + + invoke_span = invoke_spans[0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in invoke_span["data"] + + messages_data = invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) == 2 + assert "small message 4" in str(parsed_messages[0]) + assert "small message 5" in str(parsed_messages[1]) + assert tx["_meta"]["spans"]["0"]["data"]["gen_ai.request.messages"][""]["len"] == 5 From 2e259ae9e01a3240ab255415160f06c997c4422a Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Fri, 17 Oct 2025 15:24:29 +0200 Subject: [PATCH 556/591] fix(ai): add message trunction to anthropic (#4953) --- sentry_sdk/integrations/anthropic.py | 13 +++-- .../integrations/anthropic/test_anthropic.py | 49 +++++++++++++++++++ 2 files changed, 57 insertions(+), 5 deletions(-) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index 46c6b2a766..e61a3556e1 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -6,6 +6,7 @@ from sentry_sdk.ai.utils import ( set_data_normalized, normalize_message_roles, + truncate_and_annotate_messages, get_start_span_function, ) from sentry_sdk.consts import OP, SPANDATA, SPANSTATUS @@ -145,12 +146,14 @@ def _set_input_data(span, kwargs, integration): normalized_messages.append(message) role_normalized_messages = normalize_message_roles(normalized_messages) - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - role_normalized_messages, - unpack=False, + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + role_normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages_data, unpack=False + ) set_data_normalized( span, SPANDATA.GEN_AI_RESPONSE_STREAMING, kwargs.get("stream", False) diff --git a/tests/integrations/anthropic/test_anthropic.py b/tests/integrations/anthropic/test_anthropic.py index e9065e2d32..f7c2d7e8a7 100644 --- a/tests/integrations/anthropic/test_anthropic.py +++ b/tests/integrations/anthropic/test_anthropic.py @@ -945,3 +945,52 @@ def mock_messages_create(*args, **kwargs): # Verify no "ai" roles remain roles = [msg["role"] for msg in stored_messages] assert "ai" not in roles + + +def test_anthropic_message_truncation(sentry_init, capture_events): + """Test that large messages are truncated properly in Anthropic integration.""" + sentry_init( + integrations=[AnthropicIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + client = Anthropic(api_key="z") + client.messages._post = mock.Mock(return_value=EXAMPLE_MESSAGE) + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + messages = [ + {"role": "user", "content": "small message 1"}, + {"role": "assistant", "content": large_content}, + {"role": "user", "content": large_content}, + {"role": "assistant", "content": "small message 4"}, + {"role": "user", "content": "small message 5"}, + ] + + with start_transaction(): + client.messages.create(max_tokens=1024, messages=messages, model="model") + + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + chat_spans = [ + span for span in tx.get("spans", []) if span.get("op") == OP.GEN_AI_CHAT + ] + assert len(chat_spans) > 0 + + chat_span = chat_spans[0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in chat_span["data"] + + messages_data = chat_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) == 2 + assert "small message 4" in str(parsed_messages[0]) + assert "small message 5" in str(parsed_messages[1]) + assert tx["_meta"]["spans"]["0"]["data"]["gen_ai.request.messages"][""]["len"] == 5 From 1ad71634fa8d72b30c387f65966442c9438cd873 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Mon, 20 Oct 2025 08:50:39 +0200 Subject: [PATCH 557/591] fix(aws): Inject scopes in TimeoutThread exception with AWS lambda (#4914) Restore ServerlessTimeoutWarning isolation and current scope handling, so the scopes are active in an AWS lambda function and breadcrumbs and tags are applied on timeout. The behavior is restored to that before 2d392af3ea6da91ddbdde55d18e15c24dce6b59b. More information about how I found the bug is described in https://github.com/getsentry/sentry-python/issues/4912. Closes https://github.com/getsentry/sentry-python/issues/4894. --- sentry_sdk/integrations/aws_lambda.py | 2 ++ sentry_sdk/utils.py | 36 +++++++++++++++++-- .../TimeoutErrorScopeModified/.gitignore | 11 ++++++ .../TimeoutErrorScopeModified/index.py | 19 ++++++++++ .../aws_lambda/test_aws_lambda.py | 25 +++++++++++++ 5 files changed, 91 insertions(+), 2 deletions(-) create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/.gitignore create mode 100644 tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/index.py diff --git a/sentry_sdk/integrations/aws_lambda.py b/sentry_sdk/integrations/aws_lambda.py index 4990fd6e6a..85d1a6c28c 100644 --- a/sentry_sdk/integrations/aws_lambda.py +++ b/sentry_sdk/integrations/aws_lambda.py @@ -138,6 +138,8 @@ def sentry_handler(aws_event, aws_context, *args, **kwargs): timeout_thread = TimeoutThread( waiting_time, configured_time / MILLIS_TO_SECONDS, + isolation_scope=scope, + current_scope=sentry_sdk.get_current_scope(), ) # Starting the thread to raise timeout warning exception diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index cd825b29e2..3496178228 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -1484,17 +1484,37 @@ class TimeoutThread(threading.Thread): waiting_time and raises a custom ServerlessTimeout exception. """ - def __init__(self, waiting_time, configured_timeout): - # type: (float, int) -> None + def __init__( + self, waiting_time, configured_timeout, isolation_scope=None, current_scope=None + ): + # type: (float, int, Optional[sentry_sdk.Scope], Optional[sentry_sdk.Scope]) -> None threading.Thread.__init__(self) self.waiting_time = waiting_time self.configured_timeout = configured_timeout + + self.isolation_scope = isolation_scope + self.current_scope = current_scope + self._stop_event = threading.Event() def stop(self): # type: () -> None self._stop_event.set() + def _capture_exception(self): + # type: () -> ExcInfo + exc_info = sys.exc_info() + + client = sentry_sdk.get_client() + event, hint = event_from_exception( + exc_info, + client_options=client.options, + mechanism={"type": "threading", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + return exc_info + def run(self): # type: () -> None @@ -1510,6 +1530,18 @@ def run(self): integer_configured_timeout = integer_configured_timeout + 1 # Raising Exception after timeout duration is reached + if self.isolation_scope is not None and self.current_scope is not None: + with sentry_sdk.scope.use_isolation_scope(self.isolation_scope): + with sentry_sdk.scope.use_scope(self.current_scope): + try: + raise ServerlessTimeoutWarning( + "WARNING : Function is expected to get timed out. Configured timeout duration = {} seconds.".format( + integer_configured_timeout + ) + ) + except Exception: + reraise(*self._capture_exception()) + raise ServerlessTimeoutWarning( "WARNING : Function is expected to get timed out. Configured timeout duration = {} seconds.".format( integer_configured_timeout diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/.gitignore b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/.gitignore new file mode 100644 index 0000000000..1c56884372 --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/.gitignore @@ -0,0 +1,11 @@ +# Need to add some ignore rules in this directory, because the unit tests will add the Sentry SDK and its dependencies +# into this directory to create a Lambda function package that contains everything needed to instrument a Lambda function using Sentry. + +# Ignore everything +* + +# But not index.py +!index.py + +# And not .gitignore itself +!.gitignore diff --git a/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/index.py b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/index.py new file mode 100644 index 0000000000..109245b90d --- /dev/null +++ b/tests/integrations/aws_lambda/lambda_functions_with_embedded_sdk/TimeoutErrorScopeModified/index.py @@ -0,0 +1,19 @@ +import os +import time + +import sentry_sdk +from sentry_sdk.integrations.aws_lambda import AwsLambdaIntegration + +sentry_sdk.init( + dsn=os.environ.get("SENTRY_DSN"), + traces_sample_rate=1.0, + integrations=[AwsLambdaIntegration(timeout_warning=True)], +) + + +def handler(event, context): + sentry_sdk.set_tag("custom_tag", "custom_value") + time.sleep(15) + return { + "event": event, + } diff --git a/tests/integrations/aws_lambda/test_aws_lambda.py b/tests/integrations/aws_lambda/test_aws_lambda.py index 85da7e0b14..664220464c 100644 --- a/tests/integrations/aws_lambda/test_aws_lambda.py +++ b/tests/integrations/aws_lambda/test_aws_lambda.py @@ -223,6 +223,31 @@ def test_timeout_error(lambda_client, test_environment): assert exception["mechanism"]["type"] == "threading" +def test_timeout_error_scope_modified(lambda_client, test_environment): + lambda_client.invoke( + FunctionName="TimeoutErrorScopeModified", + Payload=json.dumps({}), + ) + envelopes = test_environment["server"].envelopes + + (error_event,) = envelopes + + assert error_event["level"] == "error" + assert ( + error_event["extra"]["lambda"]["function_name"] == "TimeoutErrorScopeModified" + ) + + (exception,) = error_event["exception"]["values"] + assert not exception["mechanism"]["handled"] + assert exception["type"] == "ServerlessTimeoutWarning" + assert exception["value"].startswith( + "WARNING : Function is expected to get timed out. Configured timeout duration =" + ) + assert exception["mechanism"]["type"] == "threading" + + assert error_event["tags"]["custom_tag"] == "custom_value" + + @pytest.mark.parametrize( "aws_event, has_request_data, batch_size", [ From b12823e1fd287889d09367f52fdc2cb572a0f48d Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Mon, 20 Oct 2025 10:33:59 +0200 Subject: [PATCH 558/591] fix(gcp): Inject scopes in TimeoutThread exception with GCP (#4959) Restore ServerlessTimeoutWarning isolation and current scope handling, so the scopes are active in a GCP function and breadcrumbs and tags are applied on timeout. The behavior is restored to that before 2d392af. Closes https://github.com/getsentry/sentry-python/issues/4958. --- sentry_sdk/integrations/gcp.py | 7 ++++++- tests/integrations/gcp/test_gcp.py | 3 +++ 2 files changed, 9 insertions(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/gcp.py b/sentry_sdk/integrations/gcp.py index c637b7414a..2b0441f95d 100644 --- a/sentry_sdk/integrations/gcp.py +++ b/sentry_sdk/integrations/gcp.py @@ -75,7 +75,12 @@ def sentry_func(functionhandler, gcp_event, *args, **kwargs): ): waiting_time = configured_time - TIMEOUT_WARNING_BUFFER - timeout_thread = TimeoutThread(waiting_time, configured_time) + timeout_thread = TimeoutThread( + waiting_time, + configured_time, + isolation_scope=scope, + current_scope=sentry_sdk.get_current_scope(), + ) # Starting the thread to raise timeout warning exception timeout_thread.start() diff --git a/tests/integrations/gcp/test_gcp.py b/tests/integrations/gcp/test_gcp.py index d088f134fe..c27c7653aa 100644 --- a/tests/integrations/gcp/test_gcp.py +++ b/tests/integrations/gcp/test_gcp.py @@ -196,6 +196,7 @@ def test_timeout_error(run_cloud_function): functionhandler = None event = {} def cloud_function(functionhandler, event): + sentry_sdk.set_tag("cloud_function", "true") time.sleep(10) return "3" """ @@ -219,6 +220,8 @@ def cloud_function(functionhandler, event): assert exception["mechanism"]["type"] == "threading" assert not exception["mechanism"]["handled"] + assert envelope_items[0]["tags"]["cloud_function"] == "true" + def test_performance_no_error(run_cloud_function): envelope_items, _ = run_cloud_function( From ae337ca4b54b30621778c0847b93605e57e08314 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Mon, 20 Oct 2025 11:57:40 +0000 Subject: [PATCH 559/591] release: 2.42.1 --- CHANGELOG.md | 19 +++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 22 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 4d1119ddde..2200c2f429 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,24 @@ # Changelog +## 2.42.1 + +### Various fixes & improvements + +- fix(gcp): Inject scopes in TimeoutThread exception with GCP (#4959) by @alexander-alderman-webb +- fix(aws): Inject scopes in TimeoutThread exception with AWS lambda (#4914) by @alexander-alderman-webb +- fix(ai): add message trunction to anthropic (#4953) by @shellmayr +- fix(ai): add message truncation to langgraph (#4954) by @shellmayr +- fix: Default breadcrumbs value for events without breadcrumbs (#4952) by @alexander-alderman-webb +- fix(ai): add message truncation in langchain (#4950) by @shellmayr +- fix(ai): correct size calculation, rename internal property for message truncation & add test (#4949) by @shellmayr +- fix(ai): introduce message truncation for openai (#4946) by @shellmayr +- fix(openai): Use non-deprecated Pydantic method to extract response text (#4942) by @JasonLovesDoggo +- ci: 🤖 Update test matrix with new releases (10/16) (#4945) by @github-actions +- Handle ValueError in scope resets (#4928) by @sl0thentr0py +- fix(litellm): Classify embeddings correctly (#4918) by @alexander-alderman-webb +- Generalize NOT_GIVEN check with omit for openai (#4926) by @sl0thentr0py +- ⚡️ Speed up function `_get_db_span_description` (#4924) by @misrasaurabh1 + ## 2.42.0 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index 2d54f45170..e92d95931e 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.42.0" +release = "2.42.1" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 2a3c9411be..c619faba83 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1348,4 +1348,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.42.0" +VERSION = "2.42.1" diff --git a/setup.py b/setup.py index 37c9cf54a6..e0894ae9e8 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.42.0", + version="2.42.1", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From f6d19694d1c032c59cf67d7121f4f38d6c66d62a Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 20 Oct 2025 12:00:49 +0000 Subject: [PATCH 560/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/20)=20(#4957)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/config.py | 3 ++ scripts/populate_tox/releases.jsonl | 32 ++++++++-------- tox.ini | 57 +++++++++++++++-------------- 3 files changed, 49 insertions(+), 43 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index 1f23b3fb08..ecf1d94c5c 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -48,6 +48,9 @@ "package": "apache-beam", "python": ">=3.7", "num_versions": 2, + "deps": { + "*": ["dill"], + }, }, "boto3": { "package": "boto3", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 537b3a64e8..b55e77eb51 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -20,13 +20,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.13.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.13.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.34.2", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.52.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.70.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.71.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} @@ -35,6 +35,7 @@ {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.40.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.68.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.69.0rc1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}} @@ -46,12 +47,12 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.53", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.55", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.6.0b1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.9", "version": "5.6.0b2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} @@ -66,7 +67,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.119.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.119.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} @@ -104,28 +105,27 @@ {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.10", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0a4", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.2", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.100.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.107.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.102.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.57.4", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.86.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.1.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.58.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.87.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.3.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.5.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.3.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} @@ -153,7 +153,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}} -{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.50.0", "yanked": false}} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.50.1", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}} @@ -199,9 +199,9 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.65.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.66.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.9", "version": "0.283.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.10", "version": "0.284.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} diff --git a/tox.ini b/tox.ini index 668e5888b4..dee0d83365 100644 --- a/tox.ini +++ b/tox.ini @@ -50,7 +50,7 @@ envlist = {py3.8,py3.11,py3.12}-anthropic-v0.16.0 {py3.8,py3.11,py3.12}-anthropic-v0.34.2 {py3.8,py3.11,py3.12}-anthropic-v0.52.2 - {py3.8,py3.12,py3.13}-anthropic-v0.70.0 + {py3.8,py3.12,py3.13}-anthropic-v0.71.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.10.0 @@ -77,30 +77,30 @@ envlist = {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 {py3.9,py3.12,py3.13}-langgraph-v0.6.10 - {py3.10,py3.12,py3.13}-langgraph-v1.0.0a4 + {py3.10,py3.12,py3.13}-langgraph-v1.0.0 {py3.9,py3.12,py3.13}-litellm-v1.77.7 - {py3.9,py3.12,py3.13}-litellm-v1.78.2 + {py3.9,py3.12,py3.13}-litellm-v1.78.5 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.9,py3.12,py3.13}-openai-base-v2.3.0 + {py3.9,py3.12,py3.13}-openai-base-v2.5.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.9,py3.12,py3.13}-openai-notiktoken-v2.3.0 + {py3.9,py3.12,py3.13}-openai-notiktoken-v2.5.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.3.3 + {py3.10,py3.12,py3.13}-openai_agents-v0.4.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.53 + {py3.9,py3.12,py3.13}-boto3-v1.40.55 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -141,7 +141,7 @@ envlist = {py3.9,py3.12,py3.13}-openfeature-v0.8.3 {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.65.0 + {py3.7,py3.12,py3.13}-statsig-v0.66.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.3.0 @@ -159,7 +159,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.9,py3.12,py3.13}-strawberry-v0.283.3 + {py3.10,py3.12,py3.13}-strawberry-v0.284.1 # ~~~ Network ~~~ @@ -184,10 +184,11 @@ envlist = {py3.7}-beam-v2.14.0 {py3.9,py3.12,py3.13}-beam-v2.68.0 + {py3.9,py3.12,py3.13}-beam-v2.69.0rc1 {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.8,py3.12,py3.13}-celery-v5.5.3 - {py3.8,py3.12,py3.13}-celery-v5.6.0b1 + {py3.9,py3.12,py3.13}-celery-v5.6.0b2 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.9,py3.12,py3.13}-dramatiq-v1.18.0 @@ -196,7 +197,7 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.3 {py3.9,py3.10}-ray-v2.7.2 - {py3.9,py3.12,py3.13}-ray-v2.50.0 + {py3.9,py3.12,py3.13}-ray-v2.50.1 {py3.6}-rq-v0.8.2 {py3.6,py3.7}-rq-v0.13.0 @@ -228,14 +229,14 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.119.0 + {py3.8,py3.12,py3.13}-fastapi-v0.119.1 # ~~~ Web 2 ~~~ {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.13.0 + {py3.9,py3.12,py3.13}-aiohttp-v3.13.1 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -344,7 +345,7 @@ deps = anthropic-v0.16.0: anthropic==0.16.0 anthropic-v0.34.2: anthropic==0.34.2 anthropic-v0.52.2: anthropic==0.52.2 - anthropic-v0.70.0: anthropic==0.70.0 + anthropic-v0.71.0: anthropic==0.71.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 anthropic-v0.34.2: httpx<0.28.0 @@ -384,28 +385,28 @@ deps = langchain-notiktoken-v0.3.27: langchain-community langgraph-v0.6.10: langgraph==0.6.10 - langgraph-v1.0.0a4: langgraph==1.0.0a4 + langgraph-v1.0.0: langgraph==1.0.0 litellm-v1.77.7: litellm==1.77.7 - litellm-v1.78.2: litellm==1.78.2 + litellm-v1.78.5: litellm==1.78.5 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.3.0: openai==2.3.0 + openai-base-v2.5.0: openai==2.5.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.3.0: openai==2.3.0 + openai-notiktoken-v2.5.0: openai==2.5.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.3.3: openai-agents==0.3.3 + openai_agents-v0.4.0: openai-agents==0.4.0 openai_agents: pytest-asyncio @@ -413,7 +414,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.53: boto3==1.40.53 + boto3-v1.40.55: boto3==1.40.55 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -463,7 +464,7 @@ deps = openfeature-v0.8.3: openfeature-sdk==0.8.3 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.65.0: statsig==0.65.0 + statsig-v0.66.0: statsig==0.66.0 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -490,7 +491,7 @@ deps = {py3.6}-graphene: aiocontextvars strawberry-v0.209.8: strawberry-graphql[fastapi,flask]==0.209.8 - strawberry-v0.283.3: strawberry-graphql[fastapi,flask]==0.283.3 + strawberry-v0.284.1: strawberry-graphql[fastapi,flask]==0.284.1 strawberry: httpx strawberry-v0.209.8: pydantic<2.11 @@ -530,10 +531,12 @@ deps = beam-v2.14.0: apache-beam==2.14.0 beam-v2.68.0: apache-beam==2.68.0 + beam-v2.69.0rc1: apache-beam==2.69.0rc1 + beam: dill celery-v4.4.7: celery==4.4.7 celery-v5.5.3: celery==5.5.3 - celery-v5.6.0b1: celery==5.6.0b1 + celery-v5.6.0b2: celery==5.6.0b2 celery: newrelic<10.17.0 celery: redis {py3.7}-celery: importlib-metadata<5.0 @@ -545,7 +548,7 @@ deps = huey-v2.5.3: huey==2.5.3 ray-v2.7.2: ray==2.7.2 - ray-v2.50.0: ray==2.50.0 + ray-v2.50.1: ray==2.50.1 rq-v0.8.2: rq==0.8.2 rq-v0.13.0: rq==0.13.0 @@ -617,7 +620,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.92.0: fastapi==0.92.0 fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.119.0: fastapi==0.119.0 + fastapi-v0.119.1: fastapi==0.119.1 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -633,10 +636,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.13.0: aiohttp==3.13.0 + aiohttp-v3.13.1: aiohttp==3.13.1 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.13.0: pytest-asyncio + aiohttp-v3.13.1: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.4: bottle==0.13.4 From a956be611c5e22a3df0b792b0c21ae3655955fe8 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 21 Oct 2025 10:25:39 +0200 Subject: [PATCH 561/591] ci: Run `common` test suite on Python 3.14 (#4896) ### Description Python 3.14 is out, test `common` on it. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4909 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --------- Co-authored-by: Alex Alderman Webb --- .../workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/populate_tox/tox.jinja | 9 +- tests/conftest.py | 11 --- tests/test_client.py | 87 +++++++++++-------- tox.ini | 9 +- 7 files changed, 63 insertions(+), 59 deletions(-) diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index eb5dde8b17..7d4a9a9566 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 2aaf2f07fe..365f0f47bc 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 98cf4456e8..699915c387 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index b86da57c24..ccf4c0e98b 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -18,7 +18,7 @@ requires = virtualenv<20.26.3 envlist = # === Common === - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent @@ -26,7 +26,7 @@ envlist = # === Integrations === # Asgi - {py3.7,py3.12,py3.13}-asgi + {py3.7,py3.12,py3.13,py3.14}-asgi # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -38,7 +38,7 @@ envlist = {py3.7}-gcp # OpenTelemetry (OTel) - {py3.7,py3.9,py3.12,py3.13}-opentelemetry + {py3.7,py3.9,py3.12,py3.13,py3.14}-opentelemetry # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel @@ -74,7 +74,7 @@ deps = # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest {py3.6,py3.7}-common: pytest<7.0.0 - {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common: pytest + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common: pytest # === Gevent === {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 @@ -177,6 +177,7 @@ basepython = py3.11: python3.11 py3.12: python3.12 py3.13: python3.13 + py3.14: python3.14 # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python diff --git a/tests/conftest.py b/tests/conftest.py index faa0251d9b..ebb4bba95f 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -57,17 +57,6 @@ with open(SENTRY_EVENT_SCHEMA) as f: SENTRY_EVENT_SCHEMA = json.load(f) -try: - import pytest_benchmark -except ImportError: - - @pytest.fixture - def benchmark(): - return lambda x: x() - -else: - del pytest_benchmark - from sentry_sdk import scope diff --git a/tests/test_client.py b/tests/test_client.py index ff3f61d702..a5b0b44931 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -751,7 +751,7 @@ def test_cyclic_data(sentry_init, capture_events): assert data == {"not_cyclic2": "", "not_cyclic": "", "is_cyclic": ""} -def test_databag_depth_stripping(sentry_init, capture_events, benchmark): +def test_databag_depth_stripping(sentry_init, capture_events): sentry_init() events = capture_events() @@ -759,58 +759,71 @@ def test_databag_depth_stripping(sentry_init, capture_events, benchmark): for _ in range(100000): value = [value] - @benchmark - def inner(): - del events[:] - try: - a = value # noqa - 1 / 0 - except Exception: - capture_exception() + del events[:] + try: + a = value # noqa + 1 / 0 + except Exception: + capture_exception() + + (event,) = events + + stacktrace_frame = event["exception"]["values"][0]["stacktrace"]["frames"][0] + a_var = stacktrace_frame["vars"]["a"] + + assert type(a_var) == list + assert len(a_var) == 1 and type(a_var[0]) == list + + first_level_list = a_var[0] + assert type(first_level_list) == list + assert len(first_level_list) == 1 + + second_level_list = first_level_list[0] + assert type(second_level_list) == list + assert len(second_level_list) == 1 - (event,) = events + third_level_list = second_level_list[0] + assert type(third_level_list) == list + assert len(third_level_list) == 1 - assert len(json.dumps(event)) < 10000 + inner_value_repr = third_level_list[0] + assert type(inner_value_repr) == str -def test_databag_string_stripping(sentry_init, capture_events, benchmark): +def test_databag_string_stripping(sentry_init, capture_events): sentry_init() events = capture_events() - @benchmark - def inner(): - del events[:] - try: - a = "A" * DEFAULT_MAX_VALUE_LENGTH * 10 # noqa - 1 / 0 - except Exception: - capture_exception() + del events[:] + try: + a = "A" * DEFAULT_MAX_VALUE_LENGTH * 10 # noqa + 1 / 0 + except Exception: + capture_exception() - (event,) = events + (event,) = events - assert len(json.dumps(event)) < DEFAULT_MAX_VALUE_LENGTH * 10 + assert len(json.dumps(event)) < DEFAULT_MAX_VALUE_LENGTH * 10 -def test_databag_breadth_stripping(sentry_init, capture_events, benchmark): +def test_databag_breadth_stripping(sentry_init, capture_events): sentry_init() events = capture_events() - @benchmark - def inner(): - del events[:] - try: - a = ["a"] * 1000000 # noqa - 1 / 0 - except Exception: - capture_exception() + del events[:] + try: + a = ["a"] * 1000000 # noqa + 1 / 0 + except Exception: + capture_exception() - (event,) = events + (event,) = events - assert ( - len(event["exception"]["values"][0]["stacktrace"]["frames"][0]["vars"]["a"]) - == MAX_DATABAG_BREADTH - ) - assert len(json.dumps(event)) < 10000 + assert ( + len(event["exception"]["values"][0]["stacktrace"]["frames"][0]["vars"]["a"]) + == MAX_DATABAG_BREADTH + ) + assert len(json.dumps(event)) < 10000 def test_chained_exceptions(sentry_init, capture_events): diff --git a/tox.ini b/tox.ini index dee0d83365..b5e4ec21e5 100644 --- a/tox.ini +++ b/tox.ini @@ -18,7 +18,7 @@ requires = virtualenv<20.26.3 envlist = # === Common === - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent @@ -26,7 +26,7 @@ envlist = # === Integrations === # Asgi - {py3.7,py3.12,py3.13}-asgi + {py3.7,py3.12,py3.13,py3.14}-asgi # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -38,7 +38,7 @@ envlist = {py3.7}-gcp # OpenTelemetry (OTel) - {py3.7,py3.9,py3.12,py3.13}-opentelemetry + {py3.7,py3.9,py3.12,py3.13,py3.14}-opentelemetry # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel @@ -304,7 +304,7 @@ deps = # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest {py3.6,py3.7}-common: pytest<7.0.0 - {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-common: pytest + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common: pytest # === Gevent === {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 @@ -825,6 +825,7 @@ basepython = py3.11: python3.11 py3.12: python3.12 py3.13: python3.13 + py3.14: python3.14 # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python From dcd94ed0ce7b68095120db0bf9e2998cdd11b0dc Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Tue, 21 Oct 2025 10:44:18 +0200 Subject: [PATCH 562/591] feat: Enable HTTP request code origin by default (#4967) Set the `enable_http_request_source` option to `True` by default. --- sentry_sdk/consts.py | 2 +- sentry_sdk/tracing_utils.py | 2 +- tests/integrations/aiohttp/test_aiohttp.py | 11 +++++------ tests/integrations/httpx/test_httpx.py | 17 +++++++++-------- tests/integrations/stdlib/test_httplib.py | 17 +++++++++-------- 5 files changed, 25 insertions(+), 24 deletions(-) diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index c619faba83..4b7d219245 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -909,7 +909,7 @@ def __init__( error_sampler=None, # type: Optional[Callable[[Event, Hint], Union[float, bool]]] enable_db_query_source=True, # type: bool db_query_source_threshold_ms=100, # type: int - enable_http_request_source=False, # type: bool + enable_http_request_source=True, # type: bool http_request_source_threshold_ms=100, # type: int spotlight=None, # type: Optional[Union[bool, str]] cert_file=None, # type: Optional[str] diff --git a/sentry_sdk/tracing_utils.py b/sentry_sdk/tracing_utils.py index 587133ad67..6506cca266 100644 --- a/sentry_sdk/tracing_utils.py +++ b/sentry_sdk/tracing_utils.py @@ -330,7 +330,7 @@ def add_http_request_source(span): if span.timestamp is None or span.start_timestamp is None: return - should_add_request_source = client.options.get("enable_http_request_source", False) + should_add_request_source = client.options.get("enable_http_request_source", True) if not should_add_request_source: return diff --git a/tests/integrations/aiohttp/test_aiohttp.py b/tests/integrations/aiohttp/test_aiohttp.py index 811bf7efca..849f9d017b 100644 --- a/tests/integrations/aiohttp/test_aiohttp.py +++ b/tests/integrations/aiohttp/test_aiohttp.py @@ -637,23 +637,19 @@ async def handler(request): @pytest.mark.asyncio -@pytest.mark.parametrize("enable_http_request_source", [None, False]) async def test_request_source_disabled( sentry_init, aiohttp_raw_server, aiohttp_client, capture_events, - enable_http_request_source, ): sentry_options = { "integrations": [AioHttpIntegration()], "traces_sample_rate": 1.0, + "enable_http_request_source": False, "http_request_source_threshold_ms": 0, } - if enable_http_request_source is not None: - sentry_options["enable_http_request_source"] = enable_http_request_source - sentry_init(**sentry_options) # server for making span request @@ -689,18 +685,21 @@ async def hello(request): @pytest.mark.asyncio +@pytest.mark.parametrize("enable_http_request_source", [None, True]) async def test_request_source_enabled( sentry_init, aiohttp_raw_server, aiohttp_client, capture_events, + enable_http_request_source, ): sentry_options = { "integrations": [AioHttpIntegration()], "traces_sample_rate": 1.0, - "enable_http_request_source": True, "http_request_source_threshold_ms": 0, } + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source sentry_init(**sentry_options) diff --git a/tests/integrations/httpx/test_httpx.py b/tests/integrations/httpx/test_httpx.py index 1f30fdf945..33bdc93c73 100644 --- a/tests/integrations/httpx/test_httpx.py +++ b/tests/integrations/httpx/test_httpx.py @@ -396,22 +396,18 @@ def test_omit_url_data_if_parsing_fails(sentry_init, capture_events, httpx_mock) assert SPANDATA.HTTP_QUERY not in event["breadcrumbs"]["values"][0]["data"] -@pytest.mark.parametrize("enable_http_request_source", [None, False]) @pytest.mark.parametrize( "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_request_source_disabled( - sentry_init, capture_events, enable_http_request_source, httpx_client, httpx_mock -): +def test_request_source_disabled(sentry_init, capture_events, httpx_client, httpx_mock): httpx_mock.add_response() sentry_options = { "integrations": [HttpxIntegration()], "traces_sample_rate": 1.0, + "enable_http_request_source": False, "http_request_source_threshold_ms": 0, } - if enable_http_request_source is not None: - sentry_options["enable_http_request_source"] = enable_http_request_source sentry_init(**sentry_options) @@ -438,18 +434,23 @@ def test_request_source_disabled( assert SPANDATA.CODE_FUNCTION not in data +@pytest.mark.parametrize("enable_http_request_source", [None, True]) @pytest.mark.parametrize( "httpx_client", (httpx.Client(), httpx.AsyncClient()), ) -def test_request_source_enabled(sentry_init, capture_events, httpx_client, httpx_mock): +def test_request_source_enabled( + sentry_init, capture_events, enable_http_request_source, httpx_client, httpx_mock +): httpx_mock.add_response() sentry_options = { "integrations": [HttpxIntegration()], "traces_sample_rate": 1.0, - "enable_http_request_source": True, "http_request_source_threshold_ms": 0, } + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source + sentry_init(**sentry_options) events = capture_events() diff --git a/tests/integrations/stdlib/test_httplib.py b/tests/integrations/stdlib/test_httplib.py index 9bd53d6ad1..acb115c6d4 100644 --- a/tests/integrations/stdlib/test_httplib.py +++ b/tests/integrations/stdlib/test_httplib.py @@ -376,16 +376,12 @@ def test_option_trace_propagation_targets( assert "baggage" not in request_headers -@pytest.mark.parametrize("enable_http_request_source", [None, False]) -def test_request_source_disabled( - sentry_init, capture_events, enable_http_request_source -): +def test_request_source_disabled(sentry_init, capture_events): sentry_options = { "traces_sample_rate": 1.0, + "enable_http_request_source": False, "http_request_source_threshold_ms": 0, } - if enable_http_request_source is not None: - sentry_options["enable_http_request_source"] = enable_http_request_source sentry_init(**sentry_options) @@ -409,12 +405,17 @@ def test_request_source_disabled( assert SPANDATA.CODE_FUNCTION not in data -def test_request_source_enabled(sentry_init, capture_events): +@pytest.mark.parametrize("enable_http_request_source", [None, True]) +def test_request_source_enabled( + sentry_init, capture_events, enable_http_request_source +): sentry_options = { "traces_sample_rate": 1.0, - "enable_http_request_source": True, "http_request_source_threshold_ms": 0, } + if enable_http_request_source is not None: + sentry_options["enable_http_request_source"] = enable_http_request_source + sentry_init(**sentry_options) events = capture_events() From cf989a89724d7548f4c69c672fbb0629183a8b14 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 22 Oct 2025 13:59:38 +0200 Subject: [PATCH 563/591] tests(huggingface): Support 1.0.0rc7 (#4979) Always use "hf-inference" provider for chat completions, which is the existing behavior of our tests prior to version 1.0.0rc7 of `huggingface_hub`. Closes https://github.com/getsentry/sentry-python/issues/4978 --- .../huggingface_hub/test_huggingface_hub.py | 21 +++++++++++++------ 1 file changed, 15 insertions(+), 6 deletions(-) diff --git a/tests/integrations/huggingface_hub/test_huggingface_hub.py b/tests/integrations/huggingface_hub/test_huggingface_hub.py index b9ab4df5bf..ffeb6acbb5 100644 --- a/tests/integrations/huggingface_hub/test_huggingface_hub.py +++ b/tests/integrations/huggingface_hub/test_huggingface_hub.py @@ -34,6 +34,15 @@ ) +def get_hf_provider_inference_client(): + # The provider parameter was added in version 0.28.0 of huggingface_hub + return ( + InferenceClient(model="test-model", provider="hf-inference") + if HF_VERSION >= (0, 28, 0) + else InferenceClient(model="test-model") + ) + + def _add_mock_response( httpx_mock, rsps, method, url, json=None, status=200, body=None, headers=None ): @@ -616,7 +625,7 @@ def test_chat_completion( ) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() with sentry_sdk.start_transaction(name="test"): client.chat_completion( @@ -688,7 +697,7 @@ def test_chat_completion_streaming( ) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() with sentry_sdk.start_transaction(name="test"): _ = list( @@ -752,7 +761,7 @@ def test_chat_completion_api_error( sentry_init(traces_sample_rate=1.0) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() with sentry_sdk.start_transaction(name="test"): with pytest.raises(HfHubHTTPError): @@ -804,7 +813,7 @@ def test_span_status_error(sentry_init, capture_events, mock_hf_api_with_errors) sentry_init(traces_sample_rate=1.0) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() with sentry_sdk.start_transaction(name="test"): with pytest.raises(HfHubHTTPError): @@ -849,7 +858,7 @@ def test_chat_completion_with_tools( ) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() tools = [ { @@ -938,7 +947,7 @@ def test_chat_completion_streaming_with_tools( ) events = capture_events() - client = InferenceClient(model="test-model") + client = get_hf_provider_inference_client() tools = [ { From bc740ad6e7e88417da24f8f34ff8598c262767f4 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Wed, 22 Oct 2025 14:29:02 +0200 Subject: [PATCH 564/591] Make logger template format safer to missing kwargs (#4981) #### Issues * resolves: #4975 * resolves: PY-1908 --- .gitignore | 1 + sentry_sdk/logger.py | 16 +++++++++++++-- tests/test_logs.py | 49 ++++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 64 insertions(+), 2 deletions(-) diff --git a/.gitignore b/.gitignore index 0dad53b2f4..3b1e93c41a 100644 --- a/.gitignore +++ b/.gitignore @@ -28,6 +28,7 @@ relay pip-wheel-metadata .mypy_cache .vscode/ +.claude/ # for running AWS Lambda tests using AWS SAM sam.template.yaml diff --git a/sentry_sdk/logger.py b/sentry_sdk/logger.py index 0ea7218e01..b90ac034bb 100644 --- a/sentry_sdk/logger.py +++ b/sentry_sdk/logger.py @@ -4,7 +4,7 @@ from typing import Any from sentry_sdk import get_client -from sentry_sdk.utils import safe_repr +from sentry_sdk.utils import safe_repr, capture_internal_exceptions OTEL_RANGES = [ # ((severity level range), severity text) @@ -18,10 +18,19 @@ ] +class _dict_default_key(dict): # type: ignore[type-arg] + """dict that returns the key if missing.""" + + def __missing__(self, key): + # type: (str) -> str + return "{" + key + "}" + + def _capture_log(severity_text, severity_number, template, **kwargs): # type: (str, int, str, **Any) -> None client = get_client() + body = template attrs = {} # type: dict[str, str | bool | float | int] if "attributes" in kwargs: attrs.update(kwargs.pop("attributes")) @@ -31,6 +40,9 @@ def _capture_log(severity_text, severity_number, template, **kwargs): # only attach template if there are parameters attrs["sentry.message.template"] = template + with capture_internal_exceptions(): + body = template.format_map(_dict_default_key(kwargs)) + attrs = { k: ( v @@ -51,7 +63,7 @@ def _capture_log(severity_text, severity_number, template, **kwargs): "severity_text": severity_text, "severity_number": severity_number, "attributes": attrs, - "body": template.format(**kwargs), + "body": body, "time_unix_nano": time.time_ns(), "trace_id": None, }, diff --git a/tests/test_logs.py b/tests/test_logs.py index 1e252c5bfb..7494aa01c8 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -397,3 +397,52 @@ def test_auto_flush_logs_after_5s(sentry_init, capture_envelopes): return raise AssertionError("1 logs was never flushed after 10 seconds") + + +@minimum_python_37 +@pytest.mark.parametrize( + "message,expected_body,params", + [ + ("any text with {braces} in it", "any text with {braces} in it", None), + ( + 'JSON data: {"key": "value", "number": 42}', + 'JSON data: {"key": "value", "number": 42}', + None, + ), + ("Multiple {braces} {in} {message}", "Multiple {braces} {in} {message}", None), + ("Nested {{braces}}", "Nested {{braces}}", None), + ("Empty braces: {}", "Empty braces: {}", None), + ("Braces with params: {user}", "Braces with params: alice", {"user": "alice"}), + ( + "Braces with partial params: {user1} {user2}", + "Braces with partial params: alice {user2}", + {"user1": "alice"}, + ), + ], +) +def test_logs_with_literal_braces( + sentry_init, capture_envelopes, message, expected_body, params +): + """ + Test that log messages with literal braces (like JSON) work without crashing. + This is a regression test for issue #4975. + """ + sentry_init(enable_logs=True) + envelopes = capture_envelopes() + + if params: + sentry_sdk.logger.info(message, **params) + else: + sentry_sdk.logger.info(message) + + get_client().flush() + logs = envelopes_to_logs(envelopes) + + assert len(logs) == 1 + assert logs[0]["body"] == expected_body + + # Verify template is only stored when there are parameters + if params: + assert logs[0]["attributes"]["sentry.message.template"] == message + else: + assert "sentry.message.template" not in logs[0]["attributes"] From 2fadc27bb47233d9db60288ce204294d271ac80e Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 22 Oct 2025 14:47:20 +0200 Subject: [PATCH 565/591] feat: Officially support 3.14 & run integration tests on 3.14 (#4974) ### Description Added 3.14 support to package meta and reran toxgen so that it detects we support 3.14 now and updates the test matrix. #### Issues - Closes https://github.com/getsentry/sentry-python/issues/4911 - Closes https://github.com/getsentry/sentry-python/issues/4910 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-flags.yml | 2 +- .../workflows/test-integrations-graphql.yml | 2 +- .../workflows/test-integrations-network.yml | 2 +- .github/workflows/test-integrations-web-1.yml | 2 +- scripts/populate_tox/config.py | 6 ++ scripts/populate_tox/releases.jsonl | 38 ++++---- setup.py | 1 + tests/integrations/quart/test_quart.py | 5 ++ tox.ini | 88 +++++++++---------- 11 files changed, 80 insertions(+), 70 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 1b9a341f17..940e0bed12 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 48a2881ffc..9b4b40f299 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index ce622d4258..e50e98fb54 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.8","3.9","3.12","3.13"] + python-version: ["3.7","3.8","3.9","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index f1a43dc77d..c9e858bc58 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index e8b5b36c98..2dc22c3ec1 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 455f15723a..42957e7066 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index ecf1d94c5c..e7dd48a9d8 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -204,6 +204,9 @@ ">=0.3": ["langchain-community"], }, "include": "<1.0", + "python": { + "<1.0": "<3.14", # https://github.com/langchain-ai/langchain/issues/33449#issuecomment-3408876631 + }, }, "langchain-notiktoken": { "package": "langchain", @@ -214,6 +217,9 @@ ">=0.3": ["langchain-community"], }, "include": "<1.0", + "python": { + "<1.0": "<3.14", # https://github.com/langchain-ai/langchain/issues/33449#issuecomment-3408876631 + }, }, "langgraph": { "package": "langgraph", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index b55e77eb51..fa827b1341 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -35,7 +35,7 @@ {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.40.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.68.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.69.0rc1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.69.0rc3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}} @@ -47,7 +47,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.55", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.56", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -71,9 +71,9 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.34.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.39.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.45.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.46.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} @@ -82,8 +82,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "grpcio", "requires_python": "", "version": "1.32.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.75.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.76.0rc1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.76.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}} @@ -100,32 +99,33 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc6", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc7", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.10", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.0", "yanked": false}} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.11", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.5", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.102.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.58.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.87.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.3.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.5.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.59.9", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.88.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.4.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} @@ -213,6 +213,6 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.8", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.9", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.19.2", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.20.0", "yanked": false}} diff --git a/setup.py b/setup.py index e0894ae9e8..cf66f30f23 100644 --- a/setup.py +++ b/setup.py @@ -107,6 +107,7 @@ def get_file_text(file_name): "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", + "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Libraries :: Python Modules", ], options={"bdist_wheel": {"universal": "1"}}, diff --git a/tests/integrations/quart/test_quart.py b/tests/integrations/quart/test_quart.py index 100642d245..7c027455c0 100644 --- a/tests/integrations/quart/test_quart.py +++ b/tests/integrations/quart/test_quart.py @@ -1,5 +1,6 @@ import importlib import json +import sys import threading from unittest import mock @@ -78,6 +79,10 @@ def integration_enabled_params(request): not importlib.util.find_spec("quart_flask_patch"), reason="requires quart_flask_patch", ) +@pytest.mark.skipif( + sys.version_info >= (3, 14), + reason="quart_flask_patch not working on 3.14 (yet?)", +) async def test_quart_flask_patch(sentry_init, capture_events, reset_integrations): # This testcase is forked because `import quart_flask_patch` needs to run # before anything else Quart-related is imported (since it monkeypatches diff --git a/tox.ini b/tox.ini index b5e4ec21e5..5dc4c77d41 100644 --- a/tox.ini +++ b/tox.ini @@ -58,15 +58,15 @@ envlist = {py3.9,py3.11,py3.12}-cohere-v5.19.0 {py3.9,py3.12,py3.13}-google_genai-v1.29.0 - {py3.9,py3.12,py3.13}-google_genai-v1.34.0 - {py3.9,py3.12,py3.13}-google_genai-v1.39.1 - {py3.9,py3.12,py3.13}-google_genai-v1.45.0 + {py3.9,py3.12,py3.13}-google_genai-v1.35.0 + {py3.9,py3.12,py3.13}-google_genai-v1.41.0 + {py3.9,py3.13,py3.14}-google_genai-v1.46.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 - {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc6 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc7 {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.11,py3.12}-langchain-base-v0.2.17 @@ -76,31 +76,31 @@ envlist = {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.2.17 {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 - {py3.9,py3.12,py3.13}-langgraph-v0.6.10 - {py3.10,py3.12,py3.13}-langgraph-v1.0.0 + {py3.9,py3.13,py3.14}-langgraph-v0.6.11 + {py3.10,py3.12,py3.13}-langgraph-v1.0.1 {py3.9,py3.12,py3.13}-litellm-v1.77.7 - {py3.9,py3.12,py3.13}-litellm-v1.78.5 + {py3.9,py3.12,py3.13}-litellm-v1.78.6 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.9,py3.12,py3.13}-openai-base-v2.5.0 + {py3.9,py3.12,py3.13}-openai-base-v2.6.0 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.9,py3.12,py3.13}-openai-notiktoken-v2.5.0 + {py3.9,py3.12,py3.13}-openai-notiktoken-v2.6.0 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.4.0 + {py3.10,py3.12,py3.13}-openai_agents-v0.4.1 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.12,py3.13}-boto3-v1.40.55 + {py3.9,py3.13,py3.14}-boto3-v1.40.56 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -135,13 +135,13 @@ envlist = # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 - {py3.9,py3.12,py3.13}-launchdarkly-v9.12.1 + {py3.9,py3.13,py3.14}-launchdarkly-v9.12.1 - {py3.8,py3.12,py3.13}-openfeature-v0.7.5 - {py3.9,py3.12,py3.13}-openfeature-v0.8.3 + {py3.8,py3.13,py3.14}-openfeature-v0.7.5 + {py3.9,py3.13,py3.14}-openfeature-v0.8.3 - {py3.7,py3.12,py3.13}-statsig-v0.55.3 - {py3.7,py3.12,py3.13}-statsig-v0.66.0 + {py3.7,py3.13,py3.14}-statsig-v0.55.3 + {py3.7,py3.13,py3.14}-statsig-v0.66.0 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.3.0 @@ -159,15 +159,14 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.10,py3.12,py3.13}-strawberry-v0.284.1 + {py3.10,py3.13,py3.14}-strawberry-v0.284.1 # ~~~ Network ~~~ {py3.7,py3.8}-grpc-v1.32.0 {py3.7,py3.9,py3.10}-grpc-v1.47.5 {py3.7,py3.11,py3.12}-grpc-v1.62.3 - {py3.9,py3.12,py3.13}-grpc-v1.75.1 - {py3.9,py3.12,py3.13}-grpc-v1.76.0rc1 + {py3.9,py3.13,py3.14}-grpc-v1.76.0 {py3.6,py3.8,py3.9}-httpx-v0.16.1 {py3.6,py3.9,py3.10}-httpx-v0.20.0 @@ -175,7 +174,7 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v0.28.1 {py3.6}-requests-v2.12.5 - {py3.9,py3.12,py3.13}-requests-v2.32.5 + {py3.9,py3.13,py3.14}-requests-v2.32.5 # ~~~ Tasks ~~~ @@ -184,7 +183,7 @@ envlist = {py3.7}-beam-v2.14.0 {py3.9,py3.12,py3.13}-beam-v2.68.0 - {py3.9,py3.12,py3.13}-beam-v2.69.0rc1 + {py3.9,py3.12,py3.13}-beam-v2.69.0rc3 {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.8,py3.12,py3.13}-celery-v5.5.3 @@ -218,25 +217,25 @@ envlist = {py3.12,py3.13}-django-v6.0a1 {py3.6,py3.7,py3.8}-flask-v1.1.4 - {py3.8,py3.12,py3.13}-flask-v2.3.3 - {py3.9,py3.12,py3.13}-flask-v3.1.2 + {py3.8,py3.13,py3.14}-flask-v2.3.3 + {py3.9,py3.13,py3.14}-flask-v3.1.2 {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.27.0 {py3.8,py3.12,py3.13}-starlette-v0.38.6 - {py3.9,py3.12,py3.13}-starlette-v0.48.0 + {py3.9,py3.13,py3.14}-starlette-v0.48.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.92.0 {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.12,py3.13}-fastapi-v0.119.1 + {py3.8,py3.13,py3.14}-fastapi-v0.119.1 # ~~~ Web 2 ~~~ {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.12,py3.13}-aiohttp-v3.13.1 + {py3.9,py3.13,py3.14}-aiohttp-v3.13.1 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -256,7 +255,7 @@ envlist = {py3.6,py3.10,py3.11}-pyramid-v2.0.2 {py3.7,py3.9,py3.10}-quart-v0.16.3 - {py3.9,py3.12,py3.13}-quart-v0.20.0 + {py3.9,py3.13,py3.14}-quart-v0.20.0 {py3.6}-sanic-v0.8.3 {py3.6,py3.8,py3.9}-sanic-v20.12.7 @@ -280,10 +279,10 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.9,py3.12,py3.13}-trytond-v7.6.8 + {py3.9,py3.12,py3.13}-trytond-v7.6.9 {py3.7,py3.12,py3.13}-typer-v0.15.4 - {py3.8,py3.12,py3.13}-typer-v0.19.2 + {py3.8,py3.13,py3.14}-typer-v0.20.0 @@ -356,16 +355,16 @@ deps = cohere-v5.19.0: cohere==5.19.0 google_genai-v1.29.0: google-genai==1.29.0 - google_genai-v1.34.0: google-genai==1.34.0 - google_genai-v1.39.1: google-genai==1.39.1 - google_genai-v1.45.0: google-genai==1.45.0 + google_genai-v1.35.0: google-genai==1.35.0 + google_genai-v1.41.0: google-genai==1.41.0 + google_genai-v1.46.0: google-genai==1.46.0 google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 huggingface_hub-v0.35.3: huggingface_hub==0.35.3 - huggingface_hub-v1.0.0rc6: huggingface_hub==1.0.0rc6 + huggingface_hub-v1.0.0rc7: huggingface_hub==1.0.0rc7 huggingface_hub: responses huggingface_hub: pytest-httpx @@ -384,29 +383,29 @@ deps = langchain-notiktoken: langchain-openai langchain-notiktoken-v0.3.27: langchain-community - langgraph-v0.6.10: langgraph==0.6.10 - langgraph-v1.0.0: langgraph==1.0.0 + langgraph-v0.6.11: langgraph==0.6.11 + langgraph-v1.0.1: langgraph==1.0.1 litellm-v1.77.7: litellm==1.77.7 - litellm-v1.78.5: litellm==1.78.5 + litellm-v1.78.6: litellm==1.78.6 openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.5.0: openai==2.5.0 + openai-base-v2.6.0: openai==2.6.0 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.5.0: openai==2.5.0 + openai-notiktoken-v2.6.0: openai==2.6.0 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.4.0: openai-agents==0.4.0 + openai_agents-v0.4.1: openai-agents==0.4.1 openai_agents: pytest-asyncio @@ -414,7 +413,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.55: boto3==1.40.55 + boto3-v1.40.56: boto3==1.40.56 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -500,8 +499,7 @@ deps = grpc-v1.32.0: grpcio==1.32.0 grpc-v1.47.5: grpcio==1.47.5 grpc-v1.62.3: grpcio==1.62.3 - grpc-v1.75.1: grpcio==1.75.1 - grpc-v1.76.0rc1: grpcio==1.76.0rc1 + grpc-v1.76.0: grpcio==1.76.0 grpc: protobuf grpc: mypy-protobuf grpc: types-protobuf @@ -531,7 +529,7 @@ deps = beam-v2.14.0: apache-beam==2.14.0 beam-v2.68.0: apache-beam==2.68.0 - beam-v2.69.0rc1: apache-beam==2.69.0rc1 + beam-v2.69.0rc3: apache-beam==2.69.0rc3 beam: dill celery-v4.4.7: celery==4.4.7 @@ -715,13 +713,13 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.6.8: trytond==7.6.8 + trytond-v7.6.9: trytond==7.6.9 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 typer-v0.15.4: typer==0.15.4 - typer-v0.19.2: typer==0.19.2 + typer-v0.20.0: typer==0.20.0 From 086197fa1e539d734fe3362a70207365064b063a Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 22 Oct 2025 16:07:39 +0200 Subject: [PATCH 566/591] ci: Run `common` test suite on Python 3.14t (#4969) Update `split_tox_gh_actions.py` to support the free-threading `t` suffix in versions. Upgrade `pip` in tox for free-threaded Python by setting `VIRTUALENV_PIP`. Check sys.flags.thread_inherit_context to determine if context is already propagated when creating threads in the `threading` tests. See https://docs.python.org/3/howto/free-threading-python.html#context-variables Closes https://github.com/getsentry/sentry-python/issues/4970 --- .github/workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/populate_tox/tox.jinja | 12 ++++++++---- .../split_tox_gh_actions/split_tox_gh_actions.py | 12 ++++++++++-- tests/integrations/threading/test_threading.py | 14 ++++++++++---- tox.ini | 12 ++++++++---- 7 files changed, 39 insertions(+), 17 deletions(-) diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 7d4a9a9566..0d1f2a90f8 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 365f0f47bc..56a31bff27 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 699915c387..2cd0980962 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index ccf4c0e98b..4de4d94b5f 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -18,7 +18,7 @@ requires = virtualenv<20.26.3 envlist = # === Common === - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14,py3.14t}-common # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent @@ -26,7 +26,7 @@ envlist = # === Integrations === # Asgi - {py3.7,py3.12,py3.13,py3.14}-asgi + {py3.7,py3.12,py3.13,py3.14,py3.14t}-asgi # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -38,7 +38,7 @@ envlist = {py3.7}-gcp # OpenTelemetry (OTel) - {py3.7,py3.9,py3.12,py3.13,py3.14}-opentelemetry + {py3.7,py3.9,py3.12,py3.13,py3.14,py3.14t}-opentelemetry # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel @@ -74,7 +74,7 @@ deps = # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest {py3.6,py3.7}-common: pytest<7.0.0 - {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common: pytest + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14,py3.14t}-common: pytest # === Gevent === {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 @@ -134,6 +134,9 @@ setenv = OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES COVERAGE_FILE=.coverage-sentry-{envname} py3.6: COVERAGE_RCFILE=.coveragerc36 + # Lowest version to support free-threading + # https://discuss.python.org/t/announcement-pip-24-1-release/56281 + py3.14t: VIRTUALENV_PIP=24.1 django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 @@ -178,6 +181,7 @@ basepython = py3.12: python3.12 py3.13: python3.13 py3.14: python3.14 + py3.14t: python3.14t # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 9dea95842b..12222eff1b 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -28,7 +28,7 @@ TOXENV_REGEX = re.compile( r""" - {?(?P(py\d+\.\d+,?)+)}? + {?(?P(py\d+\.\d+t?,?)+)}? -(?P[a-z](?:[a-z_]|-(?!v{?\d))*[a-z0-9]) (?:-( (v{?(?P[0-9.]+[0-9a-z,.]*}?)) @@ -250,11 +250,19 @@ def find_frameworks_missing_from_groups(py_versions): return all_frameworks - frameworks_in_a_group +def _version_key(v): + major_version, minor_version_and_suffix = v.split(".") + if minor_version_and_suffix.endswith("t"): + return int(major_version), int(minor_version_and_suffix.rstrip("t")), 1 + + return int(major_version), int(minor_version_and_suffix), 0 + + def _normalize_py_versions(py_versions): def replace_and_sort(versions): return sorted( [py.replace("py", "") for py in versions], - key=lambda v: tuple(map(int, v.split("."))), + key=_version_key, ) if isinstance(py_versions, dict): diff --git a/tests/integrations/threading/test_threading.py b/tests/integrations/threading/test_threading.py index 9c9a24aa63..788fb24fdf 100644 --- a/tests/integrations/threading/test_threading.py +++ b/tests/integrations/threading/test_threading.py @@ -1,4 +1,5 @@ import gc +import sys from concurrent import futures from textwrap import dedent from threading import Thread @@ -68,7 +69,8 @@ def stage2(): assert exception["mechanism"]["type"] == "threading" assert not exception["mechanism"]["handled"] - if propagate_hub: + # Free-threaded builds set thread_inherit_context to True, otherwise thread_inherit_context is False + if propagate_hub or getattr(sys.flags, "thread_inherit_context", None): assert event["tags"]["stage1"] == "true" else: assert "stage1" not in event.get("tags", {}) @@ -94,7 +96,8 @@ def double(number): sentry_sdk.flush() - if propagate_hub: + # Free-threaded builds set thread_inherit_context to True, otherwise thread_inherit_context is False + if propagate_hub or getattr(sys.flags, "thread_inherit_context", None): assert len(events) == 1 (event,) = events assert event["spans"][0]["trace_id"] == event["spans"][1]["trace_id"] @@ -248,7 +251,9 @@ def do_some_work(number): t.join() (event,) = events - if propagate_scope: + + # Free-threaded builds set thread_inherit_context to True, otherwise thread_inherit_context is False + if propagate_scope or getattr(sys.flags, "thread_inherit_context", None): assert render_span_tree(event) == dedent( """\ - op="outer-trx": description=null @@ -309,7 +314,8 @@ def do_some_work(number): (event,) = events - if propagate_scope: + # Free-threaded builds set thread_inherit_context to True, otherwise thread_inherit_context is False + if propagate_scope or getattr(sys.flags, "thread_inherit_context", None): assert render_span_tree(event) == dedent( """\ - op="outer-trx": description=null diff --git a/tox.ini b/tox.ini index 5dc4c77d41..ef7b0e85b9 100644 --- a/tox.ini +++ b/tox.ini @@ -18,7 +18,7 @@ requires = virtualenv<20.26.3 envlist = # === Common === - {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common + {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14,py3.14t}-common # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent @@ -26,7 +26,7 @@ envlist = # === Integrations === # Asgi - {py3.7,py3.12,py3.13,py3.14}-asgi + {py3.7,py3.12,py3.13,py3.14,py3.14t}-asgi # AWS Lambda {py3.8,py3.9,py3.11,py3.13}-aws_lambda @@ -38,7 +38,7 @@ envlist = {py3.7}-gcp # OpenTelemetry (OTel) - {py3.7,py3.9,py3.12,py3.13,py3.14}-opentelemetry + {py3.7,py3.9,py3.12,py3.13,py3.14,py3.14t}-opentelemetry # OpenTelemetry Experimental (POTel) {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13}-potel @@ -303,7 +303,7 @@ deps = # and https://github.com/pytest-dev/pytest-forked/issues/67 # for justification of the upper bound on pytest {py3.6,py3.7}-common: pytest<7.0.0 - {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-common: pytest + {py3.8,py3.9,py3.10,py3.11,py3.12,py3.13,py3.14,py3.14t}-common: pytest # === Gevent === {py3.6,py3.7,py3.8,py3.9,py3.10,py3.11}-gevent: gevent>=22.10.0, <22.11.0 @@ -728,6 +728,9 @@ setenv = OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES COVERAGE_FILE=.coverage-sentry-{envname} py3.6: COVERAGE_RCFILE=.coveragerc36 + # Lowest version to support free-threading + # https://discuss.python.org/t/announcement-pip-24-1-release/56281 + py3.14t: VIRTUALENV_PIP=24.1 django: DJANGO_SETTINGS_MODULE=tests.integrations.django.myapp.settings spark-v{3.0.3,3.5.6}: JAVA_HOME=/usr/lib/jvm/temurin-11-jdk-amd64 @@ -824,6 +827,7 @@ basepython = py3.12: python3.12 py3.13: python3.13 py3.14: python3.14 + py3.14t: python3.14t # Python version is pinned here for consistency across environments. # Tools like ruff and mypy have options that pin the target Python From 3d026cd8b1adb77f7b3086f64bb02a8bfa31e02c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Thu, 23 Oct 2025 10:58:32 +0200 Subject: [PATCH 567/591] feat(langchain): Support v1 (#4874) ### Description Adapt to import changes in new langchain 1.x #### Issues Closes https://github.com/getsentry/sentry-python/issues/4735 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- scripts/populate_tox/config.py | 4 +-- scripts/populate_tox/releases.jsonl | 10 +++--- sentry_sdk/integrations/langchain.py | 10 ++++-- .../integrations/langchain/test_langchain.py | 10 +++++- tox.ini | 32 +++++++++++-------- 5 files changed, 42 insertions(+), 24 deletions(-) diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index e7dd48a9d8..b7f6d9efe7 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -202,8 +202,8 @@ "*": ["openai", "tiktoken", "langchain-openai"], "<=0.1": ["httpx<0.28.0"], ">=0.3": ["langchain-community"], + ">=1.0": ["langchain-classic"], }, - "include": "<1.0", "python": { "<1.0": "<3.14", # https://github.com/langchain-ai/langchain/issues/33449#issuecomment-3408876631 }, @@ -215,8 +215,8 @@ "*": ["openai", "langchain-openai"], "<=0.1": ["httpx<0.28.0"], ">=0.3": ["langchain-community"], + ">=1.0": ["langchain-classic"], }, - "include": "<1.0", "python": { "<1.0": "<3.14", # https://github.com/langchain-ai/langchain/issues/33449#issuecomment-3408876631 }, diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index fa827b1341..edef42967d 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -7,7 +7,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.2.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.7", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0a1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0b1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Flask", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "1.1.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Flask", "requires_python": ">=3.9", "version": "3.1.2", "yanked": false}} @@ -47,7 +47,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.56", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.57", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -101,14 +101,14 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc7", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.2.17", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} +{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0.0,>=3.10.0", "version": "1.0.2", "yanked": false}} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.11", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.6", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.7", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} @@ -165,7 +165,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0b3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}} {"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}} diff --git a/sentry_sdk/integrations/langchain.py b/sentry_sdk/integrations/langchain.py index a8ff499831..1f5b41bd43 100644 --- a/sentry_sdk/integrations/langchain.py +++ b/sentry_sdk/integrations/langchain.py @@ -50,9 +50,15 @@ try: - from langchain.agents import AgentExecutor + # >=v1 + from langchain_classic.agents import AgentExecutor # type: ignore[import-not-found] except ImportError: - AgentExecutor = None + try: + # =3.0,<4.0 django-v1.11.29: Werkzeug<2.1.0 django-v2.2.28: djangorestframework>=3.0,<4.0 From e9738f62e1d458af20d554d26cd20bb03b3e47f6 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 23 Oct 2025 13:28:19 +0200 Subject: [PATCH 568/591] fix(ai): add message truncation to litellm (#4973) --- sentry_sdk/integrations/litellm.py | 15 ++++-- tests/integrations/litellm/test_litellm.py | 59 ++++++++++++++++++++++ 2 files changed, 70 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/litellm.py b/sentry_sdk/integrations/litellm.py index 1f047b1c1d..43661e2432 100644 --- a/sentry_sdk/integrations/litellm.py +++ b/sentry_sdk/integrations/litellm.py @@ -3,7 +3,11 @@ import sentry_sdk from sentry_sdk import consts from sentry_sdk.ai.monitoring import record_token_usage -from sentry_sdk.ai.utils import get_start_span_function, set_data_normalized +from sentry_sdk.ai.utils import ( + get_start_span_function, + set_data_normalized, + truncate_and_annotate_messages, +) from sentry_sdk.consts import SPANDATA from sentry_sdk.integrations import DidNotEnable, Integration from sentry_sdk.scope import should_send_default_pii @@ -76,9 +80,12 @@ def _input_callback(kwargs): # Record messages if allowed messages = kwargs.get("messages", []) if messages and should_send_default_pii() and integration.include_prompts: - set_data_normalized( - span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False - ) + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages(messages, span, scope) + if messages_data is not None: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages_data, unpack=False + ) # Record other parameters params = { diff --git a/tests/integrations/litellm/test_litellm.py b/tests/integrations/litellm/test_litellm.py index 19ae206c85..8e1ad21254 100644 --- a/tests/integrations/litellm/test_litellm.py +++ b/tests/integrations/litellm/test_litellm.py @@ -1,3 +1,4 @@ +import json import pytest from unittest import mock from datetime import datetime @@ -546,3 +547,61 @@ def dict(self): # Should have extracted the response message assert SPANDATA.GEN_AI_RESPONSE_TEXT in span["data"] + + +def test_litellm_message_truncation(sentry_init, capture_events): + """Test that large messages are truncated properly in LiteLLM integration.""" + sentry_init( + integrations=[LiteLLMIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + messages = [ + {"role": "user", "content": "small message 1"}, + {"role": "assistant", "content": large_content}, + {"role": "user", "content": large_content}, + {"role": "assistant", "content": "small message 4"}, + {"role": "user", "content": "small message 5"}, + ] + mock_response = MockCompletionResponse() + + with start_transaction(name="litellm test"): + kwargs = { + "model": "gpt-3.5-turbo", + "messages": messages, + } + + _input_callback(kwargs) + _success_callback( + kwargs, + mock_response, + datetime.now(), + datetime.now(), + ) + + assert len(events) > 0 + tx = events[0] + assert tx["type"] == "transaction" + + chat_spans = [ + span for span in tx.get("spans", []) if span.get("op") == OP.GEN_AI_CHAT + ] + assert len(chat_spans) > 0 + + chat_span = chat_spans[0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in chat_span["data"] + + messages_data = chat_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) == 2 + assert "small message 4" in str(parsed_messages[0]) + assert "small message 5" in str(parsed_messages[1]) + assert tx["_meta"]["spans"]["0"]["data"]["gen_ai.request.messages"][""]["len"] == 5 From 1b748278162d338b87a238c3f30da9e79dc57356 Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Thu, 23 Oct 2025 15:22:32 +0200 Subject: [PATCH 569/591] feat(integrations): pydantic-ai integration (#4906) ### Description Integration for Pydantic AI. Support for sync and async (streaming and non-streaming) #### Issues Closes https://linear.app/getsentry/issue/TET-1242/python-sdk-pydantic-ai-integration #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 + scripts/populate_tox/config.py | 6 + scripts/populate_tox/releases.jsonl | 1 + .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/__init__.py | 1 + .../integrations/pydantic_ai/__init__.py | 47 + sentry_sdk/integrations/pydantic_ai/consts.py | 1 + .../pydantic_ai/patches/__init__.py | 4 + .../pydantic_ai/patches/agent_run.py | 217 ++ .../pydantic_ai/patches/graph_nodes.py | 105 + .../pydantic_ai/patches/model_request.py | 35 + .../integrations/pydantic_ai/patches/tools.py | 75 + .../pydantic_ai/spans/__init__.py | 3 + .../pydantic_ai/spans/ai_client.py | 253 ++ .../pydantic_ai/spans/execute_tool.py | 49 + .../pydantic_ai/spans/invoke_agent.py | 112 + sentry_sdk/integrations/pydantic_ai/utils.py | 175 ++ setup.py | 1 + tests/integrations/pydantic_ai/__init__.py | 3 + .../pydantic_ai/test_pydantic_ai.py | 2425 +++++++++++++++++ tox.ini | 6 + 21 files changed, 3524 insertions(+) create mode 100644 sentry_sdk/integrations/pydantic_ai/__init__.py create mode 100644 sentry_sdk/integrations/pydantic_ai/consts.py create mode 100644 sentry_sdk/integrations/pydantic_ai/patches/__init__.py create mode 100644 sentry_sdk/integrations/pydantic_ai/patches/agent_run.py create mode 100644 sentry_sdk/integrations/pydantic_ai/patches/graph_nodes.py create mode 100644 sentry_sdk/integrations/pydantic_ai/patches/model_request.py create mode 100644 sentry_sdk/integrations/pydantic_ai/patches/tools.py create mode 100644 sentry_sdk/integrations/pydantic_ai/spans/__init__.py create mode 100644 sentry_sdk/integrations/pydantic_ai/spans/ai_client.py create mode 100644 sentry_sdk/integrations/pydantic_ai/spans/execute_tool.py create mode 100644 sentry_sdk/integrations/pydantic_ai/spans/invoke_agent.py create mode 100644 sentry_sdk/integrations/pydantic_ai/utils.py create mode 100644 tests/integrations/pydantic_ai/__init__.py create mode 100644 tests/integrations/pydantic_ai/test_pydantic_ai.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index 940e0bed12..ca1a10aa5c 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -94,6 +94,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-openai_agents" + - name: Test pydantic_ai + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-pydantic_ai" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index b7f6d9efe7..e48a1cba88 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -281,6 +281,12 @@ "package": "pure_eval", "num_versions": 2, }, + "pydantic_ai": { + "package": "pydantic-ai", + "deps": { + "*": ["pytest-asyncio"], + }, + }, "pymongo": { "package": "pymongo", "deps": { diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index edef42967d..03fd954b65 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -130,6 +130,7 @@ {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.17", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 12222eff1b..6f7015f07c 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -81,6 +81,7 @@ "openai-base", "openai-notiktoken", "openai_agents", + "pydantic_ai", ], "Cloud": [ "aws_lambda", diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 9e279b8345..45a0f53eaf 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -152,6 +152,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "openai": (1, 0, 0), "openai_agents": (0, 0, 19), "openfeature": (0, 7, 1), + "pydantic_ai": (1, 0, 0), "quart": (0, 16, 0), "ray": (2, 7, 0), "requests": (2, 0, 0), diff --git a/sentry_sdk/integrations/pydantic_ai/__init__.py b/sentry_sdk/integrations/pydantic_ai/__init__.py new file mode 100644 index 0000000000..9fccafd6d2 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/__init__.py @@ -0,0 +1,47 @@ +from sentry_sdk.integrations import DidNotEnable, Integration + + +try: + import pydantic_ai # type: ignore +except ImportError: + raise DidNotEnable("pydantic-ai not installed") + + +from .patches import ( + _patch_agent_run, + _patch_graph_nodes, + _patch_model_request, + _patch_tool_execution, +) + + +class PydanticAIIntegration(Integration): + identifier = "pydantic_ai" + origin = f"auto.ai.{identifier}" + + def __init__(self, include_prompts=True): + # type: (bool) -> None + """ + Initialize the Pydantic AI integration. + + Args: + include_prompts: Whether to include prompts and messages in span data. + Requires send_default_pii=True. Defaults to True. + """ + self.include_prompts = include_prompts + + @staticmethod + def setup_once(): + # type: () -> None + """ + Set up the pydantic-ai integration. + + This patches the key methods in pydantic-ai to create Sentry spans for: + - Agent invocations (Agent.run methods) + - Model requests (AI client calls) + - Tool executions + """ + _patch_agent_run() + _patch_graph_nodes() + _patch_model_request() + _patch_tool_execution() diff --git a/sentry_sdk/integrations/pydantic_ai/consts.py b/sentry_sdk/integrations/pydantic_ai/consts.py new file mode 100644 index 0000000000..afa66dc47d --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/consts.py @@ -0,0 +1 @@ +SPAN_ORIGIN = "auto.ai.pydantic_ai" diff --git a/sentry_sdk/integrations/pydantic_ai/patches/__init__.py b/sentry_sdk/integrations/pydantic_ai/patches/__init__.py new file mode 100644 index 0000000000..de28780728 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/patches/__init__.py @@ -0,0 +1,4 @@ +from .agent_run import _patch_agent_run # noqa: F401 +from .graph_nodes import _patch_graph_nodes # noqa: F401 +from .model_request import _patch_model_request # noqa: F401 +from .tools import _patch_tool_execution # noqa: F401 diff --git a/sentry_sdk/integrations/pydantic_ai/patches/agent_run.py b/sentry_sdk/integrations/pydantic_ai/patches/agent_run.py new file mode 100644 index 0000000000..7c403c7ba3 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/patches/agent_run.py @@ -0,0 +1,217 @@ +from functools import wraps + +import sentry_sdk +from sentry_sdk.tracing_utils import set_span_errored +from sentry_sdk.utils import event_from_exception + +from ..spans import invoke_agent_span, update_invoke_agent_span + +from typing import TYPE_CHECKING +from pydantic_ai.agent import Agent # type: ignore + +if TYPE_CHECKING: + from typing import Any, Callable, Optional + + +def _capture_exception(exc): + # type: (Any) -> None + set_span_errored() + + event, hint = event_from_exception( + exc, + client_options=sentry_sdk.get_client().options, + mechanism={"type": "pydantic_ai", "handled": False}, + ) + sentry_sdk.capture_event(event, hint=hint) + + +class _StreamingContextManagerWrapper: + """Wrapper for streaming methods that return async context managers.""" + + def __init__( + self, + agent, + original_ctx_manager, + user_prompt, + model, + model_settings, + is_streaming=True, + ): + # type: (Any, Any, Any, Any, Any, bool) -> None + self.agent = agent + self.original_ctx_manager = original_ctx_manager + self.user_prompt = user_prompt + self.model = model + self.model_settings = model_settings + self.is_streaming = is_streaming + self._isolation_scope = None # type: Any + self._span = None # type: Optional[sentry_sdk.tracing.Span] + self._result = None # type: Any + + async def __aenter__(self): + # type: () -> Any + # Set up isolation scope and invoke_agent span + self._isolation_scope = sentry_sdk.isolation_scope() + self._isolation_scope.__enter__() + + # Store agent reference and streaming flag + sentry_sdk.get_current_scope().set_context( + "pydantic_ai_agent", {"_agent": self.agent, "_streaming": self.is_streaming} + ) + + # Create invoke_agent span (will be closed in __aexit__) + self._span = invoke_agent_span( + self.user_prompt, self.agent, self.model, self.model_settings + ) + self._span.__enter__() + + # Enter the original context manager + result = await self.original_ctx_manager.__aenter__() + self._result = result + return result + + async def __aexit__(self, exc_type, exc_val, exc_tb): + # type: (Any, Any, Any) -> None + try: + # Exit the original context manager first + await self.original_ctx_manager.__aexit__(exc_type, exc_val, exc_tb) + + # Update span with output if successful + if exc_type is None and self._result and hasattr(self._result, "output"): + output = ( + self._result.output if hasattr(self._result, "output") else None + ) + if self._span is not None: + update_invoke_agent_span(self._span, output) + finally: + sentry_sdk.get_current_scope().remove_context("pydantic_ai_agent") + # Clean up invoke span + if self._span: + self._span.__exit__(exc_type, exc_val, exc_tb) + + # Clean up isolation scope + if self._isolation_scope: + self._isolation_scope.__exit__(exc_type, exc_val, exc_tb) + + +def _create_run_wrapper(original_func, is_streaming=False): + # type: (Callable[..., Any], bool) -> Callable[..., Any] + """ + Wraps the Agent.run method to create an invoke_agent span. + + Args: + original_func: The original run method + is_streaming: Whether this is a streaming method (for future use) + """ + + @wraps(original_func) + async def wrapper(self, *args, **kwargs): + # type: (Any, *Any, **Any) -> Any + # Isolate each workflow so that when agents are run in asyncio tasks they + # don't touch each other's scopes + with sentry_sdk.isolation_scope(): + # Store agent reference and streaming flag in Sentry scope for access in nested spans + # We store the full agent to allow access to tools and system prompts + sentry_sdk.get_current_scope().set_context( + "pydantic_ai_agent", {"_agent": self, "_streaming": is_streaming} + ) + + # Extract parameters for the span + user_prompt = kwargs.get("user_prompt") or (args[0] if args else None) + model = kwargs.get("model") + model_settings = kwargs.get("model_settings") + + # Create invoke_agent span + with invoke_agent_span(user_prompt, self, model, model_settings) as span: + try: + result = await original_func(self, *args, **kwargs) + + # Update span with output + output = result.output if hasattr(result, "output") else None + update_invoke_agent_span(span, output) + + return result + except Exception as exc: + _capture_exception(exc) + raise exc from None + finally: + sentry_sdk.get_current_scope().remove_context("pydantic_ai_agent") + + return wrapper + + +def _create_streaming_wrapper(original_func): + # type: (Callable[..., Any]) -> Callable[..., Any] + """ + Wraps run_stream method that returns an async context manager. + """ + + @wraps(original_func) + def wrapper(self, *args, **kwargs): + # type: (Any, *Any, **Any) -> Any + # Extract parameters for the span + user_prompt = kwargs.get("user_prompt") or (args[0] if args else None) + model = kwargs.get("model") + model_settings = kwargs.get("model_settings") + + # Call original function to get the context manager + original_ctx_manager = original_func(self, *args, **kwargs) + + # Wrap it with our instrumentation + return _StreamingContextManagerWrapper( + agent=self, + original_ctx_manager=original_ctx_manager, + user_prompt=user_prompt, + model=model, + model_settings=model_settings, + is_streaming=True, + ) + + return wrapper + + +def _create_streaming_events_wrapper(original_func): + # type: (Callable[..., Any]) -> Callable[..., Any] + """ + Wraps run_stream_events method - no span needed as it delegates to run(). + + Note: run_stream_events internally calls self.run() with an event_stream_handler, + so the invoke_agent span will be created by the run() wrapper. + """ + + @wraps(original_func) + async def wrapper(self, *args, **kwargs): + # type: (Any, *Any, **Any) -> Any + # Just call the original generator - it will call run() which has the instrumentation + try: + async for event in original_func(self, *args, **kwargs): + yield event + except Exception as exc: + _capture_exception(exc) + raise exc from None + + return wrapper + + +def _patch_agent_run(): + # type: () -> None + """ + Patches the Agent run methods to create spans for agent execution. + + This patches both non-streaming (run, run_sync) and streaming + (run_stream, run_stream_events) methods. + """ + + # Store original methods + original_run = Agent.run + original_run_stream = Agent.run_stream + original_run_stream_events = Agent.run_stream_events + + # Wrap and apply patches for non-streaming methods + Agent.run = _create_run_wrapper(original_run, is_streaming=False) + + # Wrap and apply patches for streaming methods + Agent.run_stream = _create_streaming_wrapper(original_run_stream) + Agent.run_stream_events = _create_streaming_events_wrapper( + original_run_stream_events + ) diff --git a/sentry_sdk/integrations/pydantic_ai/patches/graph_nodes.py b/sentry_sdk/integrations/pydantic_ai/patches/graph_nodes.py new file mode 100644 index 0000000000..e10770d357 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/patches/graph_nodes.py @@ -0,0 +1,105 @@ +from contextlib import asynccontextmanager +from functools import wraps + +import sentry_sdk + +from ..spans import ( + ai_client_span, + update_ai_client_span, +) +from pydantic_ai._agent_graph import ModelRequestNode # type: ignore + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable + + +def _extract_span_data(node, ctx): + # type: (Any, Any) -> tuple[list[Any], Any, Any] + """Extract common data needed for creating chat spans. + + Returns: + Tuple of (messages, model, model_settings) + """ + # Extract model and settings from context + model = None + model_settings = None + if hasattr(ctx, "deps"): + model = getattr(ctx.deps, "model", None) + model_settings = getattr(ctx.deps, "model_settings", None) + + # Build full message list: history + current request + messages = [] + if hasattr(ctx, "state") and hasattr(ctx.state, "message_history"): + messages.extend(ctx.state.message_history) + + current_request = getattr(node, "request", None) + if current_request: + messages.append(current_request) + + return messages, model, model_settings + + +def _patch_graph_nodes(): + # type: () -> None + """ + Patches the graph node execution to create appropriate spans. + + ModelRequestNode -> Creates ai_client span for model requests + CallToolsNode -> Handles tool calls (spans created in tool patching) + """ + + # Patch ModelRequestNode to create ai_client spans + original_model_request_run = ModelRequestNode.run + + @wraps(original_model_request_run) + async def wrapped_model_request_run(self, ctx): + # type: (Any, Any) -> Any + messages, model, model_settings = _extract_span_data(self, ctx) + + with ai_client_span(messages, None, model, model_settings) as span: + result = await original_model_request_run(self, ctx) + + # Extract response from result if available + model_response = None + if hasattr(result, "model_response"): + model_response = result.model_response + + update_ai_client_span(span, model_response) + return result + + ModelRequestNode.run = wrapped_model_request_run + + # Patch ModelRequestNode.stream for streaming requests + original_model_request_stream = ModelRequestNode.stream + + def create_wrapped_stream(original_stream_method): + # type: (Callable[..., Any]) -> Callable[..., Any] + """Create a wrapper for ModelRequestNode.stream that creates chat spans.""" + + @asynccontextmanager + @wraps(original_stream_method) + async def wrapped_model_request_stream(self, ctx): + # type: (Any, Any) -> Any + messages, model, model_settings = _extract_span_data(self, ctx) + + # Create chat span for streaming request + with ai_client_span(messages, None, model, model_settings) as span: + # Call the original stream method + async with original_stream_method(self, ctx) as stream: + yield stream + + # After streaming completes, update span with response data + # The ModelRequestNode stores the final response in _result + model_response = None + if hasattr(self, "_result") and self._result is not None: + # _result is a NextNode containing the model_response + if hasattr(self._result, "model_response"): + model_response = self._result.model_response + + update_ai_client_span(span, model_response) + + return wrapped_model_request_stream + + ModelRequestNode.stream = create_wrapped_stream(original_model_request_stream) diff --git a/sentry_sdk/integrations/pydantic_ai/patches/model_request.py b/sentry_sdk/integrations/pydantic_ai/patches/model_request.py new file mode 100644 index 0000000000..f4676654cd --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/patches/model_request.py @@ -0,0 +1,35 @@ +from functools import wraps +from typing import TYPE_CHECKING + +from pydantic_ai import models # type: ignore + +from ..spans import ai_client_span, update_ai_client_span + + +if TYPE_CHECKING: + from typing import Any + + +def _patch_model_request(): + # type: () -> None + """ + Patches model request execution to create AI client spans. + + In pydantic-ai, model requests are handled through the Model interface. + We need to patch the request method on models to create spans. + """ + + # Patch the base Model class's request method + if hasattr(models, "Model"): + original_request = models.Model.request + + @wraps(original_request) + async def wrapped_request(self, messages, *args, **kwargs): + # type: (Any, Any, *Any, **Any) -> Any + # Pass all messages (full conversation history) + with ai_client_span(messages, None, self, None) as span: + result = await original_request(self, messages, *args, **kwargs) + update_ai_client_span(span, result) + return result + + models.Model.request = wrapped_request diff --git a/sentry_sdk/integrations/pydantic_ai/patches/tools.py b/sentry_sdk/integrations/pydantic_ai/patches/tools.py new file mode 100644 index 0000000000..41708c1be9 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/patches/tools.py @@ -0,0 +1,75 @@ +from functools import wraps + +from pydantic_ai._tool_manager import ToolManager # type: ignore + +import sentry_sdk + +from ..spans import execute_tool_span, update_execute_tool_span + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any + +try: + from pydantic_ai.mcp import MCPServer # type: ignore + + HAS_MCP = True +except ImportError: + HAS_MCP = False + + +def _patch_tool_execution(): + # type: () -> None + """ + Patch ToolManager._call_tool to create execute_tool spans. + + This is the single point where ALL tool calls flow through in pydantic_ai, + regardless of toolset type (function, MCP, combined, wrapper, etc.). + + By patching here, we avoid: + - Patching multiple toolset classes + - Dealing with signature mismatches from instrumented MCP servers + - Complex nested toolset handling + """ + + original_call_tool = ToolManager._call_tool + + @wraps(original_call_tool) + async def wrapped_call_tool(self, call, allow_partial, wrap_validation_errors): + # type: (Any, Any, bool, bool) -> Any + + # Extract tool info before calling original + name = call.tool_name + tool = self.tools.get(name) if self.tools else None + + # Determine tool type by checking tool.toolset + tool_type = "function" # default + if tool and HAS_MCP and isinstance(tool.toolset, MCPServer): + tool_type = "mcp" + + # Get agent from Sentry scope + current_span = sentry_sdk.get_current_span() + if current_span and tool: + agent_data = ( + sentry_sdk.get_current_scope()._contexts.get("pydantic_ai_agent") or {} + ) + agent = agent_data.get("_agent") + + # Get args for span (before validation) + # call.args can be a string (JSON) or dict + args_dict = call.args if isinstance(call.args, dict) else {} + + with execute_tool_span(name, args_dict, agent, tool_type=tool_type) as span: + result = await original_call_tool( + self, call, allow_partial, wrap_validation_errors + ) + update_execute_tool_span(span, result) + return result + + # No span context - just call original + return await original_call_tool( + self, call, allow_partial, wrap_validation_errors + ) + + ToolManager._call_tool = wrapped_call_tool diff --git a/sentry_sdk/integrations/pydantic_ai/spans/__init__.py b/sentry_sdk/integrations/pydantic_ai/spans/__init__.py new file mode 100644 index 0000000000..574046d645 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/spans/__init__.py @@ -0,0 +1,3 @@ +from .ai_client import ai_client_span, update_ai_client_span # noqa: F401 +from .execute_tool import execute_tool_span, update_execute_tool_span # noqa: F401 +from .invoke_agent import invoke_agent_span, update_invoke_agent_span # noqa: F401 diff --git a/sentry_sdk/integrations/pydantic_ai/spans/ai_client.py b/sentry_sdk/integrations/pydantic_ai/spans/ai_client.py new file mode 100644 index 0000000000..735e814acd --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/spans/ai_client.py @@ -0,0 +1,253 @@ +import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.utils import safe_serialize + +from ..consts import SPAN_ORIGIN +from ..utils import ( + _set_agent_data, + _set_available_tools, + _set_model_data, + _should_send_prompts, + _get_model_name, +) + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, List, Dict + from pydantic_ai.usage import RequestUsage # type: ignore + +try: + from pydantic_ai.messages import ( # type: ignore + BaseToolCallPart, + BaseToolReturnPart, + SystemPromptPart, + UserPromptPart, + TextPart, + ThinkingPart, + ) +except ImportError: + # Fallback if these classes are not available + BaseToolCallPart = None + BaseToolReturnPart = None + SystemPromptPart = None + UserPromptPart = None + TextPart = None + ThinkingPart = None + + +def _set_usage_data(span, usage): + # type: (sentry_sdk.tracing.Span, RequestUsage) -> None + """Set token usage data on a span.""" + if usage is None: + return + + if hasattr(usage, "input_tokens") and usage.input_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_INPUT_TOKENS, usage.input_tokens) + + if hasattr(usage, "output_tokens") and usage.output_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_OUTPUT_TOKENS, usage.output_tokens) + + if hasattr(usage, "total_tokens") and usage.total_tokens is not None: + span.set_data(SPANDATA.GEN_AI_USAGE_TOTAL_TOKENS, usage.total_tokens) + + +def _set_input_messages(span, messages): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Set input messages data on a span.""" + if not _should_send_prompts(): + return + + if not messages: + return + + try: + formatted_messages = [] + system_prompt = None + + # Extract system prompt from any ModelRequest with instructions + for msg in messages: + if hasattr(msg, "instructions") and msg.instructions: + system_prompt = msg.instructions + break + + # Add system prompt as first message if present + if system_prompt: + formatted_messages.append( + {"role": "system", "content": [{"type": "text", "text": system_prompt}]} + ) + + for msg in messages: + if hasattr(msg, "parts"): + for part in msg.parts: + role = "user" + # Use isinstance checks with proper base classes + if SystemPromptPart and isinstance(part, SystemPromptPart): + role = "system" + elif ( + (TextPart and isinstance(part, TextPart)) + or (ThinkingPart and isinstance(part, ThinkingPart)) + or (BaseToolCallPart and isinstance(part, BaseToolCallPart)) + ): + role = "assistant" + elif BaseToolReturnPart and isinstance(part, BaseToolReturnPart): + role = "tool" + + content = [] # type: List[Dict[str, Any] | str] + tool_calls = None + tool_call_id = None + + # Handle ToolCallPart (assistant requesting tool use) + if BaseToolCallPart and isinstance(part, BaseToolCallPart): + tool_call_data = {} + if hasattr(part, "tool_name"): + tool_call_data["name"] = part.tool_name + if hasattr(part, "args"): + tool_call_data["arguments"] = safe_serialize(part.args) + if tool_call_data: + tool_calls = [tool_call_data] + # Handle ToolReturnPart (tool result) + elif BaseToolReturnPart and isinstance(part, BaseToolReturnPart): + if hasattr(part, "tool_name"): + tool_call_id = part.tool_name + if hasattr(part, "content"): + content.append({"type": "text", "text": str(part.content)}) + # Handle regular content + elif hasattr(part, "content"): + if isinstance(part.content, str): + content.append({"type": "text", "text": part.content}) + elif isinstance(part.content, list): + for item in part.content: + if isinstance(item, str): + content.append({"type": "text", "text": item}) + else: + content.append(safe_serialize(item)) + else: + content.append({"type": "text", "text": str(part.content)}) + + # Add message if we have content or tool calls + if content or tool_calls: + message = {"role": role} # type: Dict[str, Any] + if content: + message["content"] = content + if tool_calls: + message["tool_calls"] = tool_calls + if tool_call_id: + message["tool_call_id"] = tool_call_id + formatted_messages.append(message) + + if formatted_messages: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, formatted_messages, unpack=False + ) + except Exception: + # If we fail to format messages, just skip it + pass + + +def _set_output_data(span, response): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Set output data on a span.""" + if not _should_send_prompts(): + return + + if not response: + return + + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_MODEL, response.model_name) + try: + # Extract text from ModelResponse + if hasattr(response, "parts"): + texts = [] + tool_calls = [] + + for part in response.parts: + if TextPart and isinstance(part, TextPart) and hasattr(part, "content"): + texts.append(part.content) + elif BaseToolCallPart and isinstance(part, BaseToolCallPart): + tool_call_data = { + "type": "function", + } + if hasattr(part, "tool_name"): + tool_call_data["name"] = part.tool_name + if hasattr(part, "args"): + tool_call_data["arguments"] = safe_serialize(part.args) + tool_calls.append(tool_call_data) + + if texts: + set_data_normalized(span, SPANDATA.GEN_AI_RESPONSE_TEXT, texts) + + if tool_calls: + span.set_data( + SPANDATA.GEN_AI_RESPONSE_TOOL_CALLS, safe_serialize(tool_calls) + ) + + except Exception: + # If we fail to format output, just skip it + pass + + +def ai_client_span(messages, agent, model, model_settings): + # type: (Any, Any, Any, Any) -> sentry_sdk.tracing.Span + """Create a span for an AI client call (model request). + + Args: + messages: Full conversation history (list of messages) + agent: Agent object + model: Model object + model_settings: Model settings + """ + # Determine model name for span name + model_obj = model + if agent and hasattr(agent, "model"): + model_obj = agent.model + + model_name = _get_model_name(model_obj) or "unknown" + + span = sentry_sdk.start_span( + op=OP.GEN_AI_CHAT, + name=f"chat {model_name}", + origin=SPAN_ORIGIN, + ) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") + + _set_agent_data(span, agent) + _set_model_data(span, model, model_settings) + + # Set streaming flag + agent_data = sentry_sdk.get_current_scope()._contexts.get("pydantic_ai_agent") or {} + is_streaming = agent_data.get("_streaming", False) + span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, is_streaming) + + # Add available tools if agent is available + agent_obj = agent + if not agent_obj: + # Try to get from Sentry scope + agent_data = ( + sentry_sdk.get_current_scope()._contexts.get("pydantic_ai_agent") or {} + ) + agent_obj = agent_data.get("_agent") + + _set_available_tools(span, agent_obj) + + # Set input messages (full conversation history) + if messages: + _set_input_messages(span, messages) + + return span + + +def update_ai_client_span(span, model_response): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Update the AI client span with response data.""" + if not span: + return + + # Set usage data if available + if model_response and hasattr(model_response, "usage"): + _set_usage_data(span, model_response.usage) + + # Set output data + _set_output_data(span, model_response) diff --git a/sentry_sdk/integrations/pydantic_ai/spans/execute_tool.py b/sentry_sdk/integrations/pydantic_ai/spans/execute_tool.py new file mode 100644 index 0000000000..2094c53a40 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/spans/execute_tool.py @@ -0,0 +1,49 @@ +import sentry_sdk +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.utils import safe_serialize + +from ..consts import SPAN_ORIGIN +from ..utils import _set_agent_data, _should_send_prompts + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any + + +def execute_tool_span(tool_name, tool_args, agent, tool_type="function"): + # type: (str, Any, Any, str) -> sentry_sdk.tracing.Span + """Create a span for tool execution. + + Args: + tool_name: The name of the tool being executed + tool_args: The arguments passed to the tool + agent: The agent executing the tool + tool_type: The type of tool ("function" for regular tools, "mcp" for MCP services) + """ + span = sentry_sdk.start_span( + op=OP.GEN_AI_EXECUTE_TOOL, + name=f"execute_tool {tool_name}", + origin=SPAN_ORIGIN, + ) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "execute_tool") + span.set_data(SPANDATA.GEN_AI_TOOL_TYPE, tool_type) + span.set_data(SPANDATA.GEN_AI_TOOL_NAME, tool_name) + + _set_agent_data(span, agent) + + if _should_send_prompts() and tool_args is not None: + span.set_data(SPANDATA.GEN_AI_TOOL_INPUT, safe_serialize(tool_args)) + + return span + + +def update_execute_tool_span(span, result): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Update the execute tool span with the result.""" + if not span: + return + + if _should_send_prompts() and result is not None: + span.set_data(SPANDATA.GEN_AI_TOOL_OUTPUT, safe_serialize(result)) diff --git a/sentry_sdk/integrations/pydantic_ai/spans/invoke_agent.py b/sentry_sdk/integrations/pydantic_ai/spans/invoke_agent.py new file mode 100644 index 0000000000..d6fb86918c --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/spans/invoke_agent.py @@ -0,0 +1,112 @@ +import sentry_sdk +from sentry_sdk.ai.utils import get_start_span_function, set_data_normalized +from sentry_sdk.consts import OP, SPANDATA + +from ..consts import SPAN_ORIGIN +from ..utils import ( + _set_agent_data, + _set_available_tools, + _set_model_data, + _should_send_prompts, +) + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any + + +def invoke_agent_span(user_prompt, agent, model, model_settings): + # type: (Any, Any, Any, Any) -> sentry_sdk.tracing.Span + """Create a span for invoking the agent.""" + # Determine agent name for span + name = "agent" + if agent and getattr(agent, "name", None): + name = agent.name + + span = get_start_span_function()( + op=OP.GEN_AI_INVOKE_AGENT, + name=f"invoke_agent {name}", + origin=SPAN_ORIGIN, + ) + + span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "invoke_agent") + + _set_agent_data(span, agent) + _set_model_data(span, model, model_settings) + _set_available_tools(span, agent) + + # Add user prompt and system prompts if available and prompts are enabled + if _should_send_prompts(): + messages = [] + + # Add system prompts (both instructions and system_prompt) + system_texts = [] + + if agent: + # Check for system_prompt + system_prompts = getattr(agent, "_system_prompts", None) or [] + for prompt in system_prompts: + if isinstance(prompt, str): + system_texts.append(prompt) + + # Check for instructions (stored in _instructions) + instructions = getattr(agent, "_instructions", None) + if instructions: + if isinstance(instructions, str): + system_texts.append(instructions) + elif isinstance(instructions, (list, tuple)): + for instr in instructions: + if isinstance(instr, str): + system_texts.append(instr) + elif callable(instr): + # Skip dynamic/callable instructions + pass + + # Add all system texts as system messages + for system_text in system_texts: + messages.append( + { + "content": [{"text": system_text, "type": "text"}], + "role": "system", + } + ) + + # Add user prompt + if user_prompt: + if isinstance(user_prompt, str): + messages.append( + { + "content": [{"text": user_prompt, "type": "text"}], + "role": "user", + } + ) + elif isinstance(user_prompt, list): + # Handle list of user content + content = [] + for item in user_prompt: + if isinstance(item, str): + content.append({"text": item, "type": "text"}) + if content: + messages.append( + { + "content": content, + "role": "user", + } + ) + + if messages: + set_data_normalized( + span, SPANDATA.GEN_AI_REQUEST_MESSAGES, messages, unpack=False + ) + + return span + + +def update_invoke_agent_span(span, output): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Update and close the invoke agent span.""" + if span and _should_send_prompts() and output: + set_data_normalized( + span, SPANDATA.GEN_AI_RESPONSE_TEXT, str(output), unpack=False + ) diff --git a/sentry_sdk/integrations/pydantic_ai/utils.py b/sentry_sdk/integrations/pydantic_ai/utils.py new file mode 100644 index 0000000000..3f58869857 --- /dev/null +++ b/sentry_sdk/integrations/pydantic_ai/utils.py @@ -0,0 +1,175 @@ +import sentry_sdk +from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.consts import SPANDATA +from sentry_sdk.scope import should_send_default_pii +from sentry_sdk.utils import safe_serialize + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, List, Dict + from pydantic_ai.usage import RequestUsage # type: ignore + + +def _should_send_prompts(): + # type: () -> bool + """ + Check if prompts should be sent to Sentry. + + This checks both send_default_pii and the include_prompts integration setting. + """ + if not should_send_default_pii(): + return False + + from . import PydanticAIIntegration + + # Get the integration instance from the client + integration = sentry_sdk.get_client().get_integration(PydanticAIIntegration) + + if integration is None: + return False + + return getattr(integration, "include_prompts", False) + + +def _set_agent_data(span, agent): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Set agent-related data on a span. + + Args: + span: The span to set data on + agent: Agent object (can be None, will try to get from Sentry scope if not provided) + """ + # Extract agent name from agent object or Sentry scope + agent_obj = agent + if not agent_obj: + # Try to get from Sentry scope + agent_data = ( + sentry_sdk.get_current_scope()._contexts.get("pydantic_ai_agent") or {} + ) + agent_obj = agent_data.get("_agent") + + if agent_obj and hasattr(agent_obj, "name") and agent_obj.name: + span.set_data(SPANDATA.GEN_AI_AGENT_NAME, agent_obj.name) + + +def _get_model_name(model_obj): + # type: (Any) -> str | None + """Extract model name from a model object. + + Args: + model_obj: Model object to extract name from + + Returns: + Model name string or None if not found + """ + if not model_obj: + return None + + if hasattr(model_obj, "model_name"): + return model_obj.model_name + elif hasattr(model_obj, "name"): + try: + return model_obj.name() + except Exception: + return str(model_obj) + elif isinstance(model_obj, str): + return model_obj + else: + return str(model_obj) + + +def _set_model_data(span, model, model_settings): + # type: (sentry_sdk.tracing.Span, Any, Any) -> None + """Set model-related data on a span. + + Args: + span: The span to set data on + model: Model object (can be None, will try to get from agent if not provided) + model_settings: Model settings (can be None, will try to get from agent if not provided) + """ + # Try to get agent from Sentry scope if we need it + agent_data = sentry_sdk.get_current_scope()._contexts.get("pydantic_ai_agent") or {} + agent_obj = agent_data.get("_agent") + + # Extract model information + model_obj = model + if not model_obj and agent_obj and hasattr(agent_obj, "model"): + model_obj = agent_obj.model + + if model_obj: + # Set system from model + if hasattr(model_obj, "system"): + span.set_data(SPANDATA.GEN_AI_SYSTEM, model_obj.system) + + # Set model name + model_name = _get_model_name(model_obj) + if model_name: + span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + + # Extract model settings + settings = model_settings + if not settings and agent_obj and hasattr(agent_obj, "model_settings"): + settings = agent_obj.model_settings + + if settings: + settings_map = { + "max_tokens": SPANDATA.GEN_AI_REQUEST_MAX_TOKENS, + "temperature": SPANDATA.GEN_AI_REQUEST_TEMPERATURE, + "top_p": SPANDATA.GEN_AI_REQUEST_TOP_P, + "frequency_penalty": SPANDATA.GEN_AI_REQUEST_FREQUENCY_PENALTY, + "presence_penalty": SPANDATA.GEN_AI_REQUEST_PRESENCE_PENALTY, + } + + # ModelSettings is a TypedDict (dict at runtime), so use dict access + if isinstance(settings, dict): + for setting_name, spandata_key in settings_map.items(): + value = settings.get(setting_name) + if value is not None: + span.set_data(spandata_key, value) + else: + # Fallback for object-style settings + for setting_name, spandata_key in settings_map.items(): + if hasattr(settings, setting_name): + value = getattr(settings, setting_name) + if value is not None: + span.set_data(spandata_key, value) + + +def _set_available_tools(span, agent): + # type: (sentry_sdk.tracing.Span, Any) -> None + """Set available tools data on a span from an agent's function toolset. + + Args: + span: The span to set data on + agent: Agent object with _function_toolset attribute + """ + if not agent or not hasattr(agent, "_function_toolset"): + return + + try: + tools = [] + # Get tools from the function toolset + if hasattr(agent._function_toolset, "tools"): + for tool_name, tool in agent._function_toolset.tools.items(): + tool_info = {"name": tool_name} + + # Add description from function_schema if available + if hasattr(tool, "function_schema"): + schema = tool.function_schema + if getattr(schema, "description", None): + tool_info["description"] = schema.description + + # Add parameters from json_schema + if getattr(schema, "json_schema", None): + tool_info["parameters"] = schema.json_schema + + tools.append(tool_info) + + if tools: + span.set_data( + SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) + ) + except Exception: + # If we can't extract tools, just skip it + pass diff --git a/setup.py b/setup.py index cf66f30f23..b1662204fb 100644 --- a/setup.py +++ b/setup.py @@ -73,6 +73,7 @@ def get_file_text(file_name): "opentelemetry": ["opentelemetry-distro>=0.35b0"], "opentelemetry-experimental": ["opentelemetry-distro"], "pure-eval": ["pure_eval", "executing", "asttokens"], + "pydantic_ai": ["pydantic-ai>=1.0.0"], "pymongo": ["pymongo>=3.1"], "pyspark": ["pyspark>=2.4.4"], "quart": ["quart>=0.16.1", "blinker>=1.1"], diff --git a/tests/integrations/pydantic_ai/__init__.py b/tests/integrations/pydantic_ai/__init__.py new file mode 100644 index 0000000000..3a2ad11c0c --- /dev/null +++ b/tests/integrations/pydantic_ai/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("pydantic_ai") diff --git a/tests/integrations/pydantic_ai/test_pydantic_ai.py b/tests/integrations/pydantic_ai/test_pydantic_ai.py new file mode 100644 index 0000000000..578eca2bf6 --- /dev/null +++ b/tests/integrations/pydantic_ai/test_pydantic_ai.py @@ -0,0 +1,2425 @@ +import asyncio +import pytest + +from sentry_sdk.integrations.pydantic_ai import PydanticAIIntegration + +from pydantic_ai import Agent +from pydantic_ai.models.test import TestModel + + +@pytest.fixture +def test_agent(): + """Create a test agent with model settings.""" + return Agent( + "test", + name="test_agent", + system_prompt="You are a helpful test assistant.", + ) + + +@pytest.fixture +def test_agent_with_settings(): + """Create a test agent with explicit model settings.""" + from pydantic_ai import ModelSettings + + return Agent( + "test", + name="test_agent_settings", + system_prompt="You are a test assistant with settings.", + model_settings=ModelSettings( + temperature=0.7, + max_tokens=100, + top_p=0.9, + ), + ) + + +@pytest.mark.asyncio +async def test_agent_run_async(sentry_init, capture_events, test_agent): + """ + Test that the integration creates spans for async agent runs. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + result = await test_agent.run("Test input") + + assert result is not None + assert result.output is not None + + (transaction,) = events + spans = transaction["spans"] + + # Verify transaction (the transaction IS the invoke_agent span) + assert transaction["transaction"] == "invoke_agent test_agent" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.pydantic_ai" + + # The transaction itself should have invoke_agent data + assert transaction["contexts"]["trace"]["op"] == "gen_ai.invoke_agent" + + # Find child span types (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + # Check chat span + chat_span = chat_spans[0] + assert "chat" in chat_span["description"] + assert chat_span["data"]["gen_ai.operation.name"] == "chat" + assert chat_span["data"]["gen_ai.response.streaming"] is False + assert "gen_ai.request.messages" in chat_span["data"] + assert "gen_ai.usage.input_tokens" in chat_span["data"] + assert "gen_ai.usage.output_tokens" in chat_span["data"] + + +def test_agent_run_sync(sentry_init, capture_events, test_agent): + """ + Test that the integration creates spans for sync agent runs. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + result = test_agent.run_sync("Test input") + + assert result is not None + assert result.output is not None + + (transaction,) = events + spans = transaction["spans"] + + # Verify transaction + assert transaction["transaction"] == "invoke_agent test_agent" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.pydantic_ai" + + # Find span types + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + # Verify streaming flag is False for sync + for chat_span in chat_spans: + assert chat_span["data"]["gen_ai.response.streaming"] is False + + +@pytest.mark.asyncio +async def test_agent_run_stream(sentry_init, capture_events, test_agent): + """ + Test that the integration creates spans for streaming agent runs. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + async with test_agent.run_stream("Test input") as result: + # Consume the stream + async for _ in result.stream_output(): + pass + + (transaction,) = events + spans = transaction["spans"] + + # Verify transaction + assert transaction["transaction"] == "invoke_agent test_agent" + assert transaction["contexts"]["trace"]["origin"] == "auto.ai.pydantic_ai" + + # Find chat spans + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + # Verify streaming flag is True for streaming + for chat_span in chat_spans: + assert chat_span["data"]["gen_ai.response.streaming"] is True + assert "gen_ai.request.messages" in chat_span["data"] + assert "gen_ai.usage.input_tokens" in chat_span["data"] + # Streaming responses should still have output data + assert ( + "gen_ai.response.text" in chat_span["data"] + or "gen_ai.response.model" in chat_span["data"] + ) + + +@pytest.mark.asyncio +async def test_agent_run_stream_events(sentry_init, capture_events, test_agent): + """ + Test that run_stream_events creates spans (it uses run internally, so non-streaming). + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Consume all events + async for _ in test_agent.run_stream_events("Test input"): + pass + + (transaction,) = events + + # Verify transaction + assert transaction["transaction"] == "invoke_agent test_agent" + + # Find chat spans + spans = transaction["spans"] + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + # run_stream_events uses run() internally, so streaming should be False + for chat_span in chat_spans: + assert chat_span["data"]["gen_ai.response.streaming"] is False + + +@pytest.mark.asyncio +async def test_agent_with_tools(sentry_init, capture_events, test_agent): + """ + Test that tool execution creates execute_tool spans. + """ + + @test_agent.tool_plain + def add_numbers(a: int, b: int) -> int: + """Add two numbers together.""" + return a + b + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + result = await test_agent.run("What is 5 + 3?") + + assert result is not None + + (transaction,) = events + spans = transaction["spans"] + + # Find child span types (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + tool_spans = [s for s in spans if s["op"] == "gen_ai.execute_tool"] + + # Should have tool spans + assert len(tool_spans) >= 1 + + # Check tool span + tool_span = tool_spans[0] + assert "execute_tool" in tool_span["description"] + assert tool_span["data"]["gen_ai.operation.name"] == "execute_tool" + assert tool_span["data"]["gen_ai.tool.type"] == "function" + assert tool_span["data"]["gen_ai.tool.name"] == "add_numbers" + assert "gen_ai.tool.input" in tool_span["data"] + assert "gen_ai.tool.output" in tool_span["data"] + + # Check chat spans have available_tools + for chat_span in chat_spans: + assert "gen_ai.request.available_tools" in chat_span["data"] + available_tools_str = chat_span["data"]["gen_ai.request.available_tools"] + # Available tools is serialized as a string + assert "add_numbers" in available_tools_str + + +@pytest.mark.asyncio +async def test_agent_with_tools_streaming(sentry_init, capture_events, test_agent): + """ + Test that tool execution works correctly with streaming. + """ + + @test_agent.tool_plain + def multiply(a: int, b: int) -> int: + """Multiply two numbers.""" + return a * b + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + async with test_agent.run_stream("What is 7 times 8?") as result: + async for _ in result.stream_output(): + pass + + (transaction,) = events + spans = transaction["spans"] + + # Find span types + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + tool_spans = [s for s in spans if s["op"] == "gen_ai.execute_tool"] + + # Should have tool spans + assert len(tool_spans) >= 1 + + # Verify streaming flag is True + for chat_span in chat_spans: + assert chat_span["data"]["gen_ai.response.streaming"] is True + + # Check tool span + tool_span = tool_spans[0] + assert tool_span["data"]["gen_ai.tool.name"] == "multiply" + assert "gen_ai.tool.input" in tool_span["data"] + assert "gen_ai.tool.output" in tool_span["data"] + + +@pytest.mark.asyncio +async def test_model_settings(sentry_init, capture_events, test_agent_with_settings): + """ + Test that model settings are captured in spans. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + await test_agent_with_settings.run("Test input") + + (transaction,) = events + spans = transaction["spans"] + + # Find chat span + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + chat_span = chat_spans[0] + # Check that model settings are captured + assert chat_span["data"].get("gen_ai.request.temperature") == 0.7 + assert chat_span["data"].get("gen_ai.request.max_tokens") == 100 + assert chat_span["data"].get("gen_ai.request.top_p") == 0.9 + + +@pytest.mark.asyncio +async def test_system_prompt_in_messages(sentry_init, capture_events): + """ + Test that system prompts are included as the first message. + """ + agent = Agent( + "test", + name="test_system", + system_prompt="You are a helpful assistant specialized in testing.", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await agent.run("Hello") + + (transaction,) = events + spans = transaction["spans"] + + # The transaction IS the invoke_agent span, check for messages in chat spans instead + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + chat_span = chat_spans[0] + messages_str = chat_span["data"]["gen_ai.request.messages"] + + # Messages is serialized as a string + # Should contain system role and helpful assistant text + assert "system" in messages_str + assert "helpful assistant" in messages_str + + +@pytest.mark.asyncio +async def test_error_handling(sentry_init, capture_events): + """ + Test error handling in agent execution. + """ + # Use a simpler test that doesn't cause tool failures + # as pydantic-ai has complex error handling for tool errors + agent = Agent( + "test", + name="test_error", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + # Simple run that should succeed + await agent.run("Hello") + + # At minimum, we should have a transaction + assert len(events) >= 1 + transaction = [e for e in events if e.get("type") == "transaction"][0] + assert transaction["transaction"] == "invoke_agent test_error" + # Transaction should complete successfully (status key may not exist if no error) + trace_status = transaction["contexts"]["trace"].get("status") + assert trace_status != "error" # Could be None or some other status + + +@pytest.mark.asyncio +async def test_without_pii(sentry_init, capture_events, test_agent): + """ + Test that PII is not captured when send_default_pii is False. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=False, + ) + + events = capture_events() + + await test_agent.run("Sensitive input") + + (transaction,) = events + spans = transaction["spans"] + + # Find child spans (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Verify that messages and response text are not captured + for span in chat_spans: + assert "gen_ai.request.messages" not in span["data"] + assert "gen_ai.response.text" not in span["data"] + + +@pytest.mark.asyncio +async def test_without_pii_tools(sentry_init, capture_events, test_agent): + """ + Test that tool input/output are not captured when send_default_pii is False. + """ + + @test_agent.tool_plain + def sensitive_tool(data: str) -> str: + """A tool with sensitive data.""" + return f"Processed: {data}" + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=False, + ) + + events = capture_events() + + await test_agent.run("Use sensitive tool with private data") + + (transaction,) = events + spans = transaction["spans"] + + # Find tool spans + tool_spans = [s for s in spans if s["op"] == "gen_ai.execute_tool"] + + # If tool was executed, verify input/output are not captured + for tool_span in tool_spans: + assert "gen_ai.tool.input" not in tool_span["data"] + assert "gen_ai.tool.output" not in tool_span["data"] + + +@pytest.mark.asyncio +async def test_multiple_agents_concurrent(sentry_init, capture_events, test_agent): + """ + Test that multiple agents can run concurrently without interfering. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + async def run_agent(input_text): + return await test_agent.run(input_text) + + # Run 3 agents concurrently + results = await asyncio.gather(*[run_agent(f"Input {i}") for i in range(3)]) + + assert len(results) == 3 + assert len(events) == 3 + + # Verify each transaction is separate + for i, transaction in enumerate(events): + assert transaction["type"] == "transaction" + assert transaction["transaction"] == "invoke_agent test_agent" + # Each should have its own spans + assert len(transaction["spans"]) >= 1 + + +@pytest.mark.asyncio +async def test_message_history(sentry_init, capture_events): + """ + Test that full conversation history is captured in chat spans. + """ + agent = Agent( + "test", + name="test_history", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # First message + await agent.run("Hello, I'm Alice") + + # Second message with history + from pydantic_ai import messages + + history = [ + messages.UserPromptPart(content="Hello, I'm Alice"), + messages.ModelResponse( + parts=[messages.TextPart(content="Hello Alice! How can I help you?")], + model_name="test", + ), + ] + + await agent.run("What is my name?", message_history=history) + + # We should have 2 transactions + assert len(events) >= 2 + + # Check the second transaction has the full history + second_transaction = events[1] + spans = second_transaction["spans"] + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + if chat_spans: + chat_span = chat_spans[0] + if "gen_ai.request.messages" in chat_span["data"]: + messages_data = chat_span["data"]["gen_ai.request.messages"] + # Should have multiple messages including history + assert len(messages_data) > 1 + + +@pytest.mark.asyncio +async def test_gen_ai_system(sentry_init, capture_events, test_agent): + """ + Test that gen_ai.system is set from the model. + """ + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + await test_agent.run("Test input") + + (transaction,) = events + spans = transaction["spans"] + + # Find chat span + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + chat_span = chat_spans[0] + # gen_ai.system should be set from the model (TestModel -> 'test') + assert "gen_ai.system" in chat_span["data"] + assert chat_span["data"]["gen_ai.system"] == "test" + + +@pytest.mark.asyncio +async def test_include_prompts_false(sentry_init, capture_events, test_agent): + """ + Test that prompts are not captured when include_prompts=False. + """ + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=False)], + traces_sample_rate=1.0, + send_default_pii=True, # Even with PII enabled, prompts should not be captured + ) + + events = capture_events() + + await test_agent.run("Sensitive prompt") + + (transaction,) = events + spans = transaction["spans"] + + # Find child spans (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Verify that messages and response text are not captured + for span in chat_spans: + assert "gen_ai.request.messages" not in span["data"] + assert "gen_ai.response.text" not in span["data"] + + +@pytest.mark.asyncio +async def test_include_prompts_true(sentry_init, capture_events, test_agent): + """ + Test that prompts are captured when include_prompts=True (default). + """ + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await test_agent.run("Test prompt") + + (transaction,) = events + spans = transaction["spans"] + + # Find child spans (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Verify that messages are captured in chat spans + assert len(chat_spans) >= 1 + for chat_span in chat_spans: + assert "gen_ai.request.messages" in chat_span["data"] + + +@pytest.mark.asyncio +async def test_include_prompts_false_with_tools( + sentry_init, capture_events, test_agent +): + """ + Test that tool input/output are not captured when include_prompts=False. + """ + + @test_agent.tool_plain + def test_tool(value: int) -> int: + """A test tool.""" + return value * 2 + + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=False)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await test_agent.run("Use the test tool with value 5") + + (transaction,) = events + spans = transaction["spans"] + + # Find tool spans + tool_spans = [s for s in spans if s["op"] == "gen_ai.execute_tool"] + + # If tool was executed, verify input/output are not captured + for tool_span in tool_spans: + assert "gen_ai.tool.input" not in tool_span["data"] + assert "gen_ai.tool.output" not in tool_span["data"] + + +@pytest.mark.asyncio +async def test_include_prompts_requires_pii(sentry_init, capture_events, test_agent): + """ + Test that include_prompts requires send_default_pii=True. + """ + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=False, # PII disabled + ) + + events = capture_events() + + await test_agent.run("Test prompt") + + (transaction,) = events + spans = transaction["spans"] + + # Find child spans (invoke_agent is the transaction, not a child span) + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Even with include_prompts=True, if PII is disabled, messages should not be captured + for span in chat_spans: + assert "gen_ai.request.messages" not in span["data"] + assert "gen_ai.response.text" not in span["data"] + + +@pytest.mark.asyncio +async def test_mcp_tool_execution_spans(sentry_init, capture_events): + """ + Test that MCP (Model Context Protocol) tool calls create execute_tool spans. + + Tests MCP tools accessed through CombinedToolset, which is how they're typically + used in practice (when an agent combines regular functions with MCP servers). + """ + pytest.importorskip("mcp") + + from unittest.mock import AsyncMock, MagicMock + from pydantic_ai.mcp import MCPServerStdio + from pydantic_ai import Agent + from pydantic_ai.toolsets.combined import CombinedToolset + import sentry_sdk + + # Create mock MCP server + mock_server = MCPServerStdio( + command="python", + args=["-m", "test_server"], + ) + + # Mock the server's internal methods + mock_server._client = MagicMock() + mock_server._is_initialized = True + mock_server._server_info = MagicMock() + + # Mock tool call response + async def mock_send_request(request, response_type): + from mcp.types import CallToolResult, TextContent + + return CallToolResult( + content=[TextContent(type="text", text="MCP tool executed successfully")], + isError=False, + ) + + mock_server._client.send_request = mock_send_request + + # Mock context manager methods + async def mock_aenter(): + return mock_server + + async def mock_aexit(*args): + pass + + mock_server.__aenter__ = mock_aenter + mock_server.__aexit__ = mock_aexit + + # Mock _map_tool_result_part + async def mock_map_tool_result_part(part): + return part.text if hasattr(part, "text") else str(part) + + mock_server._map_tool_result_part = mock_map_tool_result_part + + # Create a CombinedToolset with the MCP server + # This simulates how MCP servers are typically used in practice + from pydantic_ai.toolsets.function import FunctionToolset + + function_toolset = FunctionToolset() + combined = CombinedToolset([function_toolset, mock_server]) + + # Create agent + agent = Agent( + "test", + name="test_mcp_agent", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Simulate MCP tool execution within a transaction through CombinedToolset + with sentry_sdk.start_transaction( + op="ai.run", name="invoke_agent test_mcp_agent" + ) as transaction: + # Set up the agent context + scope = sentry_sdk.get_current_scope() + scope._contexts["pydantic_ai_agent"] = { + "_agent": agent, + } + + # Create a mock tool that simulates an MCP tool from CombinedToolset + from pydantic_ai._run_context import RunContext + from pydantic_ai.result import RunUsage + from pydantic_ai.models.test import TestModel + from pydantic_ai.toolsets.combined import _CombinedToolsetTool + + ctx = RunContext( + deps=None, + model=TestModel(), + usage=RunUsage(), + retry=0, + tool_name="test_mcp_tool", + ) + + tool_name = "test_mcp_tool" + + # Create a tool that points to the MCP server + # This simulates how CombinedToolset wraps tools from different sources + tool = _CombinedToolsetTool( + toolset=combined, + tool_def=MagicMock(name=tool_name), + max_retries=0, + args_validator=MagicMock(), + source_toolset=mock_server, + source_tool=MagicMock(), + ) + + try: + await combined.call_tool(tool_name, {"query": "test"}, ctx, tool) + except Exception: + # MCP tool might raise if not fully mocked, that's okay + pass + + events_list = events + if len(events_list) == 0: + pytest.skip("No events captured, MCP test setup incomplete") + + (transaction,) = events_list + transaction["spans"] + + # Note: This test manually calls combined.call_tool which doesn't go through + # ToolManager._call_tool (which is what the integration patches). + # In real-world usage, MCP tools are called through agent.run() which uses ToolManager. + # This synthetic test setup doesn't trigger the integration's tool patches. + # We skip this test as it doesn't represent actual usage patterns. + pytest.skip( + "MCP test needs to be rewritten to use agent.run() instead of manually calling toolset methods" + ) + + +@pytest.mark.asyncio +async def test_context_cleanup_after_run(sentry_init, test_agent): + """ + Test that the pydantic_ai_agent context is properly cleaned up after agent execution. + """ + import sentry_sdk + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Verify context is not set before run + scope = sentry_sdk.get_current_scope() + assert "pydantic_ai_agent" not in scope._contexts + + # Run the agent + await test_agent.run("Test input") + + # Verify context is cleaned up after run + assert "pydantic_ai_agent" not in scope._contexts + + +def test_context_cleanup_after_run_sync(sentry_init, test_agent): + """ + Test that the pydantic_ai_agent context is properly cleaned up after sync agent execution. + """ + import sentry_sdk + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Verify context is not set before run + scope = sentry_sdk.get_current_scope() + assert "pydantic_ai_agent" not in scope._contexts + + # Run the agent synchronously + test_agent.run_sync("Test input") + + # Verify context is cleaned up after run + assert "pydantic_ai_agent" not in scope._contexts + + +@pytest.mark.asyncio +async def test_context_cleanup_after_streaming(sentry_init, test_agent): + """ + Test that the pydantic_ai_agent context is properly cleaned up after streaming execution. + """ + import sentry_sdk + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Verify context is not set before run + scope = sentry_sdk.get_current_scope() + assert "pydantic_ai_agent" not in scope._contexts + + # Run the agent with streaming + async with test_agent.run_stream("Test input") as result: + async for _ in result.stream_output(): + pass + + # Verify context is cleaned up after streaming completes + assert "pydantic_ai_agent" not in scope._contexts + + +@pytest.mark.asyncio +async def test_context_cleanup_on_error(sentry_init, test_agent): + """ + Test that the pydantic_ai_agent context is cleaned up even when an error occurs. + """ + import sentry_sdk + + # Create an agent with a tool that raises an error + @test_agent.tool_plain + def failing_tool() -> str: + """A tool that always fails.""" + raise ValueError("Tool error") + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Verify context is not set before run + scope = sentry_sdk.get_current_scope() + assert "pydantic_ai_agent" not in scope._contexts + + # Run the agent - this may or may not raise depending on pydantic-ai's error handling + try: + await test_agent.run("Use the failing tool") + except Exception: + pass + + # Verify context is cleaned up even if there was an error + assert "pydantic_ai_agent" not in scope._contexts + + +@pytest.mark.asyncio +async def test_context_isolation_concurrent_agents(sentry_init, test_agent): + """ + Test that concurrent agent executions maintain isolated contexts. + """ + import sentry_sdk + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Create a second agent + agent2 = Agent( + "test", + name="test_agent_2", + system_prompt="Second test agent.", + ) + + async def run_and_check_context(agent, agent_name): + """Run an agent and verify its context during and after execution.""" + # Before execution, context should not exist in the outer scope + outer_scope = sentry_sdk.get_current_scope() + + # Run the agent + await agent.run(f"Input for {agent_name}") + + # After execution, verify context is cleaned up + # Note: Due to isolation_scope, we can't easily check the inner scope here, + # but we can verify the outer scope remains clean + assert "pydantic_ai_agent" not in outer_scope._contexts + + return agent_name + + # Run both agents concurrently + results = await asyncio.gather( + run_and_check_context(test_agent, "agent1"), + run_and_check_context(agent2, "agent2"), + ) + + assert results == ["agent1", "agent2"] + + # Final check: outer scope should be clean + final_scope = sentry_sdk.get_current_scope() + assert "pydantic_ai_agent" not in final_scope._contexts + + +# ==================== Additional Coverage Tests ==================== + + +@pytest.mark.asyncio +async def test_invoke_agent_with_list_user_prompt(sentry_init, capture_events): + """ + Test that invoke_agent span handles list user prompts correctly. + """ + agent = Agent( + "test", + name="test_list_prompt", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Use a list as user prompt + await agent.run(["First part", "Second part"]) + + (transaction,) = events + + # Check that the invoke_agent transaction has messages data + # The invoke_agent is the transaction itself + if "gen_ai.request.messages" in transaction["contexts"]["trace"]["data"]: + messages_str = transaction["contexts"]["trace"]["data"][ + "gen_ai.request.messages" + ] + assert "First part" in messages_str + assert "Second part" in messages_str + + +@pytest.mark.asyncio +async def test_invoke_agent_with_instructions(sentry_init, capture_events): + """ + Test that invoke_agent span handles instructions correctly. + """ + from pydantic_ai import Agent + + # Create agent with instructions (can be string or list) + agent = Agent( + "test", + name="test_instructions", + ) + + # Add instructions via _instructions attribute (internal API) + agent._instructions = ["Instruction 1", "Instruction 2"] + agent._system_prompts = ["System prompt"] + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await agent.run("Test input") + + (transaction,) = events + + # Check that the invoke_agent transaction has messages data + if "gen_ai.request.messages" in transaction["contexts"]["trace"]["data"]: + messages_str = transaction["contexts"]["trace"]["data"][ + "gen_ai.request.messages" + ] + # Should contain both instructions and system prompts + assert "Instruction" in messages_str or "System prompt" in messages_str + + +@pytest.mark.asyncio +async def test_model_name_extraction_with_callable(sentry_init, capture_events): + """ + Test model name extraction when model has a callable name() method. + """ + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _get_model_name + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Test the utility function directly + mock_model = MagicMock() + # Remove model_name attribute so it checks name() next + del mock_model.model_name + mock_model.name = lambda: "custom-model-name" + + # Get model name - should call the callable name() + result = _get_model_name(mock_model) + + # Should return the result from callable + assert result == "custom-model-name" + + +@pytest.mark.asyncio +async def test_model_name_extraction_fallback_to_str(sentry_init, capture_events): + """ + Test model name extraction falls back to str() when no name attribute exists. + """ + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _get_model_name + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Test the utility function directly + mock_model = MagicMock() + # Remove name and model_name attributes + del mock_model.name + del mock_model.model_name + + # Get model name - should fall back to str() + result = _get_model_name(mock_model) + + # Should return string representation + assert result is not None + assert isinstance(result, str) + + +@pytest.mark.asyncio +async def test_model_settings_object_style(sentry_init, capture_events): + """ + Test that object-style model settings (non-dict) are handled correctly. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_model_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create mock settings object (not a dict) + mock_settings = MagicMock() + mock_settings.temperature = 0.8 + mock_settings.max_tokens = 200 + mock_settings.top_p = 0.95 + mock_settings.frequency_penalty = 0.5 + mock_settings.presence_penalty = 0.3 + + # Set model data with object-style settings + _set_model_data(span, None, mock_settings) + + span.finish() + + # Should not crash and should set the settings + assert transaction is not None + + +@pytest.mark.asyncio +async def test_usage_data_partial(sentry_init, capture_events): + """ + Test that usage data is correctly handled when only some fields are present. + """ + agent = Agent( + "test", + name="test_usage", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + await agent.run("Test input") + + (transaction,) = events + spans = transaction["spans"] + + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + assert len(chat_spans) >= 1 + + # Check that usage data fields exist (they may or may not be set depending on TestModel) + chat_span = chat_spans[0] + # At minimum, the span should have been created + assert chat_span is not None + + +@pytest.mark.asyncio +async def test_agent_data_from_scope(sentry_init, capture_events): + """ + Test that agent data can be retrieved from Sentry scope when not passed directly. + """ + import sentry_sdk + + agent = Agent( + "test", + name="test_scope_agent", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + # The integration automatically sets agent in scope during execution + await agent.run("Test input") + + (transaction,) = events + + # Verify agent name is captured + assert transaction["transaction"] == "invoke_agent test_scope_agent" + + +@pytest.mark.asyncio +async def test_available_tools_without_description( + sentry_init, capture_events, test_agent +): + """ + Test that available tools are captured even when description is missing. + """ + + @test_agent.tool_plain + def tool_without_desc(x: int) -> int: + # No docstring = no description + return x * 2 + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + await test_agent.run("Use the tool with 5") + + (transaction,) = events + spans = transaction["spans"] + + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + if chat_spans: + chat_span = chat_spans[0] + if "gen_ai.request.available_tools" in chat_span["data"]: + tools_str = chat_span["data"]["gen_ai.request.available_tools"] + assert "tool_without_desc" in tools_str + + +@pytest.mark.asyncio +async def test_output_with_tool_calls(sentry_init, capture_events, test_agent): + """ + Test that tool calls in model response are captured correctly. + """ + + @test_agent.tool_plain + def calc_tool(value: int) -> int: + """Calculate something.""" + return value + 10 + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + await test_agent.run("Use calc_tool with 5") + + (transaction,) = events + spans = transaction["spans"] + + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # At least one chat span should exist + assert len(chat_spans) >= 1 + + # Check if tool calls are captured in response + for chat_span in chat_spans: + # Tool calls may or may not be in response depending on TestModel behavior + # Just verify the span was created and has basic data + assert "gen_ai.operation.name" in chat_span["data"] + + +@pytest.mark.asyncio +async def test_message_formatting_with_different_parts(sentry_init, capture_events): + """ + Test that different message part types are handled correctly in ai_client span. + """ + from pydantic_ai import Agent, messages + + agent = Agent( + "test", + name="test_message_parts", + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Create message history with different part types + history = [ + messages.UserPromptPart(content="Hello"), + messages.ModelResponse( + parts=[ + messages.TextPart(content="Hi there!"), + ], + model_name="test", + ), + ] + + await agent.run("What did I say?", message_history=history) + + (transaction,) = events + spans = transaction["spans"] + + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Should have chat spans + assert len(chat_spans) >= 1 + + # Check that messages are captured + chat_span = chat_spans[0] + if "gen_ai.request.messages" in chat_span["data"]: + messages_data = chat_span["data"]["gen_ai.request.messages"] + # Should contain message history + assert messages_data is not None + + +@pytest.mark.asyncio +async def test_update_invoke_agent_span_with_none_output(sentry_init, capture_events): + """ + Test that update_invoke_agent_span handles None output gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.invoke_agent import ( + update_invoke_agent_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Update with None output - should not raise + update_invoke_agent_span(span, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_update_ai_client_span_with_none_response(sentry_init, capture_events): + """ + Test that update_ai_client_span handles None response gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import ( + update_ai_client_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Update with None response - should not raise + update_ai_client_span(span, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_agent_without_name(sentry_init, capture_events): + """ + Test that agent without a name is handled correctly. + """ + # Create agent without explicit name + agent = Agent("test") + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + events = capture_events() + + await agent.run("Test input") + + (transaction,) = events + + # Should still create transaction, just with default name + assert transaction["type"] == "transaction" + # Transaction name should be "invoke_agent agent" or similar default + assert "invoke_agent" in transaction["transaction"] + + +@pytest.mark.asyncio +async def test_model_response_without_parts(sentry_init, capture_events): + """ + Test handling of model response without parts attribute. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_output_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create mock response without parts + mock_response = MagicMock() + mock_response.model_name = "test-model" + del mock_response.parts # Remove parts attribute + + # Should not raise, just skip formatting + _set_output_data(span, mock_response) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_input_messages_error_handling(sentry_init, capture_events): + """ + Test that _set_input_messages handles errors gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_input_messages + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Pass invalid messages that would cause an error + invalid_messages = [object()] # Plain object without expected attributes + + # Should not raise, error is caught internally + _set_input_messages(span, invalid_messages) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_available_tools_error_handling(sentry_init, capture_events): + """ + Test that _set_available_tools handles errors gracefully. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_available_tools + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create mock agent with invalid toolset + mock_agent = MagicMock() + mock_agent._function_toolset.tools.items.side_effect = Exception("Error") + + # Should not raise, error is caught internally + _set_available_tools(span, mock_agent) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_usage_data_with_none_usage(sentry_init, capture_events): + """ + Test that _set_usage_data handles None usage gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_usage_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Pass None usage - should not raise + _set_usage_data(span, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_usage_data_with_partial_fields(sentry_init, capture_events): + """ + Test that _set_usage_data handles usage with only some fields. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_usage_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create usage object with only some fields + mock_usage = MagicMock() + mock_usage.input_tokens = 100 + mock_usage.output_tokens = None # Missing + mock_usage.total_tokens = 100 + + # Should only set the non-None fields + _set_usage_data(span, mock_usage) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_message_parts_with_tool_return(sentry_init, capture_events): + """ + Test that ToolReturnPart messages are handled correctly. + """ + from pydantic_ai import Agent, messages + + agent = Agent( + "test", + name="test_tool_return", + ) + + @agent.tool_plain + def test_tool(x: int) -> int: + """Test tool.""" + return x * 2 + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Run with history containing tool return + await agent.run("Use test_tool with 5") + + (transaction,) = events + spans = transaction["spans"] + + chat_spans = [s for s in spans if s["op"] == "gen_ai.chat"] + + # Should have chat spans + assert len(chat_spans) >= 1 + + +@pytest.mark.asyncio +async def test_message_parts_with_list_content(sentry_init, capture_events): + """ + Test that message parts with list content are handled correctly. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_input_messages + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create message with list content + mock_msg = MagicMock() + mock_part = MagicMock() + mock_part.content = ["item1", "item2", {"complex": "item"}] + mock_msg.parts = [mock_part] + mock_msg.instructions = None + + messages = [mock_msg] + + # Should handle list content + _set_input_messages(span, messages) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_output_data_with_text_and_tool_calls(sentry_init, capture_events): + """ + Test that _set_output_data handles both text and tool calls in response. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_output_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create mock response with both TextPart and ToolCallPart + from pydantic_ai import messages + + text_part = messages.TextPart(content="Here's the result") + tool_call_part = MagicMock() + tool_call_part.tool_name = "test_tool" + tool_call_part.args = {"x": 5} + + mock_response = MagicMock() + mock_response.model_name = "test-model" + mock_response.parts = [text_part, tool_call_part] + + # Should handle both text and tool calls + _set_output_data(span, mock_response) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_output_data_error_handling(sentry_init, capture_events): + """ + Test that _set_output_data handles errors in formatting gracefully. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_output_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create mock response that will cause error + mock_response = MagicMock() + mock_response.model_name = "test-model" + mock_response.parts = [MagicMock(side_effect=Exception("Error"))] + + # Should catch error and not crash + _set_output_data(span, mock_response) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_message_with_system_prompt_part(sentry_init, capture_events): + """ + Test that SystemPromptPart is handled with correct role. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_input_messages + from pydantic_ai import messages + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create message with SystemPromptPart + system_part = messages.SystemPromptPart(content="You are a helpful assistant") + + mock_msg = MagicMock() + mock_msg.parts = [system_part] + mock_msg.instructions = None + + msgs = [mock_msg] + + # Should handle system prompt + _set_input_messages(span, msgs) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_message_with_instructions(sentry_init, capture_events): + """ + Test that messages with instructions field are handled correctly. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_input_messages + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create message with instructions + mock_msg = MagicMock() + mock_msg.instructions = "System instructions here" + mock_part = MagicMock() + mock_part.content = "User message" + mock_msg.parts = [mock_part] + + msgs = [mock_msg] + + # Should extract system prompt from instructions + _set_input_messages(span, msgs) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_input_messages_without_prompts(sentry_init, capture_events): + """ + Test that _set_input_messages respects _should_send_prompts(). + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_input_messages + + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=False)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Even with messages, should not set them + messages = ["test"] + _set_input_messages(span, messages) + + span.finish() + + # Should not crash and should not set messages + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_output_data_without_prompts(sentry_init, capture_events): + """ + Test that _set_output_data respects _should_send_prompts(). + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import _set_output_data + + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=False)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Even with response, should not set output data + mock_response = MagicMock() + mock_response.model_name = "test" + _set_output_data(span, mock_response) + + span.finish() + + # Should not crash and should not set output + assert transaction is not None + + +@pytest.mark.asyncio +async def test_get_model_name_with_exception_in_callable(sentry_init, capture_events): + """ + Test that _get_model_name handles exceptions in name() callable. + """ + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _get_model_name + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Create model with callable name that raises exception + mock_model = MagicMock() + mock_model.name = MagicMock(side_effect=Exception("Error")) + + # Should fall back to str() + result = _get_model_name(mock_model) + + # Should return something (str fallback) + assert result is not None + + +@pytest.mark.asyncio +async def test_get_model_name_with_string_model(sentry_init, capture_events): + """ + Test that _get_model_name handles string models. + """ + from sentry_sdk.integrations.pydantic_ai.utils import _get_model_name + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Pass a string as model + result = _get_model_name("gpt-4") + + # Should return the string + assert result == "gpt-4" + + +@pytest.mark.asyncio +async def test_get_model_name_with_none(sentry_init, capture_events): + """ + Test that _get_model_name handles None model. + """ + from sentry_sdk.integrations.pydantic_ai.utils import _get_model_name + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Pass None + result = _get_model_name(None) + + # Should return None + assert result is None + + +@pytest.mark.asyncio +async def test_set_model_data_with_system(sentry_init, capture_events): + """ + Test that _set_model_data captures system from model. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_model_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create model with system + mock_model = MagicMock() + mock_model.system = "openai" + mock_model.model_name = "gpt-4" + + # Set model data + _set_model_data(span, mock_model, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_model_data_from_agent_scope(sentry_init, capture_events): + """ + Test that _set_model_data retrieves model from agent in scope when not passed. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_model_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Set agent in scope + scope = sentry_sdk.get_current_scope() + mock_agent = MagicMock() + mock_agent.model = MagicMock() + mock_agent.model.model_name = "test-model" + mock_agent.model_settings = {"temperature": 0.5} + scope._contexts["pydantic_ai_agent"] = {"_agent": mock_agent} + + span = sentry_sdk.start_span(op="test_span") + + # Pass None for model, should get from scope + _set_model_data(span, None, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_model_data_with_none_settings_values(sentry_init, capture_events): + """ + Test that _set_model_data skips None values in settings. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.utils import _set_model_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create settings with None values + settings = { + "temperature": 0.7, + "max_tokens": None, # Should be skipped + "top_p": None, # Should be skipped + } + + # Set model data + _set_model_data(span, None, settings) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_should_send_prompts_with_no_integration(sentry_init, capture_events): + """ + Test that _should_send_prompts returns False when integration not found. + """ + from sentry_sdk.integrations.pydantic_ai.utils import _should_send_prompts + + # Initialize without PydanticAIIntegration + sentry_init( + integrations=[], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + # Should return False + result = _should_send_prompts() + assert result is False + + +@pytest.mark.asyncio +async def test_should_send_prompts_without_pii(sentry_init, capture_events): + """ + Test that _should_send_prompts returns False when PII disabled. + """ + from sentry_sdk.integrations.pydantic_ai.utils import _should_send_prompts + + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=False, # PII disabled + ) + + # Should return False + result = _should_send_prompts() + assert result is False + + +@pytest.mark.asyncio +async def test_set_agent_data_without_agent(sentry_init, capture_events): + """ + Test that _set_agent_data handles None agent gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.utils import _set_agent_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Pass None agent, with no agent in scope + _set_agent_data(span, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_agent_data_from_scope(sentry_init, capture_events): + """ + Test that _set_agent_data retrieves agent from scope when not passed. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_agent_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Set agent in scope + scope = sentry_sdk.get_current_scope() + mock_agent = MagicMock() + mock_agent.name = "test_agent_from_scope" + scope._contexts["pydantic_ai_agent"] = {"_agent": mock_agent} + + span = sentry_sdk.start_span(op="test_span") + + # Pass None for agent, should get from scope + _set_agent_data(span, None) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_agent_data_without_name(sentry_init, capture_events): + """ + Test that _set_agent_data handles agent without name attribute. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_agent_data + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create agent without name + mock_agent = MagicMock() + mock_agent.name = None # No name + + # Should not set agent name + _set_agent_data(span, mock_agent) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_available_tools_without_toolset(sentry_init, capture_events): + """ + Test that _set_available_tools handles agent without toolset. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_available_tools + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create agent without _function_toolset + mock_agent = MagicMock() + del mock_agent._function_toolset + + # Should handle gracefully + _set_available_tools(span, mock_agent) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_set_available_tools_with_schema(sentry_init, capture_events): + """ + Test that _set_available_tools extracts tool schema correctly. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.utils import _set_available_tools + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + span = sentry_sdk.start_span(op="test_span") + + # Create agent with toolset containing schema + mock_agent = MagicMock() + mock_tool = MagicMock() + mock_schema = MagicMock() + mock_schema.description = "Test tool description" + mock_schema.json_schema = {"type": "object", "properties": {}} + mock_tool.function_schema = mock_schema + + mock_agent._function_toolset.tools = {"test_tool": mock_tool} + + # Should extract schema + _set_available_tools(span, mock_agent) + + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_execute_tool_span_creation(sentry_init, capture_events): + """ + Test direct creation of execute_tool span. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import ( + execute_tool_span, + update_execute_tool_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create execute_tool span + with execute_tool_span("test_tool", {"arg": "value"}, None, "function") as span: + # Update with result + update_execute_tool_span(span, {"result": "success"}) + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_execute_tool_span_with_mcp_type(sentry_init, capture_events): + """ + Test execute_tool span with MCP tool type. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import execute_tool_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create execute_tool span with mcp type + with execute_tool_span("mcp_tool", {"arg": "value"}, None, "mcp") as span: + # Verify type is set + assert span is not None + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_execute_tool_span_without_prompts(sentry_init, capture_events): + """ + Test that execute_tool span respects _should_send_prompts(). + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import ( + execute_tool_span, + update_execute_tool_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration(include_prompts=False)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create execute_tool span + with execute_tool_span("test_tool", {"arg": "value"}, None, "function") as span: + # Update with result - should not set input/output + update_execute_tool_span(span, {"result": "success"}) + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_execute_tool_span_with_none_args(sentry_init, capture_events): + """ + Test execute_tool span with None args. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import execute_tool_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create execute_tool span with None args + with execute_tool_span("test_tool", None, None, "function") as span: + assert span is not None + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_update_execute_tool_span_with_none_span(sentry_init, capture_events): + """ + Test that update_execute_tool_span handles None span gracefully. + """ + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import ( + update_execute_tool_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Update with None span - should not raise + update_execute_tool_span(None, {"result": "success"}) + + # Should not crash + assert True + + +@pytest.mark.asyncio +async def test_update_execute_tool_span_with_none_result(sentry_init, capture_events): + """ + Test that update_execute_tool_span handles None result gracefully. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.execute_tool import ( + execute_tool_span, + update_execute_tool_span, + ) + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create execute_tool span + with execute_tool_span("test_tool", {"arg": "value"}, None, "function") as span: + # Update with None result + update_execute_tool_span(span, None) + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_tool_execution_without_span_context(sentry_init, capture_events): + """ + Test that tool execution patch handles case when no span context exists. + This tests the code path where current_span is None in _patch_tool_execution. + """ + # Import the patching function + from unittest.mock import AsyncMock, MagicMock + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + # Create a simple agent with no tools (won't have function_toolset) + agent = Agent("test", name="test_no_span") + + # Call without span context (no transaction active) + # The patches should handle this gracefully + try: + # This will fail because we're not in a transaction, but it should not crash + await agent.run("test") + except Exception: + # Expected to fail, that's okay + pass + + # Should not crash + assert True + + +@pytest.mark.asyncio +async def test_invoke_agent_span_with_callable_instruction(sentry_init, capture_events): + """ + Test that invoke_agent_span skips callable instructions correctly. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.invoke_agent import invoke_agent_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create mock agent with callable instruction + mock_agent = MagicMock() + mock_agent.name = "test_agent" + mock_agent._system_prompts = [] + + # Add both string and callable instructions + mock_callable = lambda: "Dynamic instruction" + mock_agent._instructions = ["Static instruction", mock_callable] + + # Create span + span = invoke_agent_span("Test prompt", mock_agent, None, None) + span.finish() + + # Should not crash (callable should be skipped) + assert transaction is not None + + +@pytest.mark.asyncio +async def test_invoke_agent_span_with_string_instructions(sentry_init, capture_events): + """ + Test that invoke_agent_span handles string instructions (not list). + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.invoke_agent import invoke_agent_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Create mock agent with string instruction + mock_agent = MagicMock() + mock_agent.name = "test_agent" + mock_agent._system_prompts = [] + mock_agent._instructions = "Single instruction string" + + # Create span + span = invoke_agent_span("Test prompt", mock_agent, None, None) + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_ai_client_span_with_streaming_flag(sentry_init, capture_events): + """ + Test that ai_client_span reads streaming flag from scope. + """ + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import ai_client_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Set streaming flag in scope + scope = sentry_sdk.get_current_scope() + scope._contexts["pydantic_ai_agent"] = {"_streaming": True} + + # Create ai_client span + span = ai_client_span([], None, None, None) + span.finish() + + # Should not crash + assert transaction is not None + + +@pytest.mark.asyncio +async def test_ai_client_span_gets_agent_from_scope(sentry_init, capture_events): + """ + Test that ai_client_span gets agent from scope when not passed. + """ + import sentry_sdk + from unittest.mock import MagicMock + from sentry_sdk.integrations.pydantic_ai.spans.ai_client import ai_client_span + + sentry_init( + integrations=[PydanticAIIntegration()], + traces_sample_rate=1.0, + ) + + with sentry_sdk.start_transaction(op="test", name="test") as transaction: + # Set agent in scope + scope = sentry_sdk.get_current_scope() + mock_agent = MagicMock() + mock_agent.name = "test_agent" + mock_agent._function_toolset = MagicMock() + mock_agent._function_toolset.tools = {} + scope._contexts["pydantic_ai_agent"] = {"_agent": mock_agent} + + # Create ai_client span without passing agent + span = ai_client_span([], None, None, None) + span.finish() + + # Should not crash + assert transaction is not None diff --git a/tox.ini b/tox.ini index 05273d9e82..0e9d3a7bb7 100644 --- a/tox.ini +++ b/tox.ini @@ -95,6 +95,8 @@ envlist = {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 {py3.10,py3.12,py3.13}-openai_agents-v0.4.1 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.17 + # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 @@ -412,6 +414,9 @@ deps = openai_agents-v0.4.1: openai-agents==0.4.1 openai_agents: pytest-asyncio + pydantic_ai-v1.0.17: pydantic-ai==1.0.17 + pydantic_ai: pytest-asyncio + # ~~~ Cloud ~~~ boto3-v1.12.49: boto3==1.12.49 @@ -787,6 +792,7 @@ setenv = openai_agents: TESTPATH=tests/integrations/openai_agents openfeature: TESTPATH=tests/integrations/openfeature pure_eval: TESTPATH=tests/integrations/pure_eval + pydantic_ai: TESTPATH=tests/integrations/pydantic_ai pymongo: TESTPATH=tests/integrations/pymongo pyramid: TESTPATH=tests/integrations/pyramid quart: TESTPATH=tests/integrations/quart From 294493661b50507cae577a3dd71963546f2fd13e Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Thu, 23 Oct 2025 16:26:37 +0200 Subject: [PATCH 570/591] feat(integrations): MCP Python SDK (#4964) ### Description Adds integration for the [MCP Python SDK](https://github.com/modelcontextprotocol/python-sdk) #### Issues * Closes https://linear.app/getsentry/issue/TET-1267/mcp-python-sdk-integration * Closes https://linear.app/getsentry/issue/TET-1301/python-mcp-add-tests --------- Co-authored-by: Ivana Kellyer --- .github/workflows/test-integrations-ai.yml | 4 + scripts/populate_tox/config.py | 6 + scripts/populate_tox/releases.jsonl | 6 +- .../split_tox_gh_actions.py | 1 + sentry_sdk/consts.py | 85 ++ sentry_sdk/integrations/__init__.py | 1 + sentry_sdk/integrations/mcp.py | 552 +++++++++++ setup.py | 1 + tests/integrations/mcp/__init__.py | 3 + tests/integrations/mcp/test_mcp.py | 911 ++++++++++++++++++ tox.ini | 16 +- 11 files changed, 1583 insertions(+), 3 deletions(-) create mode 100644 sentry_sdk/integrations/mcp.py create mode 100644 tests/integrations/mcp/__init__.py create mode 100644 tests/integrations/mcp/test_mcp.py diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index ca1a10aa5c..d09213ebd6 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -82,6 +82,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-litellm" + - name: Test mcp + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-mcp" - name: Test openai-base run: | set -x # print commands that are executed diff --git a/scripts/populate_tox/config.py b/scripts/populate_tox/config.py index e48a1cba88..69263f92a3 100644 --- a/scripts/populate_tox/config.py +++ b/scripts/populate_tox/config.py @@ -242,6 +242,12 @@ "package": "loguru", "num_versions": 2, }, + "mcp": { + "package": "mcp", + "deps": { + "*": ["pytest-asyncio"], + }, + }, "openai-base": { "package": "openai", "integration_name": "openai", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 03fd954b65..ae75704553 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -98,7 +98,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.35.3", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.36.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc7", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} @@ -114,6 +114,10 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.15.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.16.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.17.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.18.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 6f7015f07c..9f9369d4ab 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -78,6 +78,7 @@ "langchain-notiktoken", "langgraph", "litellm", + "mcp", "openai-base", "openai-notiktoken", "openai_agents", diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index 4b7d219245..b041e0626b 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -749,6 +749,90 @@ class SPANDATA: Example: "MainThread" """ + MCP_TOOL_NAME = "mcp.tool.name" + """ + The name of the MCP tool being called. + Example: "get_weather" + """ + + MCP_PROMPT_NAME = "mcp.prompt.name" + """ + The name of the MCP prompt being retrieved. + Example: "code_review" + """ + + MCP_RESOURCE_URI = "mcp.resource.uri" + """ + The URI of the MCP resource being accessed. + Example: "file:///path/to/resource" + """ + + MCP_METHOD_NAME = "mcp.method.name" + """ + The MCP protocol method name being called. + Example: "tools/call", "prompts/get", "resources/read" + """ + + MCP_REQUEST_ID = "mcp.request.id" + """ + The unique identifier for the MCP request. + Example: "req_123abc" + """ + + MCP_TOOL_RESULT_CONTENT = "mcp.tool.result.content" + """ + The result/output content from an MCP tool execution. + Example: "The weather is sunny" + """ + + MCP_TOOL_RESULT_CONTENT_COUNT = "mcp.tool.result.content_count" + """ + The number of items/keys in the MCP tool result. + Example: 5 + """ + + MCP_TOOL_RESULT_IS_ERROR = "mcp.tool.result.is_error" + """ + Whether the MCP tool execution resulted in an error. + Example: True + """ + + MCP_PROMPT_RESULT_MESSAGE_CONTENT = "mcp.prompt.result.message_content" + """ + The message content from an MCP prompt retrieval. + Example: "Review the following code..." + """ + + MCP_PROMPT_RESULT_MESSAGE_ROLE = "mcp.prompt.result.message_role" + """ + The role of the message in an MCP prompt retrieval (only set for single-message prompts). + Example: "user", "assistant", "system" + """ + + MCP_PROMPT_RESULT_MESSAGE_COUNT = "mcp.prompt.result.message_count" + """ + The number of messages in an MCP prompt result. + Example: 1, 3 + """ + + MCP_RESOURCE_PROTOCOL = "mcp.resource.protocol" + """ + The protocol/scheme of the MCP resource URI. + Example: "file", "http", "https" + """ + + MCP_TRANSPORT = "mcp.transport" + """ + The transport method used for MCP communication. + Example: "pipe" (stdio), "tcp" (HTTP/WebSocket/SSE) + """ + + MCP_SESSION_ID = "mcp.session.id" + """ + The session identifier for the MCP connection. + Example: "a1b2c3d4e5f6" + """ + class SPANSTATUS: """ @@ -845,6 +929,7 @@ class OP: WEBSOCKET_SERVER = "websocket.server" SOCKET_CONNECTION = "socket.connection" SOCKET_DNS = "socket.dns" + MCP_SERVER = "mcp.server" # This type exists to trick mypy and PyCharm into thinking `init` and `Client` diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 45a0f53eaf..2d3d621ef7 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -149,6 +149,7 @@ def iter_default_integrations(with_auto_enabling_integrations): "launchdarkly": (9, 8, 0), "litellm": (1, 77, 5), "loguru": (0, 7, 0), + "mcp": (1, 15, 0), "openai": (1, 0, 0), "openai_agents": (0, 0, 19), "openfeature": (0, 7, 1), diff --git a/sentry_sdk/integrations/mcp.py b/sentry_sdk/integrations/mcp.py new file mode 100644 index 0000000000..2a2d440616 --- /dev/null +++ b/sentry_sdk/integrations/mcp.py @@ -0,0 +1,552 @@ +""" +Sentry integration for MCP (Model Context Protocol) servers. + +This integration instruments MCP servers to create spans for tool, prompt, +and resource handler execution, and captures errors that occur during execution. + +Supports the low-level `mcp.server.lowlevel.Server` API. +""" + +import inspect +from functools import wraps +from typing import TYPE_CHECKING + +import sentry_sdk +from sentry_sdk.ai.utils import get_start_span_function +from sentry_sdk.consts import OP, SPANDATA +from sentry_sdk.integrations import Integration, DidNotEnable +from sentry_sdk.utils import safe_serialize +from sentry_sdk.scope import should_send_default_pii + +try: + from mcp.server.lowlevel import Server # type: ignore[import-not-found] + from mcp.server.lowlevel.server import request_ctx # type: ignore[import-not-found] +except ImportError: + raise DidNotEnable("MCP SDK not installed") + + +if TYPE_CHECKING: + from typing import Any, Callable, Optional + + +class MCPIntegration(Integration): + identifier = "mcp" + origin = "auto.ai.mcp" + + def __init__(self, include_prompts=True): + # type: (bool) -> None + """ + Initialize the MCP integration. + + Args: + include_prompts: Whether to include prompts (tool results and prompt content) + in span data. Requires send_default_pii=True. Default is True. + """ + self.include_prompts = include_prompts + + @staticmethod + def setup_once(): + # type: () -> None + """ + Patches MCP server classes to instrument handler execution. + """ + _patch_lowlevel_server() + + +def _get_request_context_data(): + # type: () -> tuple[Optional[str], Optional[str], str] + """ + Extract request ID, session ID, and transport type from the MCP request context. + + Returns: + Tuple of (request_id, session_id, transport). + - request_id: May be None if not available + - session_id: May be None if not available + - transport: "tcp" for HTTP-based, "pipe" for stdio + """ + request_id = None # type: Optional[str] + session_id = None # type: Optional[str] + transport = "pipe" # type: str + + try: + ctx = request_ctx.get() + + if ctx is not None: + request_id = ctx.request_id + if hasattr(ctx, "request") and ctx.request is not None: + transport = "tcp" + request = ctx.request + if hasattr(request, "headers"): + session_id = request.headers.get("mcp-session-id") + + except LookupError: + # No request context available - default to pipe + pass + + return request_id, session_id, transport + + +def _get_span_config(handler_type, item_name): + # type: (str, str) -> tuple[str, str, str, Optional[str]] + """ + Get span configuration based on handler type. + + Returns: + Tuple of (span_data_key, span_name, mcp_method_name, result_data_key) + Note: result_data_key is None for resources + """ + if handler_type == "tool": + span_data_key = SPANDATA.MCP_TOOL_NAME + mcp_method_name = "tools/call" + result_data_key = SPANDATA.MCP_TOOL_RESULT_CONTENT + elif handler_type == "prompt": + span_data_key = SPANDATA.MCP_PROMPT_NAME + mcp_method_name = "prompts/get" + result_data_key = SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT + else: # resource + span_data_key = SPANDATA.MCP_RESOURCE_URI + mcp_method_name = "resources/read" + result_data_key = None # Resources don't capture result content + + span_name = f"{mcp_method_name} {item_name}" + return span_data_key, span_name, mcp_method_name, result_data_key + + +def _set_span_input_data( + span, + handler_name, + span_data_key, + mcp_method_name, + arguments, + request_id, + session_id, + transport, +): + # type: (Any, str, str, str, dict[str, Any], Optional[str], Optional[str], str) -> None + """Set input span data for MCP handlers.""" + # Set handler identifier + span.set_data(span_data_key, handler_name) + span.set_data(SPANDATA.MCP_METHOD_NAME, mcp_method_name) + + # Set transport type + span.set_data(SPANDATA.MCP_TRANSPORT, transport) + + # Set request_id if provided + if request_id: + span.set_data(SPANDATA.MCP_REQUEST_ID, request_id) + + # Set session_id if provided + if session_id: + span.set_data(SPANDATA.MCP_SESSION_ID, session_id) + + # Set request arguments (excluding common request context objects) + for k, v in arguments.items(): + span.set_data(f"mcp.request.argument.{k}", safe_serialize(v)) + + +def _extract_tool_result_content(result): + # type: (Any) -> Any + """ + Extract meaningful content from MCP tool result. + + Tool handlers can return: + - tuple (UnstructuredContent, StructuredContent): Return the structured content (dict) + - dict (StructuredContent): Return as-is + - Iterable (UnstructuredContent): Extract text from content blocks + """ + if result is None: + return None + + # Handle CombinationContent: tuple of (UnstructuredContent, StructuredContent) + if isinstance(result, tuple) and len(result) == 2: + # Return the structured content (2nd element) + return result[1] + + # Handle StructuredContent: dict + if isinstance(result, dict): + return result + + # Handle UnstructuredContent: iterable of ContentBlock objects + # Try to extract text content + if hasattr(result, "__iter__") and not isinstance(result, (str, bytes, dict)): + texts = [] + try: + for item in result: + # Try to get text attribute from ContentBlock objects + if hasattr(item, "text"): + texts.append(item.text) + elif isinstance(item, dict) and "text" in item: + texts.append(item["text"]) + except Exception: + # If extraction fails, return the original + return result + return " ".join(texts) if texts else result + + return result + + +def _set_span_output_data(span, result, result_data_key, handler_type): + # type: (Any, Any, Optional[str], str) -> None + """Set output span data for MCP handlers.""" + if result is None: + return + + # Get integration to check PII settings + integration = sentry_sdk.get_client().get_integration(MCPIntegration) + if integration is None: + return + + # Check if we should include sensitive data + should_include_data = should_send_default_pii() and integration.include_prompts + + # For tools, extract the meaningful content + if handler_type == "tool": + extracted = _extract_tool_result_content(result) + if extracted is not None and should_include_data: + span.set_data(result_data_key, safe_serialize(extracted)) + # Set content count if result is a dict + if isinstance(extracted, dict): + span.set_data(SPANDATA.MCP_TOOL_RESULT_CONTENT_COUNT, len(extracted)) + elif handler_type == "prompt": + # For prompts, count messages and set role/content only for single-message prompts + try: + messages = None # type: Optional[list[str]] + message_count = 0 + + # Check if result has messages attribute (GetPromptResult) + if hasattr(result, "messages") and result.messages: + messages = result.messages + message_count = len(messages) + # Also check if result is a dict with messages + elif isinstance(result, dict) and result.get("messages"): + messages = result["messages"] + message_count = len(messages) + + # Always set message count if we found messages + if message_count > 0: + span.set_data(SPANDATA.MCP_PROMPT_RESULT_MESSAGE_COUNT, message_count) + + # Only set role and content for single-message prompts if PII is allowed + if message_count == 1 and should_include_data and messages: + first_message = messages[0] + # Extract role + role = None + if hasattr(first_message, "role"): + role = first_message.role + elif isinstance(first_message, dict) and "role" in first_message: + role = first_message["role"] + + if role: + span.set_data(SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE, role) + + # Extract content text + content_text = None + if hasattr(first_message, "content"): + msg_content = first_message.content + # Content can be a TextContent object or similar + if hasattr(msg_content, "text"): + content_text = msg_content.text + elif isinstance(msg_content, dict) and "text" in msg_content: + content_text = msg_content["text"] + elif isinstance(msg_content, str): + content_text = msg_content + elif isinstance(first_message, dict) and "content" in first_message: + msg_content = first_message["content"] + if isinstance(msg_content, dict) and "text" in msg_content: + content_text = msg_content["text"] + elif isinstance(msg_content, str): + content_text = msg_content + + if content_text: + span.set_data(result_data_key, content_text) + except Exception: + # Silently ignore if we can't extract message info + pass + # Resources don't capture result content (result_data_key is None) + + +# Handler data preparation and wrapping + + +def _prepare_handler_data(handler_type, original_args): + # type: (str, tuple[Any, ...]) -> tuple[str, dict[str, Any], str, str, str, Optional[str]] + """ + Prepare common handler data for both async and sync wrappers. + + Returns: + Tuple of (handler_name, arguments, span_data_key, span_name, mcp_method_name, result_data_key) + """ + # Extract handler-specific data based on handler type + if handler_type == "tool": + handler_name = original_args[0] # tool_name + arguments = original_args[1] if len(original_args) > 1 else {} + elif handler_type == "prompt": + handler_name = original_args[0] # name + arguments = original_args[1] if len(original_args) > 1 else {} + # Include name in arguments dict for span data + arguments = {"name": handler_name, **(arguments or {})} + else: # resource + uri = original_args[0] + handler_name = str(uri) if uri else "unknown" + arguments = {} + + # Get span configuration + span_data_key, span_name, mcp_method_name, result_data_key = _get_span_config( + handler_type, handler_name + ) + + return ( + handler_name, + arguments, + span_data_key, + span_name, + mcp_method_name, + result_data_key, + ) + + +async def _async_handler_wrapper(handler_type, func, original_args): + # type: (str, Callable[..., Any], tuple[Any, ...]) -> Any + """ + Async wrapper for MCP handlers. + + Args: + handler_type: "tool", "prompt", or "resource" + func: The async handler function to wrap + original_args: Original arguments passed to the handler + """ + ( + handler_name, + arguments, + span_data_key, + span_name, + mcp_method_name, + result_data_key, + ) = _prepare_handler_data(handler_type, original_args) + + # Start span and execute + with get_start_span_function()( + op=OP.MCP_SERVER, + name=span_name, + origin=MCPIntegration.origin, + ) as span: + # Get request ID, session ID, and transport from context + request_id, session_id, transport = _get_request_context_data() + + # Set input span data + _set_span_input_data( + span, + handler_name, + span_data_key, + mcp_method_name, + arguments, + request_id, + session_id, + transport, + ) + + # For resources, extract and set protocol + if handler_type == "resource": + uri = original_args[0] + protocol = None + if hasattr(uri, "scheme"): + protocol = uri.scheme + elif handler_name and "://" in handler_name: + protocol = handler_name.split("://")[0] + if protocol: + span.set_data(SPANDATA.MCP_RESOURCE_PROTOCOL, protocol) + + try: + # Execute the async handler + result = await func(*original_args) + except Exception as e: + # Set error flag for tools + if handler_type == "tool": + span.set_data(SPANDATA.MCP_TOOL_RESULT_IS_ERROR, True) + sentry_sdk.capture_exception(e) + raise + + _set_span_output_data(span, result, result_data_key, handler_type) + return result + + +def _sync_handler_wrapper(handler_type, func, original_args): + # type: (str, Callable[..., Any], tuple[Any, ...]) -> Any + """ + Sync wrapper for MCP handlers. + + Args: + handler_type: "tool", "prompt", or "resource" + func: The sync handler function to wrap + original_args: Original arguments passed to the handler + """ + ( + handler_name, + arguments, + span_data_key, + span_name, + mcp_method_name, + result_data_key, + ) = _prepare_handler_data(handler_type, original_args) + + # Start span and execute + with get_start_span_function()( + op=OP.MCP_SERVER, + name=span_name, + origin=MCPIntegration.origin, + ) as span: + # Get request ID, session ID, and transport from context + request_id, session_id, transport = _get_request_context_data() + + # Set input span data + _set_span_input_data( + span, + handler_name, + span_data_key, + mcp_method_name, + arguments, + request_id, + session_id, + transport, + ) + + # For resources, extract and set protocol + if handler_type == "resource": + uri = original_args[0] + protocol = None + if hasattr(uri, "scheme"): + protocol = uri.scheme + elif handler_name and "://" in handler_name: + protocol = handler_name.split("://")[0] + if protocol: + span.set_data(SPANDATA.MCP_RESOURCE_PROTOCOL, protocol) + + try: + # Execute the sync handler + result = func(*original_args) + except Exception as e: + # Set error flag for tools + if handler_type == "tool": + span.set_data(SPANDATA.MCP_TOOL_RESULT_IS_ERROR, True) + sentry_sdk.capture_exception(e) + raise + + _set_span_output_data(span, result, result_data_key, handler_type) + return result + + +def _create_instrumented_handler(handler_type, func): + # type: (str, Callable[..., Any]) -> Callable[..., Any] + """ + Create an instrumented version of a handler function (async or sync). + + This function wraps the user's handler with a runtime wrapper that will create + Sentry spans and capture metrics when the handler is actually called. + + The wrapper preserves the async/sync nature of the original function, which is + critical for Python's async/await to work correctly. + + Args: + handler_type: "tool", "prompt", or "resource" - determines span configuration + func: The handler function to instrument (async or sync) + + Returns: + A wrapped version of func that creates Sentry spans on execution + """ + if inspect.iscoroutinefunction(func): + + @wraps(func) + async def async_wrapper(*args): + # type: (*Any) -> Any + return await _async_handler_wrapper(handler_type, func, args) + + return async_wrapper + else: + + @wraps(func) + def sync_wrapper(*args): + # type: (*Any) -> Any + return _sync_handler_wrapper(handler_type, func, args) + + return sync_wrapper + + +def _create_instrumented_decorator( + original_decorator, handler_type, *decorator_args, **decorator_kwargs +): + # type: (Callable[..., Any], str, *Any, **Any) -> Callable[..., Any] + """ + Create an instrumented version of an MCP decorator. + + This function intercepts MCP decorators (like @server.call_tool()) and injects + Sentry instrumentation into the handler registration flow. The returned decorator + will: + 1. Receive the user's handler function + 2. Wrap it with instrumentation via _create_instrumented_handler + 3. Pass the instrumented version to the original MCP decorator + + This ensures that when the handler is called at runtime, it's already wrapped + with Sentry spans and metrics collection. + + Args: + original_decorator: The original MCP decorator method (e.g., Server.call_tool) + handler_type: "tool", "prompt", or "resource" - determines span configuration + decorator_args: Positional arguments to pass to the original decorator (e.g., self) + decorator_kwargs: Keyword arguments to pass to the original decorator + + Returns: + A decorator function that instruments handlers before registering them + """ + + def instrumented_decorator(func): + # type: (Callable[..., Any]) -> Callable[..., Any] + # First wrap the handler with instrumentation + instrumented_func = _create_instrumented_handler(handler_type, func) + # Then register it with the original MCP decorator + return original_decorator(*decorator_args, **decorator_kwargs)( + instrumented_func + ) + + return instrumented_decorator + + +def _patch_lowlevel_server(): + # type: () -> None + """ + Patches the mcp.server.lowlevel.Server class to instrument handler execution. + """ + # Patch call_tool decorator + original_call_tool = Server.call_tool + + def patched_call_tool(self, **kwargs): + # type: (Server, **Any) -> Callable[[Callable[..., Any]], Callable[..., Any]] + """Patched version of Server.call_tool that adds Sentry instrumentation.""" + return lambda func: _create_instrumented_decorator( + original_call_tool, "tool", self, **kwargs + )(func) + + Server.call_tool = patched_call_tool + + # Patch get_prompt decorator + original_get_prompt = Server.get_prompt + + def patched_get_prompt(self): + # type: (Server) -> Callable[[Callable[..., Any]], Callable[..., Any]] + """Patched version of Server.get_prompt that adds Sentry instrumentation.""" + return lambda func: _create_instrumented_decorator( + original_get_prompt, "prompt", self + )(func) + + Server.get_prompt = patched_get_prompt + + # Patch read_resource decorator + original_read_resource = Server.read_resource + + def patched_read_resource(self): + # type: (Server) -> Callable[[Callable[..., Any]], Callable[..., Any]] + """Patched version of Server.read_resource that adds Sentry instrumentation.""" + return lambda func: _create_instrumented_decorator( + original_read_resource, "resource", self + )(func) + + Server.read_resource = patched_read_resource diff --git a/setup.py b/setup.py index b1662204fb..377d6c82d6 100644 --- a/setup.py +++ b/setup.py @@ -68,6 +68,7 @@ def get_file_text(file_name): "litellm": ["litellm>=1.77.5"], "litestar": ["litestar>=2.0.0"], "loguru": ["loguru>=0.5"], + "mcp": ["mcp>=1.15.0"], "openai": ["openai>=1.0.0", "tiktoken>=0.3.0"], "openfeature": ["openfeature-sdk>=0.7.1"], "opentelemetry": ["opentelemetry-distro>=0.35b0"], diff --git a/tests/integrations/mcp/__init__.py b/tests/integrations/mcp/__init__.py new file mode 100644 index 0000000000..01ef442500 --- /dev/null +++ b/tests/integrations/mcp/__init__.py @@ -0,0 +1,3 @@ +import pytest + +pytest.importorskip("mcp") diff --git a/tests/integrations/mcp/test_mcp.py b/tests/integrations/mcp/test_mcp.py new file mode 100644 index 0000000000..738fdedf48 --- /dev/null +++ b/tests/integrations/mcp/test_mcp.py @@ -0,0 +1,911 @@ +""" +Unit tests for the MCP (Model Context Protocol) integration. + +This test suite covers: +- Tool handlers (sync and async) +- Prompt handlers (sync and async) +- Resource handlers (sync and async) +- Error handling for each handler type +- Request context data extraction (request_id, session_id, transport) +- Tool result content extraction (various formats) +- Span data validation +- Origin tracking + +The tests mock the MCP server components and request context to verify +that the integration properly instruments MCP handlers with Sentry spans. +""" + +import pytest +import json +from unittest import mock + +try: + from unittest.mock import AsyncMock +except ImportError: + + class AsyncMock(mock.MagicMock): + async def __call__(self, *args, **kwargs): + return super(AsyncMock, self).__call__(*args, **kwargs) + + +from mcp.server.lowlevel import Server +from mcp.server.lowlevel.server import request_ctx + +try: + from mcp.server.lowlevel.server import request_ctx +except ImportError: + request_ctx = None + +from sentry_sdk import start_transaction +from sentry_sdk.consts import SPANDATA, OP +from sentry_sdk.integrations.mcp import MCPIntegration + + +@pytest.fixture(autouse=True) +def reset_request_ctx(): + """Reset request context before and after each test""" + if request_ctx is not None: + try: + if request_ctx.get() is not None: + request_ctx.set(None) + except LookupError: + pass + + yield + + if request_ctx is not None: + try: + request_ctx.set(None) + except LookupError: + pass + + +# Mock MCP types and structures +class MockURI: + """Mock URI object for resource testing""" + + def __init__(self, uri_string): + self.scheme = uri_string.split("://")[0] if "://" in uri_string else "" + self.path = uri_string.split("://")[1] if "://" in uri_string else uri_string + self._uri_string = uri_string + + def __str__(self): + return self._uri_string + + +class MockRequestContext: + """Mock MCP request context""" + + def __init__(self, request_id=None, session_id=None, transport="pipe"): + self.request_id = request_id + if transport == "tcp": + self.request = MockHTTPRequest(session_id) + else: + self.request = None + + +class MockHTTPRequest: + """Mock HTTP request for SSE/WebSocket transport""" + + def __init__(self, session_id=None): + self.headers = {} + if session_id: + self.headers["mcp-session-id"] = session_id + + +class MockTextContent: + """Mock TextContent object""" + + def __init__(self, text): + self.text = text + + +class MockPromptMessage: + """Mock PromptMessage object""" + + def __init__(self, role, content_text): + self.role = role + self.content = MockTextContent(content_text) + + +class MockGetPromptResult: + """Mock GetPromptResult object""" + + def __init__(self, messages): + self.messages = messages + + +def test_integration_patches_server(sentry_init): + """Test that MCPIntegration patches the Server class""" + # Get original methods before integration + original_call_tool = Server.call_tool + original_get_prompt = Server.get_prompt + original_read_resource = Server.read_resource + + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + + # After initialization, the methods should be patched + assert Server.call_tool is not original_call_tool + assert Server.get_prompt is not original_get_prompt + assert Server.read_resource is not original_read_resource + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +def test_tool_handler_sync( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test that synchronous tool handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-123", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def test_tool(tool_name, arguments): + return {"result": "success", "value": 42} + + with start_transaction(name="mcp tx"): + # Call the tool handler + result = test_tool("calculate", {"x": 10, "y": 5}) + + assert result == {"result": "success", "value": 42} + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "tools/call calculate" + assert span["origin"] == "auto.ai.mcp" + + # Check span data + assert span["data"][SPANDATA.MCP_TOOL_NAME] == "calculate" + assert span["data"][SPANDATA.MCP_METHOD_NAME] == "tools/call" + assert span["data"][SPANDATA.MCP_TRANSPORT] == "pipe" + assert span["data"][SPANDATA.MCP_REQUEST_ID] == "req-123" + assert span["data"]["mcp.request.argument.x"] == "10" + assert span["data"]["mcp.request.argument.y"] == "5" + + # Check PII-sensitive data is only present when both flags are True + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT] == json.dumps( + { + "result": "success", + "value": 42, + } + ) + assert span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT_COUNT] == 2 + else: + assert SPANDATA.MCP_TOOL_RESULT_CONTENT not in span["data"] + assert SPANDATA.MCP_TOOL_RESULT_CONTENT_COUNT not in span["data"] + + +@pytest.mark.asyncio +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +async def test_tool_handler_async( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test that async tool handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext( + request_id="req-456", session_id="session-789", transport="tcp" + ) + request_ctx.set(mock_ctx) + + @server.call_tool() + async def test_tool_async(tool_name, arguments): + return {"status": "completed"} + + with start_transaction(name="mcp tx"): + result = await test_tool_async("process", {"data": "test"}) + + assert result == {"status": "completed"} + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "tools/call process" + assert span["origin"] == "auto.ai.mcp" + + # Check span data + assert span["data"][SPANDATA.MCP_TOOL_NAME] == "process" + assert span["data"][SPANDATA.MCP_METHOD_NAME] == "tools/call" + assert span["data"][SPANDATA.MCP_TRANSPORT] == "tcp" + assert span["data"][SPANDATA.MCP_REQUEST_ID] == "req-456" + assert span["data"][SPANDATA.MCP_SESSION_ID] == "session-789" + assert span["data"]["mcp.request.argument.data"] == '"test"' + + # Check PII-sensitive data + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT] == json.dumps( + {"status": "completed"} + ) + else: + assert SPANDATA.MCP_TOOL_RESULT_CONTENT not in span["data"] + + +def test_tool_handler_with_error(sentry_init, capture_events): + """Test that tool handler errors are captured properly""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-error", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def failing_tool(tool_name, arguments): + raise ValueError("Tool execution failed") + + with start_transaction(name="mcp tx"): + with pytest.raises(ValueError): + failing_tool("bad_tool", {}) + + # Should have error event and transaction + assert len(events) == 2 + error_event, tx = events + + # Check error event + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "ValueError" + assert error_event["exception"]["values"][0]["value"] == "Tool execution failed" + + # Check transaction and span + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + span = tx["spans"][0] + + # Error flag should be set for tools + assert span["data"][SPANDATA.MCP_TOOL_RESULT_IS_ERROR] is True + assert span["tags"]["status"] == "internal_error" + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +def test_prompt_handler_sync( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test that synchronous prompt handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-prompt", transport="pipe") + request_ctx.set(mock_ctx) + + @server.get_prompt() + def test_prompt(name, arguments): + return MockGetPromptResult([MockPromptMessage("user", "Tell me about Python")]) + + with start_transaction(name="mcp tx"): + result = test_prompt("code_help", {"language": "python"}) + + assert result.messages[0].role == "user" + assert result.messages[0].content.text == "Tell me about Python" + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "prompts/get code_help" + assert span["origin"] == "auto.ai.mcp" + + # Check span data + assert span["data"][SPANDATA.MCP_PROMPT_NAME] == "code_help" + assert span["data"][SPANDATA.MCP_METHOD_NAME] == "prompts/get" + assert span["data"][SPANDATA.MCP_TRANSPORT] == "pipe" + assert span["data"][SPANDATA.MCP_REQUEST_ID] == "req-prompt" + assert span["data"]["mcp.request.argument.name"] == '"code_help"' + assert span["data"]["mcp.request.argument.language"] == '"python"' + + # Message count is always captured + assert span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_COUNT] == 1 + + # For single message prompts, role and content should be captured only with PII + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE] == "user" + assert ( + span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT] + == "Tell me about Python" + ) + else: + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE not in span["data"] + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT not in span["data"] + + +@pytest.mark.asyncio +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (True, False), (False, True), (False, False)], +) +async def test_prompt_handler_async( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test that async prompt handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext( + request_id="req-async-prompt", session_id="session-abc", transport="tcp" + ) + request_ctx.set(mock_ctx) + + @server.get_prompt() + async def test_prompt_async(name, arguments): + return MockGetPromptResult( + [ + MockPromptMessage("system", "You are a helpful assistant"), + MockPromptMessage("user", "What is MCP?"), + ] + ) + + with start_transaction(name="mcp tx"): + result = await test_prompt_async("mcp_info", {}) + + assert len(result.messages) == 2 + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "prompts/get mcp_info" + + # For multi-message prompts, count is always captured + assert span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_COUNT] == 2 + # Role/content are never captured for multi-message prompts (even with PII) + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE not in span["data"] + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT not in span["data"] + + +def test_prompt_handler_with_error(sentry_init, capture_events): + """Test that prompt handler errors are captured""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-error-prompt", transport="pipe") + request_ctx.set(mock_ctx) + + @server.get_prompt() + def failing_prompt(name, arguments): + raise RuntimeError("Prompt not found") + + with start_transaction(name="mcp tx"): + with pytest.raises(RuntimeError): + failing_prompt("missing_prompt", {}) + + # Should have error event and transaction + assert len(events) == 2 + error_event, tx = events + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "RuntimeError" + + +def test_resource_handler_sync(sentry_init, capture_events): + """Test that synchronous resource handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-resource", transport="pipe") + request_ctx.set(mock_ctx) + + @server.read_resource() + def test_resource(uri): + return {"content": "file contents", "mime_type": "text/plain"} + + with start_transaction(name="mcp tx"): + uri = MockURI("file:///path/to/file.txt") + result = test_resource(uri) + + assert result["content"] == "file contents" + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "resources/read file:///path/to/file.txt" + assert span["origin"] == "auto.ai.mcp" + + # Check span data + assert span["data"][SPANDATA.MCP_RESOURCE_URI] == "file:///path/to/file.txt" + assert span["data"][SPANDATA.MCP_METHOD_NAME] == "resources/read" + assert span["data"][SPANDATA.MCP_TRANSPORT] == "pipe" + assert span["data"][SPANDATA.MCP_REQUEST_ID] == "req-resource" + assert span["data"][SPANDATA.MCP_RESOURCE_PROTOCOL] == "file" + # Resources don't capture result content + assert SPANDATA.MCP_TOOL_RESULT_CONTENT not in span["data"] + + +@pytest.mark.asyncio +async def test_resource_handler_async(sentry_init, capture_events): + """Test that async resource handlers create proper spans""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext( + request_id="req-async-resource", session_id="session-res", transport="tcp" + ) + request_ctx.set(mock_ctx) + + @server.read_resource() + async def test_resource_async(uri): + return {"data": "resource data"} + + with start_transaction(name="mcp tx"): + uri = MockURI("https://example.com/resource") + result = await test_resource_async(uri) + + assert result["data"] == "resource data" + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 1 + + span = tx["spans"][0] + assert span["op"] == OP.MCP_SERVER + assert span["description"] == "resources/read https://example.com/resource" + + assert span["data"][SPANDATA.MCP_RESOURCE_URI] == "https://example.com/resource" + assert span["data"][SPANDATA.MCP_RESOURCE_PROTOCOL] == "https" + assert span["data"][SPANDATA.MCP_SESSION_ID] == "session-res" + + +def test_resource_handler_with_error(sentry_init, capture_events): + """Test that resource handler errors are captured""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-error-resource", transport="pipe") + request_ctx.set(mock_ctx) + + @server.read_resource() + def failing_resource(uri): + raise FileNotFoundError("Resource not found") + + with start_transaction(name="mcp tx"): + with pytest.raises(FileNotFoundError): + uri = MockURI("file:///missing.txt") + failing_resource(uri) + + # Should have error event and transaction + assert len(events) == 2 + error_event, tx = events + + assert error_event["level"] == "error" + assert error_event["exception"]["values"][0]["type"] == "FileNotFoundError" + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (False, False)], +) +def test_tool_result_extraction_tuple( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test extraction of tool results from tuple format (UnstructuredContent, StructuredContent)""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-tuple", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def test_tool_tuple(tool_name, arguments): + # Return CombinationContent: (UnstructuredContent, StructuredContent) + unstructured = [MockTextContent("Result text")] + structured = {"key": "value", "count": 5} + return (unstructured, structured) + + with start_transaction(name="mcp tx"): + test_tool_tuple("combo_tool", {}) + + (tx,) = events + span = tx["spans"][0] + + # Should extract the structured content (second element of tuple) only with PII + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT] == json.dumps( + { + "key": "value", + "count": 5, + } + ) + assert span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT_COUNT] == 2 + else: + assert SPANDATA.MCP_TOOL_RESULT_CONTENT not in span["data"] + assert SPANDATA.MCP_TOOL_RESULT_CONTENT_COUNT not in span["data"] + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (False, False)], +) +def test_tool_result_extraction_unstructured( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test extraction of tool results from UnstructuredContent (list of content blocks)""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-unstructured", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def test_tool_unstructured(tool_name, arguments): + # Return UnstructuredContent as list of content blocks + return [ + MockTextContent("First part"), + MockTextContent("Second part"), + ] + + with start_transaction(name="mcp tx"): + test_tool_unstructured("text_tool", {}) + + (tx,) = events + span = tx["spans"][0] + + # Should extract and join text from content blocks only with PII + if send_default_pii and include_prompts: + assert ( + span["data"][SPANDATA.MCP_TOOL_RESULT_CONTENT] == '"First part Second part"' + ) + else: + assert SPANDATA.MCP_TOOL_RESULT_CONTENT not in span["data"] + + +def test_request_context_no_context(sentry_init, capture_events): + """Test handling when no request context is available""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Clear request context (simulating no context available) + # This will cause a LookupError when trying to get context + request_ctx.set(None) + + @server.call_tool() + def test_tool_no_ctx(tool_name, arguments): + return {"result": "ok"} + + with start_transaction(name="mcp tx"): + # This should work even without request context + try: + test_tool_no_ctx("tool", {}) + except LookupError: + # If it raises LookupError, that's expected when context is truly missing + pass + + # Should still create span even if context is missing + (tx,) = events + span = tx["spans"][0] + + # Transport defaults to "pipe" when no context + assert span["data"][SPANDATA.MCP_TRANSPORT] == "pipe" + # Request ID and Session ID should not be present + assert SPANDATA.MCP_REQUEST_ID not in span["data"] + assert SPANDATA.MCP_SESSION_ID not in span["data"] + + +def test_span_origin(sentry_init, capture_events): + """Test that span origin is set correctly""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-origin", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def test_tool(tool_name, arguments): + return {"result": "test"} + + with start_transaction(name="mcp tx"): + test_tool("origin_test", {}) + + (tx,) = events + + assert tx["contexts"]["trace"]["origin"] == "manual" + assert tx["spans"][0]["origin"] == "auto.ai.mcp" + + +def test_multiple_handlers(sentry_init, capture_events): + """Test that multiple handler calls create multiple spans""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-multi", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def tool1(tool_name, arguments): + return {"result": "tool1"} + + @server.call_tool() + def tool2(tool_name, arguments): + return {"result": "tool2"} + + @server.get_prompt() + def prompt1(name, arguments): + return MockGetPromptResult([MockPromptMessage("user", "Test prompt")]) + + with start_transaction(name="mcp tx"): + tool1("tool_a", {}) + tool2("tool_b", {}) + prompt1("prompt_a", {}) + + (tx,) = events + assert tx["type"] == "transaction" + assert len(tx["spans"]) == 3 + + # Check that we have different span types + span_ops = [span["op"] for span in tx["spans"]] + assert all(op == OP.MCP_SERVER for op in span_ops) + + span_descriptions = [span["description"] for span in tx["spans"]] + assert "tools/call tool_a" in span_descriptions + assert "tools/call tool_b" in span_descriptions + assert "prompts/get prompt_a" in span_descriptions + + +@pytest.mark.parametrize( + "send_default_pii, include_prompts", + [(True, True), (False, False)], +) +def test_prompt_with_dict_result( + sentry_init, capture_events, send_default_pii, include_prompts +): + """Test prompt handler with dict result instead of GetPromptResult object""" + sentry_init( + integrations=[MCPIntegration(include_prompts=include_prompts)], + traces_sample_rate=1.0, + send_default_pii=send_default_pii, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-dict-prompt", transport="pipe") + request_ctx.set(mock_ctx) + + @server.get_prompt() + def test_prompt_dict(name, arguments): + # Return dict format instead of GetPromptResult object + return { + "messages": [ + {"role": "user", "content": {"text": "Hello from dict"}}, + ] + } + + with start_transaction(name="mcp tx"): + test_prompt_dict("dict_prompt", {}) + + (tx,) = events + span = tx["spans"][0] + + # Message count is always captured + assert span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_COUNT] == 1 + + # Role and content only captured with PII + if send_default_pii and include_prompts: + assert span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE] == "user" + assert ( + span["data"][SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT] + == "Hello from dict" + ) + else: + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_ROLE not in span["data"] + assert SPANDATA.MCP_PROMPT_RESULT_MESSAGE_CONTENT not in span["data"] + + +def test_resource_without_protocol(sentry_init, capture_events): + """Test resource handler with URI without protocol scheme""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-no-proto", transport="pipe") + request_ctx.set(mock_ctx) + + @server.read_resource() + def test_resource(uri): + return {"data": "test"} + + with start_transaction(name="mcp tx"): + # URI without protocol + test_resource("simple-path") + + (tx,) = events + span = tx["spans"][0] + + assert span["data"][SPANDATA.MCP_RESOURCE_URI] == "simple-path" + # No protocol should be set + assert SPANDATA.MCP_RESOURCE_PROTOCOL not in span["data"] + + +def test_tool_with_complex_arguments(sentry_init, capture_events): + """Test tool handler with complex nested arguments""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-complex", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def test_tool_complex(tool_name, arguments): + return {"processed": True} + + with start_transaction(name="mcp tx"): + complex_args = { + "nested": {"key": "value", "list": [1, 2, 3]}, + "string": "test", + "number": 42, + } + test_tool_complex("complex_tool", complex_args) + + (tx,) = events + span = tx["spans"][0] + + # Complex arguments should be serialized + assert span["data"]["mcp.request.argument.nested"] == json.dumps( + {"key": "value", "list": [1, 2, 3]} + ) + assert span["data"]["mcp.request.argument.string"] == '"test"' + assert span["data"]["mcp.request.argument.number"] == "42" + + +@pytest.mark.asyncio +async def test_async_handlers_mixed(sentry_init, capture_events): + """Test mixing sync and async handlers in the same transaction""" + sentry_init( + integrations=[MCPIntegration()], + traces_sample_rate=1.0, + ) + events = capture_events() + + server = Server("test-server") + + # Set up mock request context + mock_ctx = MockRequestContext(request_id="req-mixed", transport="pipe") + request_ctx.set(mock_ctx) + + @server.call_tool() + def sync_tool(tool_name, arguments): + return {"type": "sync"} + + @server.call_tool() + async def async_tool(tool_name, arguments): + return {"type": "async"} + + with start_transaction(name="mcp tx"): + sync_result = sync_tool("sync", {}) + async_result = await async_tool("async", {}) + + assert sync_result["type"] == "sync" + assert async_result["type"] == "async" + + (tx,) = events + assert len(tx["spans"]) == 2 + + # Both should be instrumented correctly + assert all(span["op"] == OP.MCP_SERVER for span in tx["spans"]) diff --git a/tox.ini b/tox.ini index 0e9d3a7bb7..974970f168 100644 --- a/tox.ini +++ b/tox.ini @@ -65,7 +65,7 @@ envlist = {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.35.3 + {py3.8,py3.12,py3.13}-huggingface_hub-v0.36.0 {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc7 {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 @@ -82,6 +82,11 @@ envlist = {py3.9,py3.12,py3.13}-litellm-v1.77.7 {py3.9,py3.12,py3.13}-litellm-v1.78.7 + {py3.10,py3.12,py3.13}-mcp-v1.15.0 + {py3.10,py3.12,py3.13}-mcp-v1.16.0 + {py3.10,py3.12,py3.13}-mcp-v1.17.0 + {py3.10,py3.12,py3.13}-mcp-v1.18.0 + {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 {py3.9,py3.12,py3.13}-openai-base-v2.6.0 @@ -365,7 +370,7 @@ deps = huggingface_hub-v0.24.7: huggingface_hub==0.24.7 huggingface_hub-v0.28.1: huggingface_hub==0.28.1 huggingface_hub-v0.32.6: huggingface_hub==0.32.6 - huggingface_hub-v0.35.3: huggingface_hub==0.35.3 + huggingface_hub-v0.36.0: huggingface_hub==0.36.0 huggingface_hub-v1.0.0rc7: huggingface_hub==1.0.0rc7 huggingface_hub: responses huggingface_hub: pytest-httpx @@ -395,6 +400,12 @@ deps = litellm-v1.77.7: litellm==1.77.7 litellm-v1.78.7: litellm==1.78.7 + mcp-v1.15.0: mcp==1.15.0 + mcp-v1.16.0: mcp==1.16.0 + mcp-v1.17.0: mcp==1.17.0 + mcp-v1.18.0: mcp==1.18.0 + mcp: pytest-asyncio + openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 openai-base-v2.6.0: openai==2.6.0 @@ -787,6 +798,7 @@ setenv = litellm: TESTPATH=tests/integrations/litellm litestar: TESTPATH=tests/integrations/litestar loguru: TESTPATH=tests/integrations/loguru + mcp: TESTPATH=tests/integrations/mcp openai-base: TESTPATH=tests/integrations/openai openai-notiktoken: TESTPATH=tests/integrations/openai openai_agents: TESTPATH=tests/integrations/openai_agents From ee3629cec4b0219b615e94f02b93428d2273d8d3 Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Thu, 23 Oct 2025 17:27:21 +0200 Subject: [PATCH 571/591] fix(ai): truncate messages for google genai (#4992) --- sentry_sdk/integrations/google_genai/utils.py | 22 ++++++--- .../google_genai/test_google_genai.py | 48 +++++++++++++++++++ 2 files changed, 64 insertions(+), 6 deletions(-) diff --git a/sentry_sdk/integrations/google_genai/utils.py b/sentry_sdk/integrations/google_genai/utils.py index ff973b02d9..a28c9cc47c 100644 --- a/sentry_sdk/integrations/google_genai/utils.py +++ b/sentry_sdk/integrations/google_genai/utils.py @@ -15,7 +15,11 @@ ) import sentry_sdk -from sentry_sdk.ai.utils import set_data_normalized +from sentry_sdk.ai.utils import ( + set_data_normalized, + truncate_and_annotate_messages, + normalize_message_roles, +) from sentry_sdk.consts import OP, SPANDATA from sentry_sdk.scope import should_send_default_pii from sentry_sdk.utils import ( @@ -462,12 +466,18 @@ def set_span_data_for_request(span, integration, model, contents, kwargs): messages.append({"role": "user", "content": contents_text}) if messages: - set_data_normalized( - span, - SPANDATA.GEN_AI_REQUEST_MESSAGES, - messages, - unpack=False, + normalized_messages = normalize_message_roles(messages) + scope = sentry_sdk.get_current_scope() + messages_data = truncate_and_annotate_messages( + normalized_messages, span, scope ) + if messages_data is not None: + set_data_normalized( + span, + SPANDATA.GEN_AI_REQUEST_MESSAGES, + messages_data, + unpack=False, + ) # Extract parameters directly from config (not nested under generation_config) for param, span_key in [ diff --git a/tests/integrations/google_genai/test_google_genai.py b/tests/integrations/google_genai/test_google_genai.py index 470be31944..268d7fbca9 100644 --- a/tests/integrations/google_genai/test_google_genai.py +++ b/tests/integrations/google_genai/test_google_genai.py @@ -905,3 +905,51 @@ def test_tool_calls_extraction(sentry_init, capture_events, mock_genai_client): assert tool_calls[1]["type"] == "function_call" # Arguments are serialized as JSON strings assert json.loads(tool_calls[1]["arguments"]) == {"timezone": "PST"} + + +def test_google_genai_message_truncation( + sentry_init, capture_events, mock_genai_client +): + """Test that large messages are truncated properly in Google GenAI integration.""" + sentry_init( + integrations=[GoogleGenAIIntegration(include_prompts=True)], + traces_sample_rate=1.0, + send_default_pii=True, + ) + events = capture_events() + + large_content = ( + "This is a very long message that will exceed our size limits. " * 1000 + ) + small_content = "This is a small user message" + + mock_http_response = create_mock_http_response(EXAMPLE_API_RESPONSE_JSON) + + with mock.patch.object( + mock_genai_client._api_client, "request", return_value=mock_http_response + ): + with start_transaction(name="google_genai"): + mock_genai_client.models.generate_content( + model="gemini-1.5-flash", + contents=small_content, + config=create_test_config( + system_instruction=large_content, + ), + ) + + (event,) = events + invoke_span = event["spans"][0] + assert SPANDATA.GEN_AI_REQUEST_MESSAGES in invoke_span["data"] + + messages_data = invoke_span["data"][SPANDATA.GEN_AI_REQUEST_MESSAGES] + assert isinstance(messages_data, str) + + parsed_messages = json.loads(messages_data) + assert isinstance(parsed_messages, list) + assert len(parsed_messages) == 1 + assert parsed_messages[0]["role"] == "user" + assert small_content in parsed_messages[0]["content"] + + assert ( + event["_meta"]["spans"]["0"]["data"]["gen_ai.request.messages"][""]["len"] == 2 + ) From 104db8c32487e21c2e76218d374b5aa239d40ffb Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Fri, 24 Oct 2025 08:57:59 +0200 Subject: [PATCH 572/591] ci: Prepare for new major branch, remove potel-base from actions (#5003) ### Description 1. Remove `potel-base` from GH action triggers. 2. Let's use the name `major/` for major branches from now on. Added the pattern to our test workflow files so that we don't need to add each new major branch to them by hand in the future. 3. Pulled in new test matrix updates as a side effect of regenerating the YAMLs. #### Issues Closes https://github.com/getsentry/sentry-python/issues/5004 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .../workflows/test-integrations-common.yml | 2 +- .github/workflows/test-integrations-dbs.yml | 2 +- .github/workflows/test-integrations-flags.yml | 2 +- .../workflows/test-integrations-gevent.yml | 2 +- .../workflows/test-integrations-graphql.yml | 2 +- .github/workflows/test-integrations-misc.yml | 2 +- .../workflows/test-integrations-network.yml | 2 +- .github/workflows/test-integrations-tasks.yml | 2 +- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/test-integrations-web-2.yml | 2 +- scripts/populate_tox/releases.jsonl | 17 +++++---- .../split_tox_gh_actions/templates/base.jinja | 2 +- tox.ini | 38 +++++++++++-------- 15 files changed, 45 insertions(+), 36 deletions(-) diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index d09213ebd6..b5a09634db 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 9b4b40f299..67d5da7a55 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-common.yml b/.github/workflows/test-integrations-common.yml index 0d1f2a90f8..5ac65c327f 100644 --- a/.github/workflows/test-integrations-common.yml +++ b/.github/workflows/test-integrations-common.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-dbs.yml b/.github/workflows/test-integrations-dbs.yml index a778ce86d5..7d55c10230 100644 --- a/.github/workflows/test-integrations-dbs.yml +++ b/.github/workflows/test-integrations-dbs.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index e50e98fb54..4b3b6278ef 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-gevent.yml b/.github/workflows/test-integrations-gevent.yml index 8b077f357c..998a3c0974 100644 --- a/.github/workflows/test-integrations-gevent.yml +++ b/.github/workflows/test-integrations-gevent.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index c9e858bc58..d11bbf4aaa 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 56a31bff27..5e10dc8b6c 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 2dc22c3ec1..2e5d0024d2 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-tasks.yml b/.github/workflows/test-integrations-tasks.yml index 2dfae9c9a3..d9f713177a 100644 --- a/.github/workflows/test-integrations-tasks.yml +++ b/.github/workflows/test-integrations-tasks.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index 42957e7066..f0d79d7fa8 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/.github/workflows/test-integrations-web-2.yml b/.github/workflows/test-integrations-web-2.yml index 2cd0980962..8907c9a29d 100644 --- a/.github/workflows/test-integrations-web-2.yml +++ b/.github/workflows/test-integrations-web-2.yml @@ -8,7 +8,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: # Cancel in progress workflows on pull_requests. # https://docs.github.com/en/actions/using-jobs/using-concurrency#example-using-a-fallback-value diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index ae75704553..8148cad69b 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -47,7 +47,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.57", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.58", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -66,10 +66,10 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.105.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.119.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.107.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.92.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.93.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}} @@ -94,7 +94,7 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.7.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.3", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.4", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} @@ -117,7 +117,7 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.15.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.16.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.17.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.18.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.19.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} @@ -134,7 +134,10 @@ {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.17", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.1.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.2.1", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.4.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}} diff --git a/scripts/split_tox_gh_actions/templates/base.jinja b/scripts/split_tox_gh_actions/templates/base.jinja index 5d59e63348..8d618d228c 100644 --- a/scripts/split_tox_gh_actions/templates/base.jinja +++ b/scripts/split_tox_gh_actions/templates/base.jinja @@ -11,7 +11,7 @@ on: branches: - master - release/** - - potel-base + - major/** pull_request: diff --git a/tox.ini b/tox.ini index 974970f168..a3aed00f00 100644 --- a/tox.ini +++ b/tox.ini @@ -85,7 +85,7 @@ envlist = {py3.10,py3.12,py3.13}-mcp-v1.15.0 {py3.10,py3.12,py3.13}-mcp-v1.16.0 {py3.10,py3.12,py3.13}-mcp-v1.17.0 - {py3.10,py3.12,py3.13}-mcp-v1.18.0 + {py3.10,py3.12,py3.13}-mcp-v1.19.0 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 @@ -100,14 +100,17 @@ envlist = {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 {py3.10,py3.12,py3.13}-openai_agents-v0.4.1 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.17 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.18 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.1.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.2.1 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.4.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.13,py3.14}-boto3-v1.40.57 + {py3.9,py3.13,py3.14}-boto3-v1.40.58 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -200,7 +203,7 @@ envlist = {py3.9,py3.12,py3.13}-dramatiq-v1.18.0 {py3.6,py3.7}-huey-v2.1.3 - {py3.6,py3.11,py3.12}-huey-v2.5.3 + {py3.6,py3.11,py3.12}-huey-v2.5.4 {py3.9,py3.10}-ray-v2.7.2 {py3.9,py3.12,py3.13}-ray-v2.50.1 @@ -233,9 +236,9 @@ envlist = {py3.9,py3.13,py3.14}-starlette-v0.48.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 - {py3.7,py3.10,py3.11}-fastapi-v0.92.0 - {py3.8,py3.10,py3.11}-fastapi-v0.105.0 - {py3.8,py3.13,py3.14}-fastapi-v0.119.1 + {py3.7,py3.10,py3.11}-fastapi-v0.93.0 + {py3.8,py3.10,py3.11}-fastapi-v0.107.0 + {py3.8,py3.13,py3.14}-fastapi-v0.120.0 # ~~~ Web 2 ~~~ @@ -403,7 +406,7 @@ deps = mcp-v1.15.0: mcp==1.15.0 mcp-v1.16.0: mcp==1.16.0 mcp-v1.17.0: mcp==1.17.0 - mcp-v1.18.0: mcp==1.18.0 + mcp-v1.19.0: mcp==1.19.0 mcp: pytest-asyncio openai-base-v1.0.1: openai==1.0.1 @@ -425,7 +428,10 @@ deps = openai_agents-v0.4.1: openai-agents==0.4.1 openai_agents: pytest-asyncio - pydantic_ai-v1.0.17: pydantic-ai==1.0.17 + pydantic_ai-v1.0.18: pydantic-ai==1.0.18 + pydantic_ai-v1.1.0: pydantic-ai==1.1.0 + pydantic_ai-v1.2.1: pydantic-ai==1.2.1 + pydantic_ai-v1.4.0: pydantic-ai==1.4.0 pydantic_ai: pytest-asyncio @@ -433,7 +439,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.57: boto3==1.40.57 + boto3-v1.40.58: boto3==1.40.58 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -563,7 +569,7 @@ deps = dramatiq-v1.18.0: dramatiq==1.18.0 huey-v2.1.3: huey==2.1.3 - huey-v2.5.3: huey==2.5.3 + huey-v2.5.4: huey==2.5.4 ray-v2.7.2: ray==2.7.2 ray-v2.50.1: ray==2.50.1 @@ -636,17 +642,17 @@ deps = {py3.6}-starlette: aiocontextvars fastapi-v0.79.1: fastapi==0.79.1 - fastapi-v0.92.0: fastapi==0.92.0 - fastapi-v0.105.0: fastapi==0.105.0 - fastapi-v0.119.1: fastapi==0.119.1 + fastapi-v0.93.0: fastapi==0.93.0 + fastapi-v0.107.0: fastapi==0.107.0 + fastapi-v0.120.0: fastapi==0.120.0 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart fastapi: requests fastapi: anyio<4 fastapi-v0.79.1: httpx<0.28.0 - fastapi-v0.92.0: httpx<0.28.0 - fastapi-v0.105.0: httpx<0.28.0 + fastapi-v0.93.0: httpx<0.28.0 + fastapi-v0.107.0: httpx<0.28.0 {py3.6}-fastapi: aiocontextvars From e4e11efb87f04863e80932241fa726bb8ce85edf Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue, 28 Oct 2025 08:30:04 +0100 Subject: [PATCH 573/591] build(deps): bump actions/upload-artifact from 4 to 5 (#5032) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps [actions/upload-artifact](https://github.com/actions/upload-artifact) from 4 to 5.
Release notes

Sourced from actions/upload-artifact's releases.

v5.0.0

What's Changed

BREAKING CHANGE: this update supports Node v24.x. This is not a breaking change per-se but we're treating it as such.

New Contributors

Full Changelog: https://github.com/actions/upload-artifact/compare/v4...v5.0.0

v4.6.2

What's Changed

New Contributors

Full Changelog: https://github.com/actions/upload-artifact/compare/v4...v4.6.2

v4.6.1

What's Changed

Full Changelog: https://github.com/actions/upload-artifact/compare/v4...v4.6.1

v4.6.0

What's Changed

Full Changelog: https://github.com/actions/upload-artifact/compare/v4...v4.6.0

v4.5.0

What's Changed

New Contributors

... (truncated)

Commits
  • 330a01c Merge pull request #734 from actions/danwkennedy/prepare-5.0.0
  • 03f2824 Update github.dep.yml
  • 905a1ec Prepare v5.0.0
  • 2d9f9cd Merge pull request #725 from patrikpolyak/patch-1
  • 9687587 Merge branch 'main' into patch-1
  • 2848b2c Merge pull request #727 from danwkennedy/patch-1
  • 9b51177 Spell out the first use of GHES
  • cd231ca Update GHES guidance to include reference to Node 20 version
  • de65e23 Merge pull request #712 from actions/nebuk89-patch-1
  • 8747d8c Update README.md
  • Additional commits viewable in compare view

[![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=actions/upload-artifact&package-manager=github_actions&previous-version=4&new-version=5)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) ---
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- .github/workflows/ci.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 43f9d296a1..9ad1e9b66d 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -55,7 +55,7 @@ jobs: # This will also trigger "make dist" that creates the Python packages make aws-lambda-layer - name: Upload Python Packages - uses: actions/upload-artifact@v4 + uses: actions/upload-artifact@v5 with: name: artifact-build_lambda_layer path: | @@ -79,7 +79,7 @@ jobs: make apidocs cd docs/_build && zip -r gh-pages ./ - - uses: actions/upload-artifact@v4 + - uses: actions/upload-artifact@v5 with: name: artifact-docs path: | @@ -93,7 +93,7 @@ jobs: runs-on: ubuntu-latest needs: [build_lambda_layer, docs] steps: - - uses: actions/upload-artifact/merge@v4 + - uses: actions/upload-artifact/merge@v5 with: # Craft expects release assets from github to be a single artifact named after the sha. name: ${{ github.sha }} From 8ab8f3b3895944ac52518d47a4ea436f715bab6c Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Tue, 28 Oct 2025 07:33:30 +0000 Subject: [PATCH 574/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(10/27)=20(#5033)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ivana Kellyer --- scripts/populate_tox/releases.jsonl | 11 ++++++----- tox.ini | 26 ++++++++++++++------------ 2 files changed, 20 insertions(+), 17 deletions(-) diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 8148cad69b..c639a53d1f 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -47,7 +47,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.58", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.59", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} @@ -58,7 +58,7 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.10.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.15.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.19.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.20.0", "yanked": false}} {"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}} @@ -109,6 +109,7 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.7", "yanked": false}} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.79.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} @@ -125,19 +126,19 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.88.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.4.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.0", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.1", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.1", "yanked": false}} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.2", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.1.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.2.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.4.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.6.0", "yanked": false}} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}} diff --git a/tox.ini b/tox.ini index a3aed00f00..7e5e60f39a 100644 --- a/tox.ini +++ b/tox.ini @@ -55,7 +55,7 @@ envlist = {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.10.0 {py3.9,py3.11,py3.12}-cohere-v5.15.0 - {py3.9,py3.11,py3.12}-cohere-v5.19.0 + {py3.9,py3.11,py3.12}-cohere-v5.20.0 {py3.9,py3.12,py3.13}-google_genai-v1.29.0 {py3.9,py3.12,py3.13}-google_genai-v1.35.0 @@ -81,6 +81,7 @@ envlist = {py3.9,py3.12,py3.13}-litellm-v1.77.7 {py3.9,py3.12,py3.13}-litellm-v1.78.7 + {py3.9,py3.12,py3.13}-litellm-v1.79.0 {py3.10,py3.12,py3.13}-mcp-v1.15.0 {py3.10,py3.12,py3.13}-mcp-v1.16.0 @@ -89,28 +90,28 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.9,py3.12,py3.13}-openai-base-v2.6.0 + {py3.9,py3.12,py3.13}-openai-base-v2.6.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.9,py3.12,py3.13}-openai-notiktoken-v2.6.0 + {py3.9,py3.12,py3.13}-openai-notiktoken-v2.6.1 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 {py3.10,py3.12,py3.13}-openai_agents-v0.2.11 - {py3.10,py3.12,py3.13}-openai_agents-v0.4.1 + {py3.10,py3.12,py3.13}-openai_agents-v0.4.2 {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.18 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.1.0 {py3.10,py3.12,py3.13}-pydantic_ai-v1.2.1 {py3.10,py3.12,py3.13}-pydantic_ai-v1.4.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.6.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.13,py3.14}-boto3-v1.40.58 + {py3.9,py3.13,py3.14}-boto3-v1.40.59 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -362,7 +363,7 @@ deps = cohere-v5.4.0: cohere==5.4.0 cohere-v5.10.0: cohere==5.10.0 cohere-v5.15.0: cohere==5.15.0 - cohere-v5.19.0: cohere==5.19.0 + cohere-v5.20.0: cohere==5.20.0 google_genai-v1.29.0: google-genai==1.29.0 google_genai-v1.35.0: google-genai==1.35.0 @@ -402,6 +403,7 @@ deps = litellm-v1.77.7: litellm==1.77.7 litellm-v1.78.7: litellm==1.78.7 + litellm-v1.79.0: litellm==1.79.0 mcp-v1.15.0: mcp==1.15.0 mcp-v1.16.0: mcp==1.16.0 @@ -411,27 +413,27 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.6.0: openai==2.6.0 + openai-base-v2.6.1: openai==2.6.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.6.0: openai==2.6.0 + openai-notiktoken-v2.6.1: openai==2.6.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 openai_agents-v0.0.19: openai-agents==0.0.19 openai_agents-v0.1.0: openai-agents==0.1.0 openai_agents-v0.2.11: openai-agents==0.2.11 - openai_agents-v0.4.1: openai-agents==0.4.1 + openai_agents-v0.4.2: openai-agents==0.4.2 openai_agents: pytest-asyncio pydantic_ai-v1.0.18: pydantic-ai==1.0.18 - pydantic_ai-v1.1.0: pydantic-ai==1.1.0 pydantic_ai-v1.2.1: pydantic-ai==1.2.1 pydantic_ai-v1.4.0: pydantic-ai==1.4.0 + pydantic_ai-v1.6.0: pydantic-ai==1.6.0 pydantic_ai: pytest-asyncio @@ -439,7 +441,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.58: boto3==1.40.58 + boto3-v1.40.59: boto3==1.40.59 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 From e6366194f61fdae97e7867c6457b67ac9a77014c Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Tue, 28 Oct 2025 08:35:43 +0100 Subject: [PATCH 575/591] fix(strawberry): Remove autodetection, always use sync extension (#4984) ### Description #### Problem There are two ways to instrument strawberry-graphl apps: via a sync or an async extension. We enable one or the other based on what web framework is installed (async framework -> async extension, sync framework -> sync extension). This auto-detection can be overridden via an integration option. At some point (SDK 2.0?), we added `StrawberryIntegration` to auto-enabling integrations, which means the brittle auto-detection kicks in as soon as someone has `strawberry-graphl` installed. This can lead to issues, most notably when we auto-enable the async version of the extension even though the user's Strawberry app is actually sync. #### Options 1. Removing the auto-detection, always enabling the sync version if not specified otherwise. This way we'll never mistakenly enable async code in a sync app. We also had a [report](https://github.com/getsentry/sentry-python/issues/3670#issuecomment-2436511882) at some point that the sync extension actually performs better than async, so enabling sync by default shouldn't be a big problem in async apps. 2. Removing `StrawberryIntegration` from auto-enabling integrations. Breaking change. People might just lose their traces and errors. 3. Improving the auto-detection. Best option, but out of ideas how to do this. Went with 1), all things considered it's the least breaking change (unless there's a way to do 3). Needs a big callout in the changelog anyway though. #### Issues Closes https://github.com/getsentry/sentry-python/issues/4980 #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --- sentry_sdk/integrations/strawberry.py | 19 ++++---- .../strawberry/test_strawberry.py | 45 +++++-------------- 2 files changed, 20 insertions(+), 44 deletions(-) diff --git a/sentry_sdk/integrations/strawberry.py b/sentry_sdk/integrations/strawberry.py index ae7d273079..f30e95e7f6 100644 --- a/sentry_sdk/integrations/strawberry.py +++ b/sentry_sdk/integrations/strawberry.py @@ -1,5 +1,6 @@ import functools import hashlib +import warnings from inspect import isawaitable import sentry_sdk @@ -95,17 +96,19 @@ def _sentry_patched_schema_init(self, *args, **kwargs): extensions = kwargs.get("extensions") or [] + should_use_async_extension = None # type: Optional[bool] if integration.async_execution is not None: should_use_async_extension = integration.async_execution else: # try to figure it out ourselves should_use_async_extension = _guess_if_using_async(extensions) - logger.info( - "Assuming strawberry is running %s. If not, initialize it as StrawberryIntegration(async_execution=%s).", - "async" if should_use_async_extension else "sync", - "False" if should_use_async_extension else "True", - ) + if should_use_async_extension is None: + warnings.warn( + "Assuming strawberry is running sync. If not, initialize the integration as StrawberryIntegration(async_execution=True).", + stacklevel=2, + ) + should_use_async_extension = False # remove the built in strawberry sentry extension, if present extensions = [ @@ -382,12 +385,10 @@ def inner(event, hint): def _guess_if_using_async(extensions): - # type: (List[SchemaExtension]) -> bool + # type: (List[SchemaExtension]) -> Optional[bool] if StrawberrySentryAsyncExtension in extensions: return True elif StrawberrySentrySyncExtension in extensions: return False - return bool( - {"starlette", "starlite", "litestar", "fastapi"} & set(_get_installed_modules()) - ) + return None diff --git a/tests/integrations/strawberry/test_strawberry.py b/tests/integrations/strawberry/test_strawberry.py index 7b40b238d2..ba645da257 100644 --- a/tests/integrations/strawberry/test_strawberry.py +++ b/tests/integrations/strawberry/test_strawberry.py @@ -103,49 +103,24 @@ def create_app(schema): def test_async_execution_uses_async_extension(sentry_init): sentry_init(integrations=[StrawberryIntegration(async_execution=True)]) - with mock.patch( - "sentry_sdk.integrations.strawberry._get_installed_modules", - return_value={"flask": "2.3.3"}, - ): - # actual installed modules should not matter, the explicit option takes - # precedence - schema = strawberry.Schema(Query) - assert SentryAsyncExtension in schema.extensions + schema = strawberry.Schema(Query) + assert SentryAsyncExtension in schema.extensions + assert SentrySyncExtension not in schema.extensions def test_sync_execution_uses_sync_extension(sentry_init): sentry_init(integrations=[StrawberryIntegration(async_execution=False)]) - with mock.patch( - "sentry_sdk.integrations.strawberry._get_installed_modules", - return_value={"fastapi": "0.103.1", "starlette": "0.27.0"}, - ): - # actual installed modules should not matter, the explicit option takes - # precedence - schema = strawberry.Schema(Query) - assert SentrySyncExtension in schema.extensions - - -def test_infer_execution_type_from_installed_packages_async(sentry_init): - sentry_init(integrations=[StrawberryIntegration()]) - - with mock.patch( - "sentry_sdk.integrations.strawberry._get_installed_modules", - return_value={"fastapi": "0.103.1", "starlette": "0.27.0"}, - ): - schema = strawberry.Schema(Query) - assert SentryAsyncExtension in schema.extensions + schema = strawberry.Schema(Query) + assert SentrySyncExtension in schema.extensions + assert SentryAsyncExtension not in schema.extensions -def test_infer_execution_type_from_installed_packages_sync(sentry_init): +def test_use_sync_extension_if_not_specified(sentry_init): sentry_init(integrations=[StrawberryIntegration()]) - - with mock.patch( - "sentry_sdk.integrations.strawberry._get_installed_modules", - return_value={"flask": "2.3.3"}, - ): - schema = strawberry.Schema(Query) - assert SentrySyncExtension in schema.extensions + schema = strawberry.Schema(Query) + assert SentrySyncExtension in schema.extensions + assert SentryAsyncExtension not in schema.extensions @pytest.mark.skipif( From 6e06a07467b7b808699d67e8b71bc5ec2fbc9f6f Mon Sep 17 00:00:00 2001 From: "Zhanzhao (Deo) Liang" Date: Tue, 28 Oct 2025 15:42:34 +0800 Subject: [PATCH 576/591] fix startlette deprecation warning (#5034) --- sentry_sdk/integrations/starlette.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/integrations/starlette.py b/sentry_sdk/integrations/starlette.py index f1a0e360bb..0705da3a4c 100644 --- a/sentry_sdk/integrations/starlette.py +++ b/sentry_sdk/integrations/starlette.py @@ -326,7 +326,7 @@ def _add_user_to_sentry_scope(scope): user_info.setdefault("email", starlette_user.email) sentry_scope = sentry_sdk.get_isolation_scope() - sentry_scope.user = user_info + sentry_scope.set_user(user_info) def patch_authentication_middleware(middleware_class): From fe31660c50c649f1730f4307ae6f4c10e61acd2f Mon Sep 17 00:00:00 2001 From: Kev <6111995+k-fish@users.noreply.github.com> Date: Tue, 28 Oct 2025 03:45:30 -0400 Subject: [PATCH 577/591] fix(tracemetrics): Bump metric buffer size to 1k (#5031) We've been noticing some discards from clients for likely buffer reason. Metrics were initially set to 100 to match logs but the size limit for a metric in relay is ~1000x less (~1-2kb at the moment) so the memory pressure should be fine (it'd cap in the single digit megabyte range). --- sentry_sdk/_metrics_batcher.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sentry_sdk/_metrics_batcher.py b/sentry_sdk/_metrics_batcher.py index fd9a5d732b..27d27f2c72 100644 --- a/sentry_sdk/_metrics_batcher.py +++ b/sentry_sdk/_metrics_batcher.py @@ -12,7 +12,7 @@ class MetricsBatcher: - MAX_METRICS_BEFORE_FLUSH = 100 + MAX_METRICS_BEFORE_FLUSH = 1000 FLUSH_WAIT_TIME = 5.0 def __init__( From 64c145ffb19cb7c49b679b0cdf2e7754f0f91d12 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Tue, 28 Oct 2025 09:45:06 +0100 Subject: [PATCH 578/591] chore(metrics): Rename _metrics to metrics (#5035) Moves metrics code from `_metrics.py` to `metrics.py`, and updates metrics tests to account for the namespace change. --- sentry_sdk/__init__.py | 2 ++ sentry_sdk/{_metrics.py => metrics.py} | 0 tests/test_metrics.py | 27 +++++++++++++------------- 3 files changed, 15 insertions(+), 14 deletions(-) rename sentry_sdk/{_metrics.py => metrics.py} (100%) diff --git a/sentry_sdk/__init__.py b/sentry_sdk/__init__.py index 1939be0510..e149418c38 100644 --- a/sentry_sdk/__init__.py +++ b/sentry_sdk/__init__.py @@ -1,4 +1,5 @@ from sentry_sdk import profiler +from sentry_sdk import metrics from sentry_sdk.scope import Scope from sentry_sdk.transport import Transport, HttpTransport from sentry_sdk.client import Client @@ -48,6 +49,7 @@ "trace", "monitor", "logger", + "metrics", "profiler", "start_session", "end_session", diff --git a/sentry_sdk/_metrics.py b/sentry_sdk/metrics.py similarity index 100% rename from sentry_sdk/_metrics.py rename to sentry_sdk/metrics.py diff --git a/tests/test_metrics.py b/tests/test_metrics.py index 5e774227fd..d5ca8ec572 100644 --- a/tests/test_metrics.py +++ b/tests/test_metrics.py @@ -4,7 +4,6 @@ import pytest import sentry_sdk -from sentry_sdk import _metrics from sentry_sdk import get_client from sentry_sdk.envelope import Envelope from sentry_sdk.types import Metric @@ -39,9 +38,9 @@ def test_metrics_disabled_by_default(sentry_init, capture_envelopes): envelopes = capture_envelopes() - _metrics.count("test.counter", 1) - _metrics.gauge("test.gauge", 42) - _metrics.distribution("test.distribution", 200) + sentry_sdk.metrics.count("test.counter", 1) + sentry_sdk.metrics.gauge("test.gauge", 42) + sentry_sdk.metrics.distribution("test.distribution", 200) assert len(envelopes) == 0 @@ -50,9 +49,9 @@ def test_metrics_basics(sentry_init, capture_envelopes): sentry_init(_experiments={"enable_metrics": True}) envelopes = capture_envelopes() - _metrics.count("test.counter", 1) - _metrics.gauge("test.gauge", 42, unit="millisecond") - _metrics.distribution("test.distribution", 200, unit="second") + sentry_sdk.metrics.count("test.counter", 1) + sentry_sdk.metrics.gauge("test.gauge", 42, unit="millisecond") + sentry_sdk.metrics.distribution("test.distribution", 200, unit="second") get_client().flush() metrics = envelopes_to_metrics(envelopes) @@ -81,7 +80,7 @@ def test_metrics_experimental_option(sentry_init, capture_envelopes): sentry_init(_experiments={"enable_metrics": True}) envelopes = capture_envelopes() - _metrics.count("test.counter", 5) + sentry_sdk.metrics.count("test.counter", 5) get_client().flush() @@ -99,7 +98,7 @@ def test_metrics_with_attributes(sentry_init, capture_envelopes): ) envelopes = capture_envelopes() - _metrics.count( + sentry_sdk.metrics.count( "test.counter", 1, attributes={"endpoint": "/api/test", "status": "success"} ) @@ -121,7 +120,7 @@ def test_metrics_with_user(sentry_init, capture_envelopes): sentry_sdk.set_user( {"id": "user-123", "email": "test@example.com", "username": "testuser"} ) - _metrics.count("test.user.counter", 1) + sentry_sdk.metrics.count("test.user.counter", 1) get_client().flush() @@ -138,7 +137,7 @@ def test_metrics_with_span(sentry_init, capture_envelopes): envelopes = capture_envelopes() with sentry_sdk.start_transaction(op="test", name="test-span"): - _metrics.count("test.span.counter", 1) + sentry_sdk.metrics.count("test.span.counter", 1) get_client().flush() @@ -154,7 +153,7 @@ def test_metrics_tracing_without_performance(sentry_init, capture_envelopes): sentry_init(_experiments={"enable_metrics": True}) envelopes = capture_envelopes() - _metrics.count("test.span.counter", 1) + sentry_sdk.metrics.count("test.span.counter", 1) get_client().flush() @@ -197,8 +196,8 @@ def _before_metric(record, hint): ) envelopes = capture_envelopes() - _metrics.count("test.skip", 1) - _metrics.count("test.keep", 1) + sentry_sdk.metrics.count("test.skip", 1) + sentry_sdk.metrics.count("test.keep", 1) get_client().flush() From d7ccf06aea28c709abd3d6e7002951c4fa169fe2 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 29 Oct 2025 08:34:09 +0100 Subject: [PATCH 579/591] fix(django): Improve logic for classifying cache hits and misses (#5029) Determine cache hits and misses independently for `django.core.cache.get()` and `django.core.cache.get_many()`. When calling `get_many()`, only register a cache miss when the returned value is an empty dictionary. When calling `get()`, register a cache miss when - the return value is `None` and no default value is provided; or - the return value is equal to a provided default value. Closes https://github.com/getsentry/sentry-python/issues/5027 --- sentry_sdk/integrations/django/caching.py | 19 ++++- .../integrations/django/test_cache_module.py | 76 ++++++++++++++++++- 2 files changed, 88 insertions(+), 7 deletions(-) diff --git a/sentry_sdk/integrations/django/caching.py b/sentry_sdk/integrations/django/caching.py index 7985611761..82b602f9b5 100644 --- a/sentry_sdk/integrations/django/caching.py +++ b/sentry_sdk/integrations/django/caching.py @@ -45,7 +45,8 @@ def _instrument_call( ): # type: (CacheHandler, str, Callable[..., Any], tuple[Any, ...], dict[str, Any], Optional[str], Optional[int]) -> Any is_set_operation = method_name.startswith("set") - is_get_operation = not is_set_operation + is_get_method = method_name == "get" + is_get_many_method = method_name == "get_many" op = OP.CACHE_PUT if is_set_operation else OP.CACHE_GET description = _get_span_description(method_name, args, kwargs) @@ -69,8 +70,20 @@ def _instrument_call( span.set_data(SPANDATA.CACHE_KEY, key) item_size = None - if is_get_operation: - if value: + if is_get_many_method: + if value != {}: + item_size = len(str(value)) + span.set_data(SPANDATA.CACHE_HIT, True) + else: + span.set_data(SPANDATA.CACHE_HIT, False) + elif is_get_method: + default_value = None + if len(args) >= 2: + default_value = args[1] + elif "default" in kwargs: + default_value = kwargs["default"] + + if value != default_value: item_size = len(str(value)) span.set_data(SPANDATA.CACHE_HIT, True) else: diff --git a/tests/integrations/django/test_cache_module.py b/tests/integrations/django/test_cache_module.py index bc58dd8471..01b97c1302 100644 --- a/tests/integrations/django/test_cache_module.py +++ b/tests/integrations/django/test_cache_module.py @@ -223,8 +223,8 @@ def test_cache_spans_middleware( assert second_event["spans"][0]["data"]["cache.key"][0].startswith( "views.decorators.cache.cache_header." ) - assert not second_event["spans"][0]["data"]["cache.hit"] - assert "cache.item_size" not in second_event["spans"][0]["data"] + assert second_event["spans"][0]["data"]["cache.hit"] + assert second_event["spans"][0]["data"]["cache.item_size"] == 2 # second_event - cache.get 2 assert second_event["spans"][1]["op"] == "cache.get" assert second_event["spans"][1]["description"].startswith( @@ -501,14 +501,76 @@ def test_cache_spans_item_size(sentry_init, client, capture_events, use_django_c assert len(second_event["spans"]) == 2 assert second_event["spans"][0]["op"] == "cache.get" - assert not second_event["spans"][0]["data"]["cache.hit"] - assert "cache.item_size" not in second_event["spans"][0]["data"] + assert second_event["spans"][0]["data"]["cache.hit"] + assert second_event["spans"][0]["data"]["cache.item_size"] == 2 assert second_event["spans"][1]["op"] == "cache.get" assert second_event["spans"][1]["data"]["cache.hit"] assert second_event["spans"][1]["data"]["cache.item_size"] == 58 +@pytest.mark.forked +@pytest_mark_django_db_decorator() +def test_cache_spans_get_custom_default( + sentry_init, capture_events, use_django_caching +): + sentry_init( + integrations=[ + DjangoIntegration( + cache_spans=True, + middleware_spans=False, + signals_spans=False, + ) + ], + traces_sample_rate=1.0, + ) + events = capture_events() + + id = os.getpid() + + from django.core.cache import cache + + with sentry_sdk.start_transaction(): + cache.set(f"S{id}", "Sensitive1") + cache.set(f"S{id + 1}", "") + + cache.get(f"S{id}", "null") + cache.get(f"S{id}", default="null") + + cache.get(f"S{id + 1}", "null") + cache.get(f"S{id + 1}", default="null") + + cache.get(f"S{id + 2}", "null") + cache.get(f"S{id + 2}", default="null") + + (transaction,) = events + assert len(transaction["spans"]) == 8 + + assert transaction["spans"][0]["op"] == "cache.put" + assert transaction["spans"][0]["description"] == f"S{id}" + + assert transaction["spans"][1]["op"] == "cache.put" + assert transaction["spans"][1]["description"] == f"S{id + 1}" + + for span in (transaction["spans"][2], transaction["spans"][3]): + assert span["op"] == "cache.get" + assert span["description"] == f"S{id}" + assert span["data"]["cache.hit"] + assert span["data"]["cache.item_size"] == 10 + + for span in (transaction["spans"][4], transaction["spans"][5]): + assert span["op"] == "cache.get" + assert span["description"] == f"S{id + 1}" + assert span["data"]["cache.hit"] + assert span["data"]["cache.item_size"] == 0 + + for span in (transaction["spans"][6], transaction["spans"][7]): + assert span["op"] == "cache.get" + assert span["description"] == f"S{id + 2}" + assert not span["data"]["cache.hit"] + assert "cache.item_size" not in span["data"] + + @pytest.mark.forked @pytest_mark_django_db_decorator() def test_cache_spans_get_many(sentry_init, capture_events, use_django_caching): @@ -538,24 +600,30 @@ def test_cache_spans_get_many(sentry_init, capture_events, use_django_caching): assert transaction["spans"][0]["op"] == "cache.get" assert transaction["spans"][0]["description"] == f"S{id}, S{id + 1}" + assert not transaction["spans"][0]["data"]["cache.hit"] assert transaction["spans"][1]["op"] == "cache.get" assert transaction["spans"][1]["description"] == f"S{id}" + assert not transaction["spans"][1]["data"]["cache.hit"] assert transaction["spans"][2]["op"] == "cache.get" assert transaction["spans"][2]["description"] == f"S{id + 1}" + assert not transaction["spans"][2]["data"]["cache.hit"] assert transaction["spans"][3]["op"] == "cache.put" assert transaction["spans"][3]["description"] == f"S{id}" assert transaction["spans"][4]["op"] == "cache.get" assert transaction["spans"][4]["description"] == f"S{id}, S{id + 1}" + assert transaction["spans"][4]["data"]["cache.hit"] assert transaction["spans"][5]["op"] == "cache.get" assert transaction["spans"][5]["description"] == f"S{id}" + assert transaction["spans"][5]["data"]["cache.hit"] assert transaction["spans"][6]["op"] == "cache.get" assert transaction["spans"][6]["description"] == f"S{id + 1}" + assert not transaction["spans"][6]["data"]["cache.hit"] @pytest.mark.forked From b3b2eb62d9f1d9132aa8114a4d2eecf7e5fc517c Mon Sep 17 00:00:00 2001 From: Fabian Schindler Date: Wed, 29 Oct 2025 09:43:55 +0100 Subject: [PATCH 580/591] fix(integrations): hooking into error tracing function to find out if an execute tool span should be set to error (#4986) --- .../integrations/openai_agents/__init__.py | 2 + .../openai_agents/patches/__init__.py | 1 + .../openai_agents/patches/error_tracing.py | 77 +++++++++++++ .../openai_agents/test_openai_agents.py | 107 ++++++++++++++++++ 4 files changed, 187 insertions(+) create mode 100644 sentry_sdk/integrations/openai_agents/patches/error_tracing.py diff --git a/sentry_sdk/integrations/openai_agents/__init__.py b/sentry_sdk/integrations/openai_agents/__init__.py index 06b6459441..7e2dee0f66 100644 --- a/sentry_sdk/integrations/openai_agents/__init__.py +++ b/sentry_sdk/integrations/openai_agents/__init__.py @@ -5,6 +5,7 @@ _create_get_all_tools_wrapper, _create_run_wrapper, _patch_agent_run, + _patch_error_tracing, ) try: @@ -48,6 +49,7 @@ class OpenAIAgentsIntegration(Integration): @staticmethod def setup_once(): # type: () -> None + _patch_error_tracing() _patch_tools() _patch_model() _patch_runner() diff --git a/sentry_sdk/integrations/openai_agents/patches/__init__.py b/sentry_sdk/integrations/openai_agents/patches/__init__.py index 06bb1711f8..33058f01a1 100644 --- a/sentry_sdk/integrations/openai_agents/patches/__init__.py +++ b/sentry_sdk/integrations/openai_agents/patches/__init__.py @@ -2,3 +2,4 @@ from .tools import _create_get_all_tools_wrapper # noqa: F401 from .runner import _create_run_wrapper # noqa: F401 from .agent_run import _patch_agent_run # noqa: F401 +from .error_tracing import _patch_error_tracing # noqa: F401 diff --git a/sentry_sdk/integrations/openai_agents/patches/error_tracing.py b/sentry_sdk/integrations/openai_agents/patches/error_tracing.py new file mode 100644 index 0000000000..7d145267fc --- /dev/null +++ b/sentry_sdk/integrations/openai_agents/patches/error_tracing.py @@ -0,0 +1,77 @@ +from functools import wraps + +import sentry_sdk +from sentry_sdk.consts import SPANSTATUS +from sentry_sdk.tracing_utils import set_span_errored + +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from typing import Any, Callable, Optional + + +def _patch_error_tracing(): + # type: () -> None + """ + Patches agents error tracing function to inject our span error logic + when a tool execution fails. + + In newer versions, the function is at: agents.util._error_tracing.attach_error_to_current_span + In older versions, it was at: agents._utils.attach_error_to_current_span + + This works even when the module or function doesn't exist. + """ + error_tracing_module = None + + # Try newer location first (agents.util._error_tracing) + try: + from agents.util import _error_tracing + + error_tracing_module = _error_tracing + except (ImportError, AttributeError): + pass + + # Try older location (agents._utils) + if error_tracing_module is None: + try: + import agents._utils + + error_tracing_module = agents._utils + except (ImportError, AttributeError): + # Module doesn't exist in either location, nothing to patch + return + + # Check if the function exists + if not hasattr(error_tracing_module, "attach_error_to_current_span"): + return + + original_attach_error = error_tracing_module.attach_error_to_current_span + + @wraps(original_attach_error) + def sentry_attach_error_to_current_span(error, *args, **kwargs): + # type: (Any, *Any, **Any) -> Any + """ + Wraps agents' error attachment to also set Sentry span status to error. + This allows us to properly track tool execution errors even though + the agents library swallows exceptions. + """ + # Set the current Sentry span to errored + current_span = sentry_sdk.get_current_span() + if current_span is not None: + set_span_errored(current_span) + current_span.set_data("span.status", "error") + + # Optionally capture the error details if we have them + if hasattr(error, "__class__"): + current_span.set_data("error.type", error.__class__.__name__) + if hasattr(error, "__str__"): + error_message = str(error) + if error_message: + current_span.set_data("error.message", error_message) + + # Call the original function + return original_attach_error(error, *args, **kwargs) + + error_tracing_module.attach_error_to_current_span = ( + sentry_attach_error_to_current_span + ) diff --git a/tests/integrations/openai_agents/test_openai_agents.py b/tests/integrations/openai_agents/test_openai_agents.py index e647ce9fad..bc1de4e95b 100644 --- a/tests/integrations/openai_agents/test_openai_agents.py +++ b/tests/integrations/openai_agents/test_openai_agents.py @@ -1077,3 +1077,110 @@ def test_openai_agents_message_role_mapping(sentry_init, capture_events): # Verify no "ai" roles remain in any message for message in stored_messages: assert message["role"] != "ai" + + +@pytest.mark.asyncio +async def test_tool_execution_error_tracing(sentry_init, capture_events, test_agent): + """ + Test that tool execution errors are properly tracked via error tracing patch. + + This tests the patch of agents error tracing function to ensure execute_tool + spans are set to error status when tool execution fails. + + The function location varies by version: + - Newer versions: agents.util._error_tracing.attach_error_to_current_span + - Older versions: agents._utils.attach_error_to_current_span + """ + + @agents.function_tool + def failing_tool(message: str) -> str: + """A tool that fails""" + raise ValueError("Tool execution failed") + + # Create agent with the failing tool + agent_with_tool = test_agent.clone(tools=[failing_tool]) + + with patch.dict(os.environ, {"OPENAI_API_KEY": "test-key"}): + with patch( + "agents.models.openai_responses.OpenAIResponsesModel.get_response" + ) as mock_get_response: + # Create a mock response that includes tool call + tool_call = ResponseFunctionToolCall( + id="call_123", + call_id="call_123", + name="failing_tool", + type="function_call", + arguments='{"message": "test"}', + function=MagicMock( + name="failing_tool", arguments='{"message": "test"}' + ), + ) + + # First response with tool call + tool_response = ModelResponse( + output=[tool_call], + usage=Usage( + requests=1, input_tokens=10, output_tokens=5, total_tokens=15 + ), + response_id="resp_tool_123", + ) + + # Second response after tool error (agents library handles the error and continues) + final_response = ModelResponse( + output=[ + ResponseOutputMessage( + id="msg_final", + type="message", + status="completed", + content=[ + ResponseOutputText( + text="An error occurred while running the tool", + type="output_text", + annotations=[], + ) + ], + role="assistant", + ) + ], + usage=Usage( + requests=1, input_tokens=15, output_tokens=10, total_tokens=25 + ), + response_id="resp_final_123", + ) + + mock_get_response.side_effect = [tool_response, final_response] + + sentry_init( + integrations=[OpenAIAgentsIntegration()], + traces_sample_rate=1.0, + send_default_pii=True, + ) + + events = capture_events() + + # Note: The agents library catches tool exceptions internally, + # so we don't expect this to raise + await agents.Runner.run( + agent_with_tool, + "Please use the failing tool", + run_config=test_run_config, + ) + + (transaction,) = events + spans = transaction["spans"] + + # Find the execute_tool span + execute_tool_span = None + for span in spans: + if span.get("description", "").startswith("execute_tool failing_tool"): + execute_tool_span = span + break + + # Verify the execute_tool span was created + assert execute_tool_span is not None, "execute_tool span was not created" + assert execute_tool_span["description"] == "execute_tool failing_tool" + assert execute_tool_span["data"]["gen_ai.tool.name"] == "failing_tool" + + # Verify error status was set (this is the key test for our patch) + # The span should be marked as error because the tool execution failed + assert execute_tool_span["tags"]["status"] == "error" From 76cc4163087c6e89f406b89440be45d87ff8e148 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 29 Oct 2025 10:31:20 +0100 Subject: [PATCH 581/591] fix(google-genai): Set agent name (#5038) ### Description Cherry-picked off of https://github.com/getsentry/sentry-python/pull/5030 #### Issues Ref https://linear.app/getsentry/issue/TET-1293/make-sure-that-agent-name-is-set-on-all-of-its-gen-ai-children #### Reminders - Please add tests to validate your changes, and lint your code using `tox -e linters`. - Add GH Issue ID _&_ Linear ID (if applicable) - PR title should use [conventional commit](https://develop.sentry.dev/engineering-practices/commit-messages/#type) style (`feat:`, `fix:`, `ref:`, `meta:`) - For external contributors: [CONTRIBUTING.md](https://github.com/getsentry/sentry-python/blob/master/CONTRIBUTING.md), [Sentry SDK development docs](https://develop.sentry.dev/sdk/), [Discord community](https://discord.gg/Ww9hbqr) --------- Co-authored-by: Fabian Schindler --- sentry_sdk/integrations/google_genai/__init__.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/sentry_sdk/integrations/google_genai/__init__.py b/sentry_sdk/integrations/google_genai/__init__.py index 7175b64340..8f2d5df477 100644 --- a/sentry_sdk/integrations/google_genai/__init__.py +++ b/sentry_sdk/integrations/google_genai/__init__.py @@ -92,6 +92,7 @@ def new_generate_content_stream(self, *args, **kwargs): chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) set_span_data_for_request(chat_span, integration, model_name, contents, kwargs) chat_span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + chat_span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) try: stream = f(self, *args, **kwargs) @@ -165,6 +166,7 @@ async def new_async_generate_content_stream(self, *args, **kwargs): chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) set_span_data_for_request(chat_span, integration, model_name, contents, kwargs) chat_span.set_data(SPANDATA.GEN_AI_RESPONSE_STREAMING, True) + chat_span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) try: stream = await f(self, *args, **kwargs) @@ -233,6 +235,7 @@ def new_generate_content(self, *args, **kwargs): chat_span.set_data(SPANDATA.GEN_AI_OPERATION_NAME, "chat") chat_span.set_data(SPANDATA.GEN_AI_SYSTEM, GEN_AI_SYSTEM) chat_span.set_data(SPANDATA.GEN_AI_REQUEST_MODEL, model_name) + chat_span.set_data(SPANDATA.GEN_AI_AGENT_NAME, model_name) set_span_data_for_request( chat_span, integration, model_name, contents, kwargs ) From a7a3fb17956790a0ddba575c0436e6409e0ddec5 Mon Sep 17 00:00:00 2001 From: getsentry-bot Date: Wed, 29 Oct 2025 10:14:37 +0000 Subject: [PATCH 582/591] release: 2.43.0 --- CHANGELOG.md | 27 +++++++++++++++++++++++++++ docs/conf.py | 2 +- sentry_sdk/consts.py | 2 +- setup.py | 2 +- 4 files changed, 30 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2200c2f429..b9f7071bed 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,32 @@ # Changelog +## 2.43.0 + +### Various fixes & improvements + +- fix(google-genai): Set agent name (#5038) by @sentrivana +- fix(integrations): hooking into error tracing function to find out if an execute tool span should be set to error (#4986) by @constantinius +- fix(django): Improve logic for classifying cache hits and misses (#5029) by @alexander-alderman-webb +- chore(metrics): Rename \_metrics to metrics (#5035) by @alexander-alderman-webb +- fix(tracemetrics): Bump metric buffer size to 1k (#5031) by @k-fish +- fix startlette deprecation warning (#5034) by @DeoLeung +- fix(strawberry): Remove autodetection, always use sync extension (#4984) by @sentrivana +- ci: 🤖 Update test matrix with new releases (10/27) (#5033) by @github-actions +- build(deps): bump actions/upload-artifact from 4 to 5 (#5032) by @dependabot +- ci: Prepare for new major branch, remove potel-base from actions (#5003) by @sentrivana +- fix(ai): truncate messages for google genai (#4992) by @shellmayr +- feat(integrations): MCP Python SDK (#4964) by @constantinius +- feat(integrations): pydantic-ai integration (#4906) by @constantinius +- fix(ai): add message truncation to litellm (#4973) by @shellmayr +- feat(langchain): Support v1 (#4874) by @sentrivana +- ci: Run `common` test suite on Python 3.14t (#4969) by @alexander-alderman-webb +- feat: Officially support 3.14 & run integration tests on 3.14 (#4974) by @sentrivana +- Make logger template format safer to missing kwargs (#4981) by @sl0thentr0py +- tests(huggingface): Support 1.0.0rc7 (#4979) by @alexander-alderman-webb +- feat: Enable HTTP request code origin by default (#4967) by @alexander-alderman-webb +- ci: Run `common` test suite on Python 3.14 (#4896) by @sentrivana +- ci: 🤖 Update test matrix with new releases (10/20) (#4957) by @github-actions + ## 2.42.1 ### Various fixes & improvements diff --git a/docs/conf.py b/docs/conf.py index e92d95931e..565af1a51c 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -31,7 +31,7 @@ copyright = "2019-{}, Sentry Team and Contributors".format(datetime.now().year) author = "Sentry Team and Contributors" -release = "2.42.1" +release = "2.43.0" version = ".".join(release.split(".")[:2]) # The short X.Y version. diff --git a/sentry_sdk/consts.py b/sentry_sdk/consts.py index b041e0626b..33900acd50 100644 --- a/sentry_sdk/consts.py +++ b/sentry_sdk/consts.py @@ -1433,4 +1433,4 @@ def _get_default_options(): del _get_default_options -VERSION = "2.42.1" +VERSION = "2.43.0" diff --git a/setup.py b/setup.py index 377d6c82d6..b3b7ebb737 100644 --- a/setup.py +++ b/setup.py @@ -21,7 +21,7 @@ def get_file_text(file_name): setup( name="sentry-sdk", - version="2.42.1", + version="2.43.0", author="Sentry Team and Contributors", author_email="hello@sentry.io", url="https://github.com/getsentry/sentry-python", From b069aa24fdf3c52a9e8b75f4f83d5fee035c3234 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 29 Oct 2025 11:27:42 +0100 Subject: [PATCH 583/591] Update CHANGELOG.md --- CHANGELOG.md | 47 +++++++++++++++++++++++++++++++++++++++++------ 1 file changed, 41 insertions(+), 6 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index b9f7071bed..a5f78e7fb6 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,19 +4,55 @@ ### Various fixes & improvements -- fix(google-genai): Set agent name (#5038) by @sentrivana +- Pydantic AI integration (#4906) by @constantinius + + Enable the new Pydantic AI integration with the code snippet below, and you can use the Sentry AI dashboards to observe your AI calls: + + ```python + import sentry_sdk + from sentry_sdk.integrations.pydantic_ai import PydanticAIIntegration + sentry_sdk.init( + dsn="", + # Set traces_sample_rate to 1.0 to capture 100% + # of transactions for tracing. + traces_sample_rate=1.0, + # Add data like inputs and responses; + # see https://docs.sentry.io/platforms/python/data-management/data-collected/ for more info + send_default_pii=True, + integrations=[ + PydanticAIIntegration(), + ], + ) + ``` +- MCP Python SDK (#4964) by @constantinius + + Enable the new Python MCP integration with the code snippet below: + + ```python + import sentry_sdk + from sentry_sdk.integrations.mcp import MCPIntegration + sentry_sdk.init( + dsn="", + # Set traces_sample_rate to 1.0 to capture 100% + # of transactions for tracing. + traces_sample_rate=1.0, + # Add data like inputs and responses; + # see https://docs.sentry.io/platforms/python/data-management/data-collected/ for more info + send_default_pii=True, + integrations=[ + MCPIntegration(), + ], + ) + ``` +- fix(google-genai): Set agent name (#5038) by @constantinius - fix(integrations): hooking into error tracing function to find out if an execute tool span should be set to error (#4986) by @constantinius - fix(django): Improve logic for classifying cache hits and misses (#5029) by @alexander-alderman-webb - chore(metrics): Rename \_metrics to metrics (#5035) by @alexander-alderman-webb - fix(tracemetrics): Bump metric buffer size to 1k (#5031) by @k-fish - fix startlette deprecation warning (#5034) by @DeoLeung - fix(strawberry): Remove autodetection, always use sync extension (#4984) by @sentrivana -- ci: 🤖 Update test matrix with new releases (10/27) (#5033) by @github-actions - build(deps): bump actions/upload-artifact from 4 to 5 (#5032) by @dependabot -- ci: Prepare for new major branch, remove potel-base from actions (#5003) by @sentrivana - fix(ai): truncate messages for google genai (#4992) by @shellmayr -- feat(integrations): MCP Python SDK (#4964) by @constantinius -- feat(integrations): pydantic-ai integration (#4906) by @constantinius - fix(ai): add message truncation to litellm (#4973) by @shellmayr - feat(langchain): Support v1 (#4874) by @sentrivana - ci: Run `common` test suite on Python 3.14t (#4969) by @alexander-alderman-webb @@ -25,7 +61,6 @@ - tests(huggingface): Support 1.0.0rc7 (#4979) by @alexander-alderman-webb - feat: Enable HTTP request code origin by default (#4967) by @alexander-alderman-webb - ci: Run `common` test suite on Python 3.14 (#4896) by @sentrivana -- ci: 🤖 Update test matrix with new releases (10/20) (#4957) by @github-actions ## 2.42.1 From 32e97968773795408cdddbf86cba6591d09fdd54 Mon Sep 17 00:00:00 2001 From: Ivana Kellyer Date: Wed, 29 Oct 2025 13:14:54 +0100 Subject: [PATCH 584/591] docs: Elaborate on Strawberry autodetection in changelog (#5039) Forgot to call this out in the changelog. Fixing that. --- CHANGELOG.md | 27 ++++++++++++++++++++------- 1 file changed, 20 insertions(+), 7 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a5f78e7fb6..723d0a077a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,7 +5,7 @@ ### Various fixes & improvements - Pydantic AI integration (#4906) by @constantinius - + Enable the new Pydantic AI integration with the code snippet below, and you can use the Sentry AI dashboards to observe your AI calls: ```python @@ -25,7 +25,7 @@ ) ``` - MCP Python SDK (#4964) by @constantinius - + Enable the new Python MCP integration with the code snippet below: ```python @@ -44,13 +44,26 @@ ], ) ``` +- fix(strawberry): Remove autodetection, always use sync extension (#4984) by @sentrivana + + Previously, `StrawberryIntegration` would try to guess whether it should install the sync or async version of itself. This auto-detection was very brittle and could lead to us auto-enabling async code in a sync context. With this change, `StrawberryIntegration` remains an auto-enabling integration, but it'll enable the sync version by default. If you want to enable the async version, pass the option explicitly: + + ```python + sentry_sdk.init( + # ... + integrations=[ + StrawberryIntegration( + async_execution=True + ), + ], + ) + ``` - fix(google-genai): Set agent name (#5038) by @constantinius - fix(integrations): hooking into error tracing function to find out if an execute tool span should be set to error (#4986) by @constantinius - fix(django): Improve logic for classifying cache hits and misses (#5029) by @alexander-alderman-webb - chore(metrics): Rename \_metrics to metrics (#5035) by @alexander-alderman-webb - fix(tracemetrics): Bump metric buffer size to 1k (#5031) by @k-fish - fix startlette deprecation warning (#5034) by @DeoLeung -- fix(strawberry): Remove autodetection, always use sync extension (#4984) by @sentrivana - build(deps): bump actions/upload-artifact from 4 to 5 (#5032) by @dependabot - fix(ai): truncate messages for google genai (#4992) by @shellmayr - fix(ai): add message truncation to litellm (#4973) by @shellmayr @@ -113,7 +126,7 @@ ### Various fixes & improvements - feat: Add `concurrent.futures` patch to threading integration (#4770) by @alexander-alderman-webb - + The SDK now makes sure to automatically preserve span relationships when using `ThreadPoolExecutor`. - chore: Remove old metrics code (#4899) by @sentrivana @@ -128,7 +141,7 @@ - Add LiteLLM integration (#4864) by @constantinius Once you've enabled the [new LiteLLM integration](https://docs.sentry.io/platforms/python/integrations/litellm/), you can use the Sentry AI Agents Monitoring, a Sentry dashboard that helps you understand what's going on with your AI requests: - + ```python import sentry_sdk from sentry_sdk.integrations.litellm import LiteLLMIntegration @@ -151,10 +164,10 @@ - Also emit spans for MCP tool calls done by the LLM (#4875) by @constantinius - Option to not trace HTTP requests based on status codes (#4869) by @alexander-alderman-webb You can now disable transactions for incoming requests with specific HTTP status codes. The [new `trace_ignore_status_codes` option](https://docs.sentry.io/platforms/python/configuration/options/#trace_ignore_status_codes) accepts a `set` of status codes as integers. If a transaction wraps a request that results in one of the provided status codes, the transaction will be unsampled. - + ```python import sentry_sdk - + sentry_sdk.init( trace_ignore_status_codes={301, 302, 303, *range(305, 400), 404}, ) From 3e86962e267e86f51e8483c77dc1ad116f316d1c Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Wed, 29 Oct 2025 14:03:38 +0100 Subject: [PATCH 585/591] ci: Run integration tests on Python 3.14t (#4995) Modifies `populate_tox.py` to add free-threading Python to the test suite for integrations where we can expect the package to support free-threading. Packages are considered compatible with free-threading if the package itself, and all its dependencies, support free-threading. Pure Python packages are considered compatible if their dependencies support free-threading. The support for free-threading of packages with extension modules is determined based on the ABI tag in the wheel name. A dry-run of `pip install` determines the ABI tags of dependencies. Relevant output from the command is cached analogously to responses of PyPI REST calls. Closes https://github.com/getsentry/sentry-python/issues/4971 --- .github/workflows/test-integrations-ai.yml | 2 +- .github/workflows/test-integrations-cloud.yml | 2 +- .github/workflows/test-integrations-flags.yml | 2 +- .../workflows/test-integrations-graphql.yml | 2 +- .../workflows/test-integrations-network.yml | 2 +- .github/workflows/test-integrations-web-1.yml | 2 +- .github/workflows/update-tox.yml | 2 +- .../populate_tox/package_dependencies.jsonl | 18 + scripts/populate_tox/populate_tox.py | 304 +++++++++++- scripts/populate_tox/releases.jsonl | 453 +++++++++--------- .../launchdarkly/test_launchdarkly.py | 8 +- tox.ini | 82 ++-- 12 files changed, 586 insertions(+), 293 deletions(-) create mode 100644 scripts/populate_tox/package_dependencies.jsonl diff --git a/.github/workflows/test-integrations-ai.yml b/.github/workflows/test-integrations-ai.yml index b5a09634db..e0a4950824 100644 --- a/.github/workflows/test-integrations-ai.yml +++ b/.github/workflows/test-integrations-ai.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-cloud.yml b/.github/workflows/test-integrations-cloud.yml index 67d5da7a55..00323b44e8 100644 --- a/.github/workflows/test-integrations-cloud.yml +++ b/.github/workflows/test-integrations-cloud.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-flags.yml b/.github/workflows/test-integrations-flags.yml index 4b3b6278ef..0f232aeb29 100644 --- a/.github/workflows/test-integrations-flags.yml +++ b/.github/workflows/test-integrations-flags.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.7","3.8","3.9","3.12","3.13","3.14"] + python-version: ["3.7","3.8","3.9","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-graphql.yml b/.github/workflows/test-integrations-graphql.yml index d11bbf4aaa..9504e84a6f 100644 --- a/.github/workflows/test-integrations-graphql.yml +++ b/.github/workflows/test-integrations-graphql.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-network.yml b/.github/workflows/test-integrations-network.yml index 2e5d0024d2..170f407b31 100644 --- a/.github/workflows/test-integrations-network.yml +++ b/.github/workflows/test-integrations-network.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/test-integrations-web-1.yml b/.github/workflows/test-integrations-web-1.yml index f0d79d7fa8..9bbc57d079 100644 --- a/.github/workflows/test-integrations-web-1.yml +++ b/.github/workflows/test-integrations-web-1.yml @@ -29,7 +29,7 @@ jobs: strategy: fail-fast: false matrix: - python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14"] + python-version: ["3.6","3.7","3.8","3.9","3.10","3.11","3.12","3.13","3.14","3.14t"] # python3.6 reached EOL and is no longer being supported on # new versions of hosted runners on Github Actions # ubuntu-20.04 is the last version that supported python3.6 diff --git a/.github/workflows/update-tox.yml b/.github/workflows/update-tox.yml index 0d2640bb47..ebb62a44c2 100644 --- a/.github/workflows/update-tox.yml +++ b/.github/workflows/update-tox.yml @@ -20,7 +20,7 @@ jobs: - name: Setup Python uses: actions/setup-python@v6 with: - python-version: 3.13 + python-version: 3.14t - name: Checkout repo uses: actions/checkout@v5.0.0 diff --git a/scripts/populate_tox/package_dependencies.jsonl b/scripts/populate_tox/package_dependencies.jsonl new file mode 100644 index 0000000000..8217222282 --- /dev/null +++ b/scripts/populate_tox/package_dependencies.jsonl @@ -0,0 +1,18 @@ +{"name": "boto3", "version": "1.40.61", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/61/24/3bf865b07d15fea85b63504856e137029b6acbc73762496064219cdb265d/boto3-1.40.61-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/38/c5/f6ce561004db45f0b847c2cd9b19c67c6bf348a82018a48cb718be6b58b0/botocore-1.40.61-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/f0/ae7ca09223a81a1d890b2557186ea015f6e0502e9b8cb8e1813f1d8cfa4e/s3transfer-0.14.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}]} +{"name": "django", "version": "6.0b1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/0e/1a/306fda7e62e27ccbcb92d97f67f1094352a9f22c62f3c2b238fa50eb82d7/django-6.0b1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/17/9c/fc2331f538fbf7eedba64b2052e99ccf9ba9d6888e2f41441ee28847004b/asgiref-3.10.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a9/5c/bfd6bd0bf979426d405cc6e71eceb8701b148b16c21d2dc3c261efc61c7b/sqlparse-0.5.3-py3-none-any.whl"}}]} +{"name": "fastapi", "version": "0.120.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/7e/bb/1a74dbe87e9a595bf63052c886dfef965dc5b91d149456a8301eb3d41ce2/fastapi-0.120.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/51/da/545b75d420bb23b5d494b0517757b351963e974e79933f01e05c929f20a6/starlette-0.49.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/02/b7/cf592cb5de5cb3bade3357f8d2cf42bf103bbe39f459824b4939fd212911/annotated_doc-0.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "flask", "version": "2.3.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/fd/56/26f0be8adc2b4257df20c1c4260ddd0aa396cf8e75d90ab2f7ff99bc34f9/flask-2.3.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} +{"name": "flask", "version": "3.1.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} +{"name": "google-genai", "version": "1.46.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/db/79/8993ec6cbf56e5c8f88c165380e55de34ec74f7b928bc302ff5c370f9c4e/google_genai-1.46.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/87/24/ec82aee6ba1a076288818fe5cc5125f4d93fffdc68bb7b381c68286c8aaa/google_auth-2.42.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/96/c5/1e741d26306c42e2bf6ab740b2202872727e0f606033c9dd713f8b93f5a8/cachetools-6.2.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/8d/0133e4eb4beed9e425d9a98ed6e081a55d195481b7632472be1af08d2f6b/rsa-4.9.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fa/a8/5b41e0da817d64113292ab1f8247140aac61cbf6cfd085d6a0fa77f4984f/websockets-15.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c8/f1/d6a797abb14f6283c0ddff96bbdd46937f64122b8c925cab503dd37f8214/pyasn1-0.6.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/47/8d/d529b5d697919ba8c11ad626e835d4039be708a35b0d22de83a269a6682c/pyasn1_modules-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "langchain", "version": "1.0.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/8c/06/0e03587da37173c29a58bf17312793c2453df9ca2912e9adfe869c120437/langchain-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c0/54/3cdbe9d151d06cd689b5aa937ac11403b64bbfe76486fda6431a24061721/langchain_core-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b1/3c/acc0956a0da96b25a2c5c1a85168eacf1253639a04ed391d7a7bcaae5d6c/langgraph-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/68/47/9ffd10882403020ea866e381de7f8e504a78f606a914af7f8244456c7783/langgraph_prebuilt-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b4/2b/7e0248f65e35800ea8e4e3dbb3bcc36c61b81f5b8abeddaceec8320ab491/langsmith-0.4.38-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "langgraph", "version": "0.6.11", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/df/94/430f0341c5c2fe3e3b9f5ab2622f35e2bda12c4a7d655c519468e853d1b0/langgraph-0.6.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8e/d1/e4727f4822943befc3b7046f79049b1086c9493a34b4d44a1adf78577693/langgraph_prebuilt-0.6.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c0/54/3cdbe9d151d06cd689b5aa937ac11403b64bbfe76486fda6431a24061721/langchain_core-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b4/2b/7e0248f65e35800ea8e4e3dbb3bcc36c61b81f5b8abeddaceec8320ab491/langsmith-0.4.38-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "launchdarkly-server-sdk", "version": "9.12.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/d8/63/30b2a1ec9994d04d82355663022734919903b54c7497b302206f67316237/launchdarkly_server_sdk-9.12.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/95/c20ba9cd28951f360b222da09138daf9169ad951be2a0d500eb0f0c84614/launchdarkly_eventsource-1.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/cb/84/a04c59324445f4bcc98dc05b39a1cd07c242dde643c1a3c21e4f7beaf2f2/expiringdict-1.2.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/34/90/0200184d2124484f918054751ef997ed6409cb05b7e8dcbf5a22da4c4748/pyrfc3339-2.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl"}}]} +{"name": "openfeature-sdk", "version": "0.7.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9b/20/cb043f54b11505d993e4dd84652cfc44c1260dc94b7f41aa35489af58277/openfeature_sdk-0.7.5-py3-none-any.whl"}}]} +{"name": "openfeature-sdk", "version": "0.8.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/f0/f5/707a5b144115de1a49bf5761a63af2545fef0a1824f72db39ddea0a3438f/openfeature_sdk-0.8.3-py3-none-any.whl"}}]} +{"name": "quart", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/7e/e9/cc28f21f52913adf333f653b9e0a3bf9cb223f5083a26422968ba73edd8d/quart-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/3b/dfa13a8d96aa24e40ea74a975a9906cfdc2ab2f4e3b498862a57052f04eb/hypercorn-0.17.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/69/b2/119f6e6dcbd96f9069ce9a2665e0146588dc9f88f29549711853645e736a/h2-4.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/c6/80c95b1b2b94682a72cbdbfb85b81ae2daffa4291fbfa1b1464502ede10d/hpack-4.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/30/47d0bf6072f7252e6521f3447ccfa40b421b6824517f82854703d0f5a98b/hyperframe-6.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/58/e860788190eba3bcce367f74d29c4675466ce8dddfba85f7827588416f01/wsproto-1.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/bc/8a/340a1555ae33d7354dbca4faa54948d76d89a27ceef032c8c3bc661d003e/aiofiles-25.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/5e/5f/82c8074f7e84978129347c2c6ec8b6c59f3584ff1a20bc3c940a3e061790/priority-2.0.0-py3-none-any.whl"}}]} +{"name": "requests", "version": "2.32.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}]} +{"name": "starlette", "version": "0.49.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/51/da/545b75d420bb23b5d494b0517757b351963e974e79933f01e05c929f20a6/starlette-0.49.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "statsig", "version": "0.55.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9d/17/de62fdea8aab8aa7c4a833378e0e39054b728dfd45ef279e975ed5ef4e86/statsig-0.55.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} +{"name": "statsig", "version": "0.66.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/dd/a3/d448734d1921c08b1b00fd69ef6e383b1a747ce9a32fc8e65aa3700ccb8c/statsig-0.66.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2f/c2/f9e977608bdf958650638c3f1e28f85a1b075f075ebbe77db8555463787b/Brotli-1.1.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} +{"name": "strawberry-graphql", "version": "0.284.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/eb/9d/03a4225721f82dc8d25de894ae0b29ede0c5710de88a1710a8c081c88a4c/strawberry_graphql-0.284.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ae/4f/7297663840621022bc73c22d7d9d80dbc78b4db6297f764b545cd5dd462d/graphql_core-3.2.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/00/f2/c68a97c727c795119f1056ad2b7e716c23f26f004292517c435accf90b5c/lia_web-0.2.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} +{"name": "typer", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/78/64/7713ffe4b5983314e9d436a90d5bd4f63b6054e2aca783a3cfc44cb95bbf/typer-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/25/7a/b0178788f8dc6cafce37a212c99565fa1fe7872c70c6c9c1e1a372d9d88f/rich-14.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/54/e7d793b573f298e1c9013b8c4dade17d481164aa517d1d7148619c2cedbf/markdown_it_py-4.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 453823f39d..0700a82637 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -11,6 +11,7 @@ import subprocess import sys import time +from dataclasses import dataclass from bisect import bisect_left from collections import defaultdict from datetime import datetime, timedelta, timezone # noqa: F401 @@ -39,6 +40,7 @@ TOX_FILE = Path(__file__).resolve().parent.parent.parent / "tox.ini" RELEASES_CACHE_FILE = Path(__file__).resolve().parent / "releases.jsonl" +DEPENDENCIES_CACHE_FILE = Path(__file__).resolve().parent / "package_dependencies.jsonl" ENV = Environment( loader=FileSystemLoader(Path(__file__).resolve().parent), trim_blocks=True, @@ -52,6 +54,7 @@ CLASSIFIER_PREFIX = "Programming Language :: Python :: " CACHE = defaultdict(dict) +DEPENDENCIES_CACHE = defaultdict(dict) IGNORE = { # Do not try auto-generating the tox entries for these. They will be @@ -67,6 +70,26 @@ "potel", } +# Free-threading is experimentally supported in 3.13, and officially supported in 3.14. +MIN_FREE_THREADING_SUPPORT = Version("3.14") + + +@dataclass(order=True) +class ThreadedVersion: + version: Version + no_gil: bool + + def __init__(self, version: str | Version, no_gil=False): + self.version = Version(version) if isinstance(version, str) else version + self.no_gil = no_gil + + def __str__(self): + version = f"py{self.version.major}.{self.version.minor}" + if self.no_gil: + version += "t" + + return version + def _fetch_sdk_metadata() -> PackageMetadata: (dist,) = distributions( @@ -112,6 +135,38 @@ def fetch_release(package: str, version: Version) -> Optional[dict]: return release +@functools.cache +def fetch_package_dependencies(package: str, version: Version) -> dict: + """Fetch package dependencies metadata from cache or, failing that, PyPI.""" + package_dependencies = _fetch_package_dependencies_from_cache(package, version) + if package_dependencies is not None: + return package_dependencies + + # Removing non-report output with -qqq may be brittle, but avoids file I/O. + # Currently -qqq supresses all non-report output that would break json.loads(). + pip_report = subprocess.run( + [ + sys.executable, + "-m", + "pip", + "install", + f"{package}=={str(version)}", + "--dry-run", + "--ignore-installed", + "--report", + "-", + "-qqq", + ], + capture_output=True, + text=True, + ).stdout.strip() + + dependencies_info = json.loads(pip_report)["install"] + _save_to_package_dependencies_cache(package, version, dependencies_info) + + return dependencies_info + + def _fetch_from_cache(package: str, version: Version) -> Optional[dict]: package = _normalize_name(package) if package in CACHE and str(version) in CACHE[package]: @@ -121,6 +176,17 @@ def _fetch_from_cache(package: str, version: Version) -> Optional[dict]: return None +def _fetch_package_dependencies_from_cache( + package: str, version: Version +) -> Optional[dict]: + package = _normalize_name(package) + if package in DEPENDENCIES_CACHE and str(version) in DEPENDENCIES_CACHE[package]: + DEPENDENCIES_CACHE[package][str(version)]["_accessed"] = True + return DEPENDENCIES_CACHE[package][str(version)]["dependencies"] + + return None + + def _save_to_cache(package: str, version: Version, release: Optional[dict]) -> None: with open(RELEASES_CACHE_FILE, "a") as releases_cache: releases_cache.write(json.dumps(_normalize_release(release)) + "\n") @@ -129,6 +195,23 @@ def _save_to_cache(package: str, version: Version, release: Optional[dict]) -> N CACHE[_normalize_name(package)][str(version)]["_accessed"] = True +def _save_to_package_dependencies_cache( + package: str, version: Version, release: Optional[dict] +) -> None: + with open(DEPENDENCIES_CACHE_FILE, "a") as releases_cache: + line = { + "name": package, + "version": str(version), + "dependencies": _normalize_package_dependencies(release), + } + releases_cache.write(json.dumps(line) + "\n") + + DEPENDENCIES_CACHE[_normalize_name(package)][str(version)] = { + "info": release, + "_accessed": True, + } + + def _prefilter_releases( integration: str, releases: dict[str, dict], @@ -404,12 +487,19 @@ def supported_python_versions( return supported -def pick_python_versions_to_test(python_versions: list[Version]) -> list[Version]: +def pick_python_versions_to_test( + python_versions: list[Version], + python_versions_with_supported_free_threaded_wheel: set[Version], +) -> list[ThreadedVersion]: """ Given a list of Python versions, pick those that make sense to test on. Currently, this is the oldest, the newest, and the second newest Python version. + + A free-threaded variant is also chosen for the newest Python version for which + - a free-threaded wheel is distributed; and + - the SDK supports free-threading. """ filtered_python_versions = { python_versions[0], @@ -421,7 +511,21 @@ def pick_python_versions_to_test(python_versions: list[Version]) -> list[Version except IndexError: pass - return sorted(filtered_python_versions) + versions_to_test = sorted( + ThreadedVersion(version) for version in filtered_python_versions + ) + + for python_version in reversed(python_versions): + if python_version < MIN_FREE_THREADING_SUPPORT: + break + + if python_version in python_versions_with_supported_free_threaded_wheel: + versions_to_test.append( + ThreadedVersion(versions_to_test[-1].version, no_gil=True) + ) + break + + return versions_to_test def _parse_python_versions_from_classifiers(classifiers: list[str]) -> list[Version]: @@ -475,12 +579,103 @@ def determine_python_versions(pypi_data: dict) -> Union[SpecifierSet, list[Versi return [] -def _render_python_versions(python_versions: list[Version]) -> str: - return ( - "{" - + ",".join(f"py{version.major}.{version.minor}" for version in python_versions) - + "}" - ) +def _get_abi_tag(wheel_filename: str) -> str: + return wheel_filename.removesuffix(".whl").split("-")[-2] + + +def _get_abi_tag_version(python_version: Version): + return f"{python_version.major}{python_version.minor}" + + +@functools.cache +def _has_free_threading_dependencies( + package_name: str, release: Version, python_version: Version +) -> bool: + """ + Checks if all dependencies of a version of a package support free-threading. + + A dependency supports free-threading if + - the dependency is pure Python, indicated by a "none" abi tag in its wheel name; or + - the abi tag of one of its wheels has a "t" suffix to indicate a free-threaded build; or + - no wheel targets the platform on which the script is run, but PyPI distributes a wheel + satisfying one of the above conditions. + """ + dependencies_info = fetch_package_dependencies(package_name, release) + + for dependency_info in dependencies_info: + wheel_filename = dependency_info["download_info"]["url"].split("/")[-1] + + if wheel_filename.endswith(".tar.gz"): + package_release = wheel_filename.rstrip(".tar.gz") + dependency_name, dependency_version = package_release.split("-") + + pypi_data = fetch_release(dependency_name, Version(dependency_version)) + supports_free_threading = False + + for download in pypi_data["urls"]: + abi_tag = _get_abi_tag(download["filename"]) + abi_tag_version = _get_abi_tag_version(python_version) + + if download["packagetype"] == "bdist_wheel" and ( + ( + abi_tag.endswith("t") + and abi_tag.startswith(f"cp{abi_tag_version}") + ) + or abi_tag == "none" + ): + supports_free_threading = True + + if not supports_free_threading: + return False + + elif wheel_filename.endswith(".whl"): + abi_tag = _get_abi_tag(wheel_filename) + if abi_tag != "none" and not abi_tag.endswith("t"): + return False + + else: + raise Exception( + f"Wheel filename with unhandled extension: {wheel_filename}" + ) + + return True + + +def _supports_free_threading( + package_name: str, release: Version, python_version: Version, pypi_data: dict +) -> bool: + """ + Check if the package version supports free-threading on the given Python minor + version. + + There are two cases in which we assume a package has free-threading + support: + - The package is pure Python, indicated by a "none" abi tag in its wheel name, + and has dependencies supporting free-threading; or + - the abi tag of one of its wheels has a "t" suffix to indicate a free-threaded build. + + See https://peps.python.org/pep-0427/#file-name-convention + """ + for download in pypi_data["urls"]: + if download["packagetype"] == "bdist_wheel": + abi_tag = _get_abi_tag(download["filename"]) + + abi_tag_version = _get_abi_tag_version(python_version) + if ( + abi_tag.endswith("t") and abi_tag.startswith(f"cp{abi_tag_version}") + ) or ( + abi_tag == "none" + and _has_free_threading_dependencies( + package_name, release, python_version + ) + ): + return True + + return False + + +def _render_python_versions(python_versions: list[ThreadedVersion]) -> str: + return "{" + ",".join(str(version) for version in python_versions) + "}" def _render_dependencies(integration: str, releases: list[Version]) -> list[str]: @@ -585,12 +780,22 @@ def _add_python_versions_to_release( TEST_SUITE_CONFIG[integration].get("python") ) + supported_py_versions = supported_python_versions( + determine_python_versions(release_pypi_data), + target_python_versions, + release, + ) + + py_versions_with_supported_free_threaded_wheel = set( + version + for version in supported_py_versions + if version >= MIN_FREE_THREADING_SUPPORT + and _supports_free_threading(package, release, version, release_pypi_data) + ) + release.python_versions = pick_python_versions_to_test( - supported_python_versions( - determine_python_versions(release_pypi_data), - target_python_versions, - release, - ) + supported_py_versions, + py_versions_with_supported_free_threaded_wheel, ) release.rendered_python_versions = _render_python_versions(release.python_versions) @@ -647,8 +852,16 @@ def _normalize_name(package: str) -> str: return package.lower().replace("-", "_") +def _extract_wheel_info_to_cache(wheel: dict): + return { + "packagetype": wheel["packagetype"], + "filename": wheel["filename"], + } + + def _normalize_release(release: dict) -> dict: """Filter out unneeded parts of the release JSON.""" + urls = [_extract_wheel_info_to_cache(wheel) for wheel in release["urls"]] normalized = { "info": { "classifiers": release["info"]["classifiers"], @@ -657,15 +870,36 @@ def _normalize_release(release: dict) -> dict: "version": release["info"]["version"], "yanked": release["info"]["yanked"], }, + "urls": urls, } return normalized +def _normalize_package_dependencies(package_dependencies: list[dict]) -> list[dict]: + """Filter out unneeded parts of the package dependencies JSON.""" + normalized = [ + { + "download_info": {"url": depedency["download_info"]["url"]}, + } + for depedency in package_dependencies + ] + + return normalized + + +def _exit_if_not_free_threaded_interpreter(): + if "free-threading build" not in sys.version: + raise Exception("Running with a free-threaded interpreter is required.") + + def main() -> dict[str, list]: """ Generate tox.ini from the tox.jinja template. """ global MIN_PYTHON_VERSION, MAX_PYTHON_VERSION + + _exit_if_not_free_threaded_interpreter() + meta = _fetch_sdk_metadata() sdk_python_versions = _parse_python_versions_from_classifiers( meta.get_all("Classifier") @@ -676,6 +910,20 @@ def main() -> dict[str, list]: f"The SDK supports Python versions {MIN_PYTHON_VERSION} - {MAX_PYTHON_VERSION}." ) + if not RELEASES_CACHE_FILE.exists(): + print( + f"Creating {RELEASES_CACHE_FILE.name}." + "You should only see this message if you cleared the cache by removing the file." + ) + RELEASES_CACHE_FILE.write_text("") + + if not DEPENDENCIES_CACHE_FILE.exists(): + print( + f"Creating {DEPENDENCIES_CACHE_FILE.name}." + "You should only see this message if you cleared the cache by removing the file." + ) + DEPENDENCIES_CACHE_FILE.write_text("") + # Load file cache global CACHE @@ -689,6 +937,19 @@ def main() -> dict[str, list]: False # for cleaning up unused cache entries ) + # Load package dependencies cache + global DEPENDENCIES_CACHE + + with open(DEPENDENCIES_CACHE_FILE) as dependencies_cache: + for line in dependencies_cache: + release = json.loads(line) + name = _normalize_name(release["name"]) + version = release["version"] + DEPENDENCIES_CACHE[name][version] = { + "dependencies": release["dependencies"], + "_accessed": False, + } + # Process packages packages = defaultdict(list) @@ -764,9 +1025,24 @@ def main() -> dict[str, list]: ): releases_cache.write(json.dumps(release) + "\n") + # Sort the dependencies file + releases = [] + with open(DEPENDENCIES_CACHE_FILE) as releases_cache: + releases = [json.loads(line) for line in releases_cache] + releases.sort(key=lambda r: (r["name"], r["version"])) + with open(DEPENDENCIES_CACHE_FILE, "w") as releases_cache: + for release in releases: + if ( + DEPENDENCIES_CACHE[_normalize_name(release["name"])][ + release["version"] + ]["_accessed"] + is True + ): + releases_cache.write(json.dumps(release) + "\n") + print( "Done generating tox.ini. Make sure to also update the CI YAML " - "files to reflect the new test targets." + "files by executing split_tox_gh_actions.py." ) return packages diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index c639a53d1f..fd475ed1ef 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -1,227 +1,226 @@ -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.10.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.11.29", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.8.19", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.4", "version": "2.0.13", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.5", "version": "2.2.28", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.1.14", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.2.25", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.25", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.7", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0b1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Flask", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "1.1.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Flask", "requires_python": ">=3.9", "version": "3.1.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.16.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Quart", "requires_python": ">=3.9", "version": "0.20.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "", "version": "1.2.19", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "1.3.24", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7", "version": "1.4.54", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.44", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.13.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.34.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.52.2", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.71.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.15.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.19.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.40.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.68.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.69.0rc3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.8", "version": "0.26.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.5.0", "version": "0.23.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.6.0", "version": "0.25.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.7.0", "version": "0.27.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.8.0", "version": "0.30.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.59", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.9", "version": "5.6.0b2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.10.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.15.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.20.0", "yanked": false}} -{"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": "", "version": "1.4.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.107.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.93.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.46.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": "", "version": "3.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": null, "version": "3.4.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "grpcio", "requires_python": "", "version": "1.32.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.76.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.7", "version": "0.24.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.26.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.28.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": null, "version": "0.1.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.10.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.7.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "2.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.4", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.28.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.32.6", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.36.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.0rc7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}} -{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}} -{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0.0,>=3.10.0", "version": "1.0.2", "yanked": false}} -{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.11", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.7", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.79.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.15.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.16.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.17.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.19.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.59.9", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.88.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.4.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.1", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.2", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.2.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.6.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "3.2.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.5.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.6.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.3", "yanked": false}} -{"info": {"classifiers": ["Framework :: Pylons", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": null, "version": "1.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.10.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.6.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.7.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.8.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.9.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=3.6", "version": "2.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.3.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.4.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "3.0.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}} -{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.50.1", "yanked": false}} -{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3"], "name": "redis", "requires_python": null, "version": "2.9.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.2.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "3.5.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}} -{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}} -{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis-py-cluster", "requires_python": "", "version": "1.3.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4", "version": "2.1.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3"], "name": "requests", "requires_python": null, "version": "2.0.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.10.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.11.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.12.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.16.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development :: Libraries"], "name": "requests", "requires_python": ">=3.9", "version": "2.32.5", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5"], "name": "requests", "requires_python": null, "version": "2.8.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.10.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.13.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.6.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.7.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.8.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=2.7", "version": "1.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.7", "version": "1.16.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.5", "version": "1.8.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.9", "version": "2.6.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "sanic", "requires_python": "", "version": "0.8.3", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.6", "version": "20.12.7", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "23.12.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "25.3.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.6", "version": "0.16.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.7", "version": "0.27.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.8", "version": "0.38.6", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.9", "version": "0.48.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.66.0", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.10", "version": "0.284.1", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}} -{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.4.27", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.6.22", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.8.18", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.4", "version": "5.0.63", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.5", "version": "5.4.20", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.9", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.20.0", "yanked": false}} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX :: Linux", "Programming Language :: C", "Programming Language :: C++", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Unix Shell", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Archiving", "Topic :: System :: Archiving :: Compression", "Topic :: Text Processing :: Fonts", "Topic :: Utilities"], "name": "Brotli", "requires_python": "", "version": "1.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "Brotli-1.1.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "Brotli-1.1.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.10.8", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-1.10.8-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-1.10.8.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.11.29", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-1.11.29-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-1.11.29.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": "", "version": "1.8.19", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-1.8.19-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-1.8.19.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.4", "version": "2.0.13", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-2.0.13-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-2.0.13.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.5", "version": "2.2.28", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-2.2.28-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-2.2.28.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.1.14", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-3.1.14-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-3.1.14.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.6", "version": "3.2.25", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Django-3.2.25-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Django-3.2.25.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.8", "version": "4.2.25", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "django-4.2.25-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "django-4.2.25.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.10", "version": "5.2.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "django-5.2.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "django-5.2.7.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Django", "requires_python": ">=3.12", "version": "6.0b1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "django-6.0b1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "django-6.0b1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Flask", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "1.1.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Flask-1.1.4-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Flask-1.1.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "Flask", "requires_python": ">=3.8", "version": "2.3.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "flask-2.3.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "flask-2.3.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Flask", "requires_python": ">=3.9", "version": "3.1.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "flask-3.1.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "flask-3.1.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "Quart", "requires_python": ">=3.7", "version": "0.16.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "Quart-0.16.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "Quart-0.16.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "Quart", "requires_python": ">=3.9", "version": "0.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "quart-0.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "quart-0.20.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "", "version": "1.2.19", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "SQLAlchemy-1.2.19.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "1.3.24", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp27-cp27m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.3.24-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "SQLAlchemy-1.3.24.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7", "version": "1.4.54", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-macosx_12_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-manylinux1_x86_64.manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_5_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp36-cp36m-manylinux1_x86_64.manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_5_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-manylinux1_x86_64.manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_5_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-macosx_12_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-manylinux1_x86_64.manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_5_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-macosx_12_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-manylinux1_x86_64.manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_5_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-1.4.54-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "sqlalchemy-1.4.54.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database :: Front-Ends"], "name": "SQLAlchemy", "requires_python": ">=3.7", "version": "2.0.44", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "SQLAlchemy-2.0.44-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "sqlalchemy-2.0.44-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "sqlalchemy-2.0.44.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "UnleashClient-6.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "unleashclient-6.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Typing :: Typed"], "name": "UnleashClient", "requires_python": ">=3.8", "version": "6.3.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "unleashclient-6.3.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "unleashclient-6.3.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.8", "version": "3.10.11", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.10.11-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "aiohttp-3.10.11.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.9", "version": "3.13.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-macosx_10_13_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314t-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_riscv64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.13.2-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "aiohttp-3.13.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.5.3", "version": "3.4.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-macosx_10_11_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-macosx_10_11_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-macosx_10_11_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.4.4-cp37-cp37m-win_amd64.whl"}, {"packagetype": "sdist", "filename": "aiohttp-3.4.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "aiohttp", "requires_python": ">=3.6", "version": "3.7.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "aiohttp-3.7.4-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "aiohttp-3.7.4.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "anthropic-0.16.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "anthropic-0.16.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.7", "version": "0.35.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "anthropic-0.35.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "anthropic-0.35.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.54.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "anthropic-0.54.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "anthropic-0.54.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "anthropic", "requires_python": ">=3.8", "version": "0.72.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "anthropic-0.72.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "anthropic-0.72.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.12.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp27-cp27m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp35-cp35m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp36-cp36m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp37-cp37m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.12.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.12.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.13.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp27-cp27m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp35-cp35m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp36-cp36m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp37-cp37m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.13.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.13.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.14.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp27-cp27m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp35-cp35m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp36-cp36m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp37-cp37m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.14.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.14.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.15.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp27-cp27m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp35-cp35m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp36-cp36m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp37-cp37m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.15.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.15.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*", "version": "2.19.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp27-cp27m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp35-cp35m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp36-cp36m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp37-cp37m-macosx_10_6_intel.macosx_10_9_intel.macosx_10_9_x86_64.macosx_10_10_intel.macosx_10_10_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.19.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.19.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.6", "version": "2.26.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.26.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.26.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.7", "version": "2.41.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.41.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "apache-beam-2.41.0.zip"}]} +{"info": {"classifiers": ["Intended Audience :: End Users/Desktop", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "apache-beam", "requires_python": ">=3.9", "version": "2.69.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "apache_beam-2.69.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "apache_beam-2.69.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": "", "version": "0.20.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ariadne-0.20.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "ariadne-0.20.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "ariadne", "requires_python": ">=3.9", "version": "0.26.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ariadne-0.26.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "ariadne-0.26.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.6", "version": "0.23", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "arq-0.23-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "arq-0.23.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Framework :: AsyncIO", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Clustering", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "arq", "requires_python": ">=3.8", "version": "0.26.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "arq-0.26.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "arq-0.26.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.5.0", "version": "0.23.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp35-cp35m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp37-cp37m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp38-cp38-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp39-cp39-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.23.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "asyncpg-0.23.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.6.0", "version": "0.25.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp36-cp36m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp36-cp36m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.25.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "asyncpg-0.25.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.7.0", "version": "0.27.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp37-cp37m-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.27.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "asyncpg-0.27.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Database :: Front-Ends"], "name": "asyncpg", "requires_python": ">=3.8.0", "version": "0.30.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "asyncpg-0.30.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "asyncpg-0.30.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.12.49-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.12.49.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.20.54-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.20.54.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.28.85-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.28.85.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.61", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.40.61-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.40.61.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.12.25-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.12.25.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.13.4-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.13.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-4.4.7-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-4.4.7.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-5.5.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-5.5.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.9", "version": "5.6.0b2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-5.6.0b2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-5.6.0b2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "chalice-1.16.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "chalice-1.16.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "chalice-1.32.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "chalice-1.32.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-win_amd64.whl"}, {"packagetype": "sdist", "filename": "clickhouse-driver-0.2.9.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.10.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "cohere-5.10.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "cohere-5.10.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.15.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "cohere-5.15.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "cohere-5.15.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "cohere", "requires_python": "<4.0,>=3.9", "version": "5.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "cohere-5.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "cohere-5.20.0.tar.gz"}]} +{"info": {"classifiers": ["Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "cohere", "requires_python": "<4.0,>=3.8", "version": "5.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "cohere-5.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "cohere-5.4.0.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.9", "version": "1.18.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "dramatiq-1.18.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "dramatiq-1.18.0.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: System :: Distributed Computing"], "name": "dramatiq", "requires_python": ">=3.5", "version": "1.9.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "dramatiq-1.9.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "dramatiq-1.9.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": "", "version": "1.4.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-1.4.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "falcon-1.4.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp34-cp34m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp34-cp34m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-2.0.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "falcon-2.0.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "falcon-3.1.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "falcon-4.1.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.107.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.107.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.107.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.120.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.120.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.79.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.79.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.93.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.93.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.93.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.29.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.29.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.35.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.35.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.41.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.41.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.46.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.46.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.46.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-3.4.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-3.4.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.0.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.0.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.2.0b0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.2.0b0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": "", "version": "3.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "graphene-3.3-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "graphene-3.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries"], "name": "graphene", "requires_python": null, "version": "3.4.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "graphene-3.4.3-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "graphene-3.4.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "grpcio", "requires_python": "", "version": "1.32.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27mu-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27mu-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27mu-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-macosx_10_7_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-manylinux2010_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-manylinux2010_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.32.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "sdist", "filename": "grpcio-1.32.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.6", "version": "1.47.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-macosx_12_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.47.5-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "grpcio-1.47.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.7", "version": "1.62.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-macosx_12_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-macosx_10_10_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.62.3-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "grpcio-1.62.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "grpcio", "requires_python": ">=3.9", "version": "1.76.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-linux_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-macosx_11_0_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "grpcio-1.76.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "grpcio-1.76.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.16.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.16.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.16.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.20.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.6", "version": "0.22.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.22.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.22.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.7", "version": "0.24.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.24.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.24.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.26.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.26.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.26.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Trio", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "httpx", "requires_python": ">=3.8", "version": "0.28.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "httpx-0.28.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "httpx-0.28.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": null, "version": "0.1.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "huey-0.1.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.10.5", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "huey-1.10.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "1.7.0", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "huey-1.7.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "huey", "requires_python": "", "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "huey-2.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "huey", "requires_python": "", "version": "2.1.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "huey-2.1.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "huey", "requires_python": null, "version": "2.5.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "huey-2.5.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "huey-2.5.4.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.24.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "huggingface_hub-0.24.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "huggingface_hub-0.24.7.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.8.0", "version": "0.36.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "huggingface_hub-0.36.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "huggingface_hub-0.36.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "huggingface_hub-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "huggingface_hub-1.0.1.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-0.1.20-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-0.1.20.tar.gz"}]} +{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-0.3.27-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-0.3.27.tar.gz"}]} +{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0.0,>=3.10.0", "version": "1.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-1.0.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-1.0.2.tar.gz"}]} +{"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.11", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langgraph-0.6.11-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langgraph-0.6.11.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langgraph-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langgraph-1.0.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "launchdarkly_server_sdk-9.12.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "launchdarkly_server_sdk-9.12.2.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "launchdarkly_server_sdk-9.8.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "launchdarkly_server_sdk-9.8.1.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.77.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.77.7.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.78.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.78.7.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.79.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.79.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.79.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.12.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.12.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.18.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.18.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.6.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.6.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "loguru-0.7.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "loguru-0.7.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.15.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.15.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.15.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.16.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.16.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.17.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.17.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.17.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.19.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.19.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.19.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.0.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.103.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.103.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.109.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.109.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.59.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.59.9-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.59.9.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.88.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.88.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.88.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.0.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.4.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.6.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.6.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.0.19-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.0.19.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.1.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.2.11-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.2.11.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.4.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.4.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.4.2.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.8", "version": "0.7.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openfeature_sdk-0.7.5-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openfeature_sdk-0.7.5.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3"], "name": "openfeature-sdk", "requires_python": ">=3.9", "version": "0.8.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openfeature_sdk-0.8.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openfeature_sdk-0.8.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX :: Linux", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.15", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Rust", "Typing :: Typed"], "name": "orjson", "requires_python": ">=3.9", "version": "3.11.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-macosx_15_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp311-cp311-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-macosx_15_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp312-cp312-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-macosx_15_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp313-cp313-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-macosx_15_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp314-cp314-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-musllinux_1_2_armv7l.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "orjson-3.11.4-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "orjson-3.11.4.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pure_eval-0.0.3.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pure_eval-0.2.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pure_eval-0.2.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.0.18-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.0.18.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.2.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.2.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.2.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.4.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.7.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.7.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.7.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}, "urls": [{"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.5-macosx-10.5-i386.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.6-macosx-10.5-i386.egg"}, {"packagetype": "sdist", "filename": "pymongo-0.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-cp32m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.8-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.9-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-2.8.1.tar.gz"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.5.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.4.exe"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.13.0-py2.7-macosx-10.14-intel.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.13.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "3.2.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-macosx_10_11_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp26-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-macosx_10_11_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp27-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp32-cp32m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp32-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp32-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp33-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp33-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp34-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp34-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp35-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.2.2-cp35-none-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-macosx-10.11-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-macosx-10.11-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.2-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.2-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.2-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.4-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.5-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.2.2-py3.5-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.2.2.tar.gz"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py3.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win32-py3.5.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py3.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-3.2.2.win-amd64-py3.5.exe"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26m-macosx_10_11_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26m-macosx_10_12_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-cp26mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp26-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27m-macosx_10_11_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27m-macosx_10_12_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp27-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp33-cp33m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp33-cp33m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp33-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp33-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp34-cp34m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp34-cp34m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp34-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp34-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp35-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.4.0-cp35-none-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.6-macosx-10.11-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.6-macosx-10.12-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.7-macosx-10.11-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.7-macosx-10.12-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.4-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.5-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.4.0-py3.5-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.4.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.5.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26m-macosx_10_12_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp26-cp26m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp27-none-macosx_10_12_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp33-cp33m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp33-cp33m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp33-cp33m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp33-cp33m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp34-cp34m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp34-cp34m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp34-cp34m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp34-cp34m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.5.1-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.6-macosx-10.12-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.7-macosx-10.12-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.4-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.5-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.5-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.6-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.5.1-py3.6-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.5.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.6.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp26-cp26m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27m-macosx_10_13_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp34-cp34m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp34-cp34m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp34-cp34m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp34-cp34m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.6.1-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py2.7-macosx-10.13-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.4-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.5-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.5-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.6-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.6.1-py3.6-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.6.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.0.2-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "pymongo-4.0.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database", "Typing :: Typed"], "name": "pymongo", "requires_python": ">=3.9", "version": "4.15.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp310-cp310-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp311-cp311-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp312-cp312-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp313-cp313-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314t-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp314-cp314-win_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-4.15.3-cp39-cp39-win_arm64.whl"}, {"packagetype": "sdist", "filename": "pymongo-4.15.3.tar.gz"}]} +{"info": {"classifiers": ["Framework :: Pylons", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": null, "version": "1.0.2", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyramid-1.0.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.10.8", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-1.10.8-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-1.10.8.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.6.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-1.6.5-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-1.6.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.7.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-1.7.6-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-1.7.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": "", "version": "1.8.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-1.8.6-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-1.8.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*", "version": "1.9.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-1.9.4-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-1.9.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 6 - Mature", "Framework :: Pyramid", "Intended Audience :: Developers", "License :: Repoze Public License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: WSGI"], "name": "pyramid", "requires_python": ">=3.6", "version": "2.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pyramid-2.0.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pyramid-2.0.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.1.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-2.1.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.3.4", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-2.3.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "2.4.8", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-2.4.8.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "pyspark", "requires_python": "", "version": "3.0.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-3.0.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-3.1.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-3.5.7.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-4.0.1.tar.gz"}]} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.51.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-win_amd64.whl"}]} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-win_amd64.whl"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-0.6.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-2.10.6-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-2.10.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3"], "name": "redis", "requires_python": null, "version": "2.9.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-2.9.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-3.0.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-3.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "3.2.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-3.2.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-3.2.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "3.5.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-3.5.3-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-3.5.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.6", "version": "4.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-4.0.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-4.0.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.7", "version": "4.6.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-4.6.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-4.6.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.8", "version": "5.3.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-5.3.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-5.3.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "6.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-6.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-6.4.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "redis", "requires_python": ">=3.9", "version": "7.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-7.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-7.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4"], "name": "redis-py-cluster", "requires_python": null, "version": "0.1.0", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-py-cluster-0.1.0.tar.gz"}]} +{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.1.0", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-py-cluster-1.1.0.tar.gz"}]} +{"info": {"classifiers": [], "name": "redis-py-cluster", "requires_python": null, "version": "1.2.0", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-py-cluster-1.2.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis-py-cluster", "requires_python": "", "version": "1.3.6", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-py-cluster-1.3.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*", "version": "2.0.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis_py_cluster-2.0.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-py-cluster-2.0.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "redis-py-cluster", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4", "version": "2.1.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis_py_cluster-2.1.3-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-py-cluster-2.1.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3"], "name": "requests", "requires_python": null, "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.0.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.0.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.10.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.10.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.10.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": null, "version": "2.11.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.11.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.11.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.12.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.12.5-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.12.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "requests", "requires_python": "", "version": "2.16.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.16.5-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.16.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development :: Libraries"], "name": "requests", "requires_python": ">=3.9", "version": "2.32.5", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.32.5-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.32.5.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5"], "name": "requests", "requires_python": null, "version": "2.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "requests-2.8.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "requests-2.8.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.10.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-0.10.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-0.10.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.13.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-0.13.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-0.13.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.6.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-0.6.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-0.6.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.7.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-0.7.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-0.7.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": "", "version": "0.8.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-0.8.2-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-0.8.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=2.7", "version": "1.0", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "rq-1.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.7", "version": "1.16.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-1.16.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-1.16.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.5", "version": "1.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-1.8.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-1.8.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Operating System :: MacOS", "Operating System :: POSIX", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Distributed Computing", "Topic :: System :: Monitoring", "Topic :: System :: Systems Administration"], "name": "rq", "requires_python": ">=3.9", "version": "2.6.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "rq-2.6.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "rq-2.6.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "sanic", "requires_python": "", "version": "0.8.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "sanic-0.8.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "sanic-0.8.3.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.6", "version": "20.12.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "sanic-20.12.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "sanic-20.12.7.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "23.12.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "sanic-23.12.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "sanic-23.12.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "sanic", "requires_python": ">=3.8", "version": "25.3.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "sanic-25.3.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "sanic-25.3.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.6", "version": "0.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.16.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.16.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.7", "version": "0.27.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.27.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.27.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.8", "version": "0.38.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.38.6-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.38.6.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.9", "version": "0.49.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.49.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.49.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlite-1.48.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlite-1.48.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlite-1.51.16-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlite-1.51.16.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "statsig-0.55.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "statsig-0.55.3.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.66.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "statsig-0.66.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "statsig-0.66.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "strawberry_graphql-0.209.8-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "strawberry_graphql-0.209.8.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.10", "version": "0.284.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "strawberry_graphql-0.284.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "strawberry_graphql-0.284.1.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp38-cp38-win_amd64.whl"}, {"packagetype": "sdist", "filename": "tornado-6.0.4.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">=3.9", "version": "6.5.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-musllinux_1_2_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.5.2-cp39-abi3-win_arm64.whl"}, {"packagetype": "sdist", "filename": "tornado-6.5.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License (GPL)", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": null, "version": "1.2.10", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "trytond-1.2.10.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.4.27", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-4.4.27-py2-none-any.whl"}, {"packagetype": "bdist_wheel", "filename": "trytond-4.4.27-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-4.4.27.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.6.22", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-4.6.22-py2-none-any.whl"}, {"packagetype": "bdist_wheel", "filename": "trytond-4.6.22-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-4.6.22.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": "", "version": "4.8.18", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-4.8.18-py2-none-any.whl"}, {"packagetype": "bdist_wheel", "filename": "trytond-4.8.18-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-4.8.18.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Operating System :: OS Independent", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.4", "version": "5.0.63", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-5.0.63-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-5.0.63.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.5", "version": "5.4.20", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-5.4.20-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-5.4.20.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-5.8.16-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-5.8.16.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-6.2.14-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-6.2.14.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-6.8.17-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-6.8.17.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-7.6.9-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-7.6.9.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "typer-0.15.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "typer-0.15.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "typer-0.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "typer-0.20.0.tar.gz"}]} diff --git a/tests/integrations/launchdarkly/test_launchdarkly.py b/tests/integrations/launchdarkly/test_launchdarkly.py index 20bb4d031f..e588b596d3 100644 --- a/tests/integrations/launchdarkly/test_launchdarkly.py +++ b/tests/integrations/launchdarkly/test_launchdarkly.py @@ -199,8 +199,14 @@ def test_launchdarkly_integration_did_not_enable(monkeypatch): with pytest.raises(DidNotEnable): LaunchDarklyIntegration() + td = TestData.data_source() + # Disable background requests as we aren't using a server. + # Required because we corrupt the internal state above. + config = Config( + "sdk-key", update_processor_class=td, diagnostic_opt_out=True, send_events=False + ) # Client not initialized. - client = LDClient(config=Config("sdk-key")) + client = LDClient(config=config) monkeypatch.setattr(client, "is_initialized", lambda: False) with pytest.raises(DidNotEnable): LaunchDarklyIntegration(ld_client=client) diff --git a/tox.ini b/tox.ini index 7e5e60f39a..9a663a12c7 100644 --- a/tox.ini +++ b/tox.ini @@ -48,9 +48,9 @@ envlist = # ~~~ AI ~~~ {py3.8,py3.11,py3.12}-anthropic-v0.16.0 - {py3.8,py3.11,py3.12}-anthropic-v0.34.2 - {py3.8,py3.11,py3.12}-anthropic-v0.52.2 - {py3.8,py3.12,py3.13}-anthropic-v0.71.0 + {py3.8,py3.11,py3.12}-anthropic-v0.35.0 + {py3.8,py3.11,py3.12}-anthropic-v0.54.0 + {py3.8,py3.12,py3.13}-anthropic-v0.72.0 {py3.9,py3.10,py3.11}-cohere-v5.4.0 {py3.9,py3.11,py3.12}-cohere-v5.10.0 @@ -60,13 +60,11 @@ envlist = {py3.9,py3.12,py3.13}-google_genai-v1.29.0 {py3.9,py3.12,py3.13}-google_genai-v1.35.0 {py3.9,py3.12,py3.13}-google_genai-v1.41.0 - {py3.9,py3.13,py3.14}-google_genai-v1.46.0 + {py3.9,py3.13,py3.14,py3.14t}-google_genai-v1.46.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.28.1 - {py3.8,py3.12,py3.13}-huggingface_hub-v0.32.6 {py3.8,py3.12,py3.13}-huggingface_hub-v0.36.0 - {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.0rc7 + {py3.9,py3.12,py3.13}-huggingface_hub-v1.0.1 {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.12,py3.13}-langchain-base-v0.3.27 @@ -104,14 +102,14 @@ envlist = {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.18 {py3.10,py3.12,py3.13}-pydantic_ai-v1.2.1 {py3.10,py3.12,py3.13}-pydantic_ai-v1.4.0 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.6.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.7.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.13,py3.14}-boto3-v1.40.59 + {py3.9,py3.13,py3.14,py3.14t}-boto3-v1.40.61 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -134,7 +132,7 @@ envlist = {py3.7,py3.10,py3.11}-redis-v4.6.0 {py3.8,py3.11,py3.12}-redis-v5.3.1 {py3.9,py3.12,py3.13}-redis-v6.4.0 - {py3.9,py3.12,py3.13}-redis-v7.0.0 + {py3.9,py3.12,py3.13}-redis-v7.0.1 {py3.6}-redis_py_cluster_legacy-v1.3.6 {py3.6,py3.7,py3.8}-redis_py_cluster_legacy-v2.1.3 @@ -146,10 +144,10 @@ envlist = # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 - {py3.9,py3.13,py3.14}-launchdarkly-v9.12.1 + {py3.9,py3.13,py3.14,py3.14t}-launchdarkly-v9.12.2 - {py3.8,py3.13,py3.14}-openfeature-v0.7.5 - {py3.9,py3.13,py3.14}-openfeature-v0.8.3 + {py3.8,py3.13,py3.14,py3.14t}-openfeature-v0.7.5 + {py3.9,py3.13,py3.14,py3.14t}-openfeature-v0.8.3 {py3.7,py3.13,py3.14}-statsig-v0.55.3 {py3.7,py3.13,py3.14}-statsig-v0.66.0 @@ -170,7 +168,7 @@ envlist = {py3.8,py3.12,py3.13}-graphene-v3.4.3 {py3.8,py3.10,py3.11}-strawberry-v0.209.8 - {py3.10,py3.13,py3.14}-strawberry-v0.284.1 + {py3.10,py3.13,py3.14,py3.14t}-strawberry-v0.284.1 # ~~~ Network ~~~ @@ -185,7 +183,7 @@ envlist = {py3.9,py3.11,py3.12}-httpx-v0.28.1 {py3.6}-requests-v2.12.5 - {py3.9,py3.13,py3.14}-requests-v2.32.5 + {py3.9,py3.13,py3.14,py3.14t}-requests-v2.32.5 # ~~~ Tasks ~~~ @@ -193,8 +191,7 @@ envlist = {py3.8,py3.11,py3.12}-arq-v0.26.3 {py3.7}-beam-v2.14.0 - {py3.9,py3.12,py3.13}-beam-v2.68.0 - {py3.9,py3.12,py3.13}-beam-v2.69.0rc3 + {py3.9,py3.12,py3.13}-beam-v2.69.0 {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.8,py3.12,py3.13}-celery-v5.5.3 @@ -207,7 +204,7 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.4 {py3.9,py3.10}-ray-v2.7.2 - {py3.9,py3.12,py3.13}-ray-v2.50.1 + {py3.9,py3.12,py3.13}-ray-v2.51.0 {py3.6}-rq-v0.8.2 {py3.6,py3.7}-rq-v0.13.0 @@ -225,28 +222,28 @@ envlist = {py3.6,py3.9,py3.10}-django-v3.2.25 {py3.8,py3.11,py3.12}-django-v4.2.25 {py3.10,py3.12,py3.13}-django-v5.2.7 - {py3.12,py3.13,py3.14}-django-v6.0b1 + {py3.12,py3.13,py3.14,py3.14t}-django-v6.0b1 {py3.6,py3.7,py3.8}-flask-v1.1.4 - {py3.8,py3.13,py3.14}-flask-v2.3.3 - {py3.9,py3.13,py3.14}-flask-v3.1.2 + {py3.8,py3.13,py3.14,py3.14t}-flask-v2.3.3 + {py3.9,py3.13,py3.14,py3.14t}-flask-v3.1.2 {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.27.0 {py3.8,py3.12,py3.13}-starlette-v0.38.6 - {py3.9,py3.13,py3.14}-starlette-v0.48.0 + {py3.9,py3.13,py3.14,py3.14t}-starlette-v0.49.1 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.93.0 {py3.8,py3.10,py3.11}-fastapi-v0.107.0 - {py3.8,py3.13,py3.14}-fastapi-v0.120.0 + {py3.8,py3.13,py3.14,py3.14t}-fastapi-v0.120.1 # ~~~ Web 2 ~~~ {py3.7}-aiohttp-v3.4.4 {py3.7,py3.8,py3.9}-aiohttp-v3.7.4 {py3.8,py3.12,py3.13}-aiohttp-v3.10.11 - {py3.9,py3.13,py3.14}-aiohttp-v3.13.1 + {py3.9,py3.13,py3.14,py3.14t}-aiohttp-v3.13.2 {py3.6,py3.7}-bottle-v0.12.25 {py3.8,py3.12,py3.13}-bottle-v0.13.4 @@ -266,7 +263,7 @@ envlist = {py3.6,py3.10,py3.11}-pyramid-v2.0.2 {py3.7,py3.9,py3.10}-quart-v0.16.3 - {py3.9,py3.13,py3.14}-quart-v0.20.0 + {py3.9,py3.13,py3.14,py3.14t}-quart-v0.20.0 {py3.6}-sanic-v0.8.3 {py3.6,py3.8,py3.9}-sanic-v20.12.7 @@ -293,7 +290,7 @@ envlist = {py3.9,py3.12,py3.13}-trytond-v7.6.9 {py3.7,py3.12,py3.13}-typer-v0.15.4 - {py3.8,py3.13,py3.14}-typer-v0.20.0 + {py3.8,py3.13,py3.14,py3.14t}-typer-v0.20.0 @@ -353,12 +350,12 @@ deps = # ~~~ AI ~~~ anthropic-v0.16.0: anthropic==0.16.0 - anthropic-v0.34.2: anthropic==0.34.2 - anthropic-v0.52.2: anthropic==0.52.2 - anthropic-v0.71.0: anthropic==0.71.0 + anthropic-v0.35.0: anthropic==0.35.0 + anthropic-v0.54.0: anthropic==0.54.0 + anthropic-v0.72.0: anthropic==0.72.0 anthropic: pytest-asyncio anthropic-v0.16.0: httpx<0.28.0 - anthropic-v0.34.2: httpx<0.28.0 + anthropic-v0.35.0: httpx<0.28.0 cohere-v5.4.0: cohere==5.4.0 cohere-v5.10.0: cohere==5.10.0 @@ -372,10 +369,8 @@ deps = google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 - huggingface_hub-v0.28.1: huggingface_hub==0.28.1 - huggingface_hub-v0.32.6: huggingface_hub==0.32.6 huggingface_hub-v0.36.0: huggingface_hub==0.36.0 - huggingface_hub-v1.0.0rc7: huggingface_hub==1.0.0rc7 + huggingface_hub-v1.0.1: huggingface_hub==1.0.1 huggingface_hub: responses huggingface_hub: pytest-httpx @@ -433,7 +428,7 @@ deps = pydantic_ai-v1.0.18: pydantic-ai==1.0.18 pydantic_ai-v1.2.1: pydantic-ai==1.2.1 pydantic_ai-v1.4.0: pydantic-ai==1.4.0 - pydantic_ai-v1.6.0: pydantic-ai==1.6.0 + pydantic_ai-v1.7.0: pydantic-ai==1.7.0 pydantic_ai: pytest-asyncio @@ -441,7 +436,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.59: boto3==1.40.59 + boto3-v1.40.61: boto3==1.40.61 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -468,7 +463,7 @@ deps = redis-v4.6.0: redis==4.6.0 redis-v5.3.1: redis==5.3.1 redis-v6.4.0: redis==6.4.0 - redis-v7.0.0: redis==7.0.0 + redis-v7.0.1: redis==7.0.1 redis: fakeredis!=1.7.4 redis: pytest<8.0.0 redis-v4.6.0: fakeredis<2.31.0 @@ -485,7 +480,7 @@ deps = # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 - launchdarkly-v9.12.1: launchdarkly-server-sdk==9.12.1 + launchdarkly-v9.12.2: launchdarkly-server-sdk==9.12.2 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.3: openfeature-sdk==0.8.3 @@ -556,8 +551,7 @@ deps = arq-v0.23: pydantic<2 beam-v2.14.0: apache-beam==2.14.0 - beam-v2.68.0: apache-beam==2.68.0 - beam-v2.69.0rc3: apache-beam==2.69.0rc3 + beam-v2.69.0: apache-beam==2.69.0 beam: dill celery-v4.4.7: celery==4.4.7 @@ -574,7 +568,7 @@ deps = huey-v2.5.4: huey==2.5.4 ray-v2.7.2: ray==2.7.2 - ray-v2.50.1: ray==2.50.1 + ray-v2.51.0: ray==2.51.0 rq-v0.8.2: rq==0.8.2 rq-v0.13.0: rq==0.13.0 @@ -632,7 +626,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.27.0: starlette==0.27.0 starlette-v0.38.6: starlette==0.38.6 - starlette-v0.48.0: starlette==0.48.0 + starlette-v0.49.1: starlette==0.49.1 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -646,7 +640,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.93.0: fastapi==0.93.0 fastapi-v0.107.0: fastapi==0.107.0 - fastapi-v0.120.0: fastapi==0.120.0 + fastapi-v0.120.1: fastapi==0.120.1 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -662,10 +656,10 @@ deps = aiohttp-v3.4.4: aiohttp==3.4.4 aiohttp-v3.7.4: aiohttp==3.7.4 aiohttp-v3.10.11: aiohttp==3.10.11 - aiohttp-v3.13.1: aiohttp==3.13.1 + aiohttp-v3.13.2: aiohttp==3.13.2 aiohttp: pytest-aiohttp aiohttp-v3.10.11: pytest-asyncio - aiohttp-v3.13.1: pytest-asyncio + aiohttp-v3.13.2: pytest-asyncio bottle-v0.12.25: bottle==0.12.25 bottle-v0.13.4: bottle==0.13.4 From 2d49b74b0c19556f889836bf78773b2183044145 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Thu, 30 Oct 2025 14:31:58 +0100 Subject: [PATCH 586/591] Allow new integration setup on the instance with config options (#5047) ### Description Currently `setup_once` is called without any arguments so cannot access either the SDK init `options` or the integration specific init arguments on the instance. This PR enables a separate backwards compat optional `setup_once_with_options` as an **instance** method. Note that we we will use this rarely but I need it to unblock the OTLPIntegration since I need the DSN there. #### Example before ```python sentry_sdk.init(dsn="") class FooIntegration(Integration): def __init__(self, bar=42): self.bar = bar @staticmethod def setup_once(): self.bar # Not available options.dsn # Not available ``` after ```python def setup_once_with_options(self, options=None): self.bar options.dsn ``` --------- Co-authored-by: Ivana Kellyer --- .gitignore | 1 + sentry_sdk/client.py | 1 + sentry_sdk/integrations/__init__.py | 20 +++++++++++++++----- 3 files changed, 17 insertions(+), 5 deletions(-) diff --git a/.gitignore b/.gitignore index 3b1e93c41a..a37ad4a60f 100644 --- a/.gitignore +++ b/.gitignore @@ -29,6 +29,7 @@ pip-wheel-metadata .mypy_cache .vscode/ .claude/ +.tool-versions # for running AWS Lambda tests using AWS SAM sam.template.yaml diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index 91096c6b4f..af8a1c1157 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -409,6 +409,7 @@ def _capture_envelope(envelope): "auto_enabling_integrations" ], disabled_integrations=self.options["disabled_integrations"], + options=self.options, ) spotlight_config = self.options.get("spotlight") diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index 2d3d621ef7..af4206e208 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -15,6 +15,7 @@ from typing import Set from typing import Type from typing import Union + from typing import Any _DEFAULT_FAILED_REQUEST_STATUS_CODES = frozenset(range(500, 600)) @@ -171,12 +172,13 @@ def iter_default_integrations(with_auto_enabling_integrations): def setup_integrations( - integrations, - with_defaults=True, - with_auto_enabling_integrations=False, - disabled_integrations=None, + integrations, # type: Sequence[Integration] + with_defaults=True, # type: bool + with_auto_enabling_integrations=False, # type: bool + disabled_integrations=None, # type: Optional[Sequence[Union[type[Integration], Integration]]] + options=None, # type: Optional[Dict[str, Any]] ): - # type: (Sequence[Integration], bool, bool, Optional[Sequence[Union[type[Integration], Integration]]]) -> Dict[str, Integration] + # type: (...) -> Dict[str, Integration] """ Given a list of integration instances, this installs them all. @@ -221,6 +223,7 @@ def setup_integrations( ) try: type(integration).setup_once() + integration.setup_once_with_options(options) except DidNotEnable as e: if identifier not in used_as_default_integration: raise @@ -300,3 +303,10 @@ def setup_once(): instance again. """ pass + + def setup_once_with_options(self, options=None): + # type: (Optional[Dict[str, Any]]) -> None + """ + Called after setup_once in rare cases on the instance and with options since we don't have those available above. + """ + pass From 2397b15bd2e0018c52853563bb520159f5b5b4b6 Mon Sep 17 00:00:00 2001 From: Alex Alderman Webb Date: Fri, 31 Oct 2025 12:58:48 +0100 Subject: [PATCH 587/591] chore: Remove `enable_metrics` option (#5046) Remove the `enable_metrics` option. Users can capture metrics with `sentry_sdk.metrics` functions without further opt-in. Closes https://github.com/getsentry/sentry-python/issues/5048 --- sentry_sdk/client.py | 11 +++-------- sentry_sdk/utils.py | 8 -------- tests/test_metrics.py | 27 ++++++--------------------- 3 files changed, 9 insertions(+), 37 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index af8a1c1157..c711097d4c 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -10,6 +10,7 @@ import sentry_sdk from sentry_sdk._compat import PY37, check_uwsgi_thread_support +from sentry_sdk._metrics_batcher import MetricsBatcher from sentry_sdk.utils import ( AnnotatedValue, ContextVar, @@ -26,7 +27,6 @@ get_before_send_log, get_before_send_metric, has_logs_enabled, - has_metrics_enabled, ) from sentry_sdk.serializer import serialize from sentry_sdk.tracing import trace @@ -374,12 +374,7 @@ def _capture_envelope(envelope): self.log_batcher = LogBatcher(capture_func=_capture_envelope) - self.metrics_batcher = None - - if has_metrics_enabled(self.options): - from sentry_sdk._metrics_batcher import MetricsBatcher - - self.metrics_batcher = MetricsBatcher(capture_func=_capture_envelope) + self.metrics_batcher = MetricsBatcher(capture_func=_capture_envelope) max_request_body_size = ("always", "never", "small", "medium") if self.options["max_request_body_size"] not in max_request_body_size: @@ -979,7 +974,7 @@ def _capture_log(self, log): def _capture_metric(self, metric): # type: (Optional[Metric]) -> None - if not has_metrics_enabled(self.options) or metric is None: + if metric is None: return isolation_scope = sentry_sdk.get_isolation_scope() diff --git a/sentry_sdk/utils.py b/sentry_sdk/utils.py index 3496178228..93c35cee8a 100644 --- a/sentry_sdk/utils.py +++ b/sentry_sdk/utils.py @@ -2047,14 +2047,6 @@ def get_before_send_log(options): ) -def has_metrics_enabled(options): - # type: (Optional[dict[str, Any]]) -> bool - if options is None: - return False - - return bool(options["_experiments"].get("enable_metrics", False)) - - def get_before_send_metric(options): # type: (Optional[dict[str, Any]]) -> Optional[Callable[[Metric, Hint], Optional[Metric]]] if options is None: diff --git a/tests/test_metrics.py b/tests/test_metrics.py index d5ca8ec572..b0e92c9320 100644 --- a/tests/test_metrics.py +++ b/tests/test_metrics.py @@ -33,20 +33,8 @@ def envelopes_to_metrics(envelopes): return res -def test_metrics_disabled_by_default(sentry_init, capture_envelopes): - sentry_init() - - envelopes = capture_envelopes() - - sentry_sdk.metrics.count("test.counter", 1) - sentry_sdk.metrics.gauge("test.gauge", 42) - sentry_sdk.metrics.distribution("test.distribution", 200) - - assert len(envelopes) == 0 - - def test_metrics_basics(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_metrics": True}) + sentry_init() envelopes = capture_envelopes() sentry_sdk.metrics.count("test.counter", 1) @@ -77,7 +65,7 @@ def test_metrics_basics(sentry_init, capture_envelopes): def test_metrics_experimental_option(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_metrics": True}) + sentry_init() envelopes = capture_envelopes() sentry_sdk.metrics.count("test.counter", 5) @@ -93,9 +81,7 @@ def test_metrics_experimental_option(sentry_init, capture_envelopes): def test_metrics_with_attributes(sentry_init, capture_envelopes): - sentry_init( - _experiments={"enable_metrics": True}, release="1.0.0", environment="test" - ) + sentry_init(release="1.0.0", environment="test") envelopes = capture_envelopes() sentry_sdk.metrics.count( @@ -114,7 +100,7 @@ def test_metrics_with_attributes(sentry_init, capture_envelopes): def test_metrics_with_user(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_metrics": True}) + sentry_init() envelopes = capture_envelopes() sentry_sdk.set_user( @@ -133,7 +119,7 @@ def test_metrics_with_user(sentry_init, capture_envelopes): def test_metrics_with_span(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_metrics": True}, traces_sample_rate=1.0) + sentry_init(traces_sample_rate=1.0) envelopes = capture_envelopes() with sentry_sdk.start_transaction(op="test", name="test-span"): @@ -150,7 +136,7 @@ def test_metrics_with_span(sentry_init, capture_envelopes): def test_metrics_tracing_without_performance(sentry_init, capture_envelopes): - sentry_init(_experiments={"enable_metrics": True}) + sentry_init() envelopes = capture_envelopes() sentry_sdk.metrics.count("test.span.counter", 1) @@ -190,7 +176,6 @@ def _before_metric(record, hint): sentry_init( _experiments={ - "enable_metrics": True, "before_send_metric": _before_metric, }, ) From 89658215506c42f6049b5d2d3e4a71cac22cf65d Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Fri, 31 Oct 2025 15:28:29 +0100 Subject: [PATCH 588/591] Add external_propagation_context support (#5051) If we are on an external tracing system like otel, we allow registering a new source of `trace_id/span_id` that takes precedence over the scope's propagation_context. * Also reworked logs and metrics to use `get_trace_context` * Cleaned up handling of `get_trace_context` that is still messy but a bit more clearer now regarding which underlying `propagation_context` is used --- sentry_sdk/client.py | 39 ++++++++------- sentry_sdk/scope.py | 56 ++++++++++++++++------ tests/integrations/logging/test_logging.py | 4 ++ tests/integrations/loguru/test_loguru.py | 4 ++ tests/test_logs.py | 4 +- tests/test_metrics.py | 16 ++++--- tests/test_scope.py | 46 ++++++++++++++++++ 7 files changed, 126 insertions(+), 43 deletions(-) diff --git a/sentry_sdk/client.py b/sentry_sdk/client.py index c711097d4c..24824e0050 100644 --- a/sentry_sdk/client.py +++ b/sentry_sdk/client.py @@ -928,17 +928,18 @@ def _capture_log(self, log): if release is not None and "sentry.release" not in log["attributes"]: log["attributes"]["sentry.release"] = release - span = current_scope.span - if span is not None and "sentry.trace.parent_span_id" not in log["attributes"]: - log["attributes"]["sentry.trace.parent_span_id"] = span.span_id - - if log.get("trace_id") is None: - transaction = current_scope.transaction - propagation_context = isolation_scope.get_active_propagation_context() - if transaction is not None: - log["trace_id"] = transaction.trace_id - elif propagation_context is not None: - log["trace_id"] = propagation_context.trace_id + trace_context = current_scope.get_trace_context() + trace_id = trace_context.get("trace_id") + span_id = trace_context.get("span_id") + + if trace_id is not None and log.get("trace_id") is None: + log["trace_id"] = trace_id + + if ( + span_id is not None + and "sentry.trace.parent_span_id" not in log["attributes"] + ): + log["attributes"]["sentry.trace.parent_span_id"] = span_id # The user, if present, is always set on the isolation scope. if isolation_scope._user is not None: @@ -977,6 +978,7 @@ def _capture_metric(self, metric): if metric is None: return + current_scope = sentry_sdk.get_current_scope() isolation_scope = sentry_sdk.get_isolation_scope() metric["attributes"]["sentry.sdk.name"] = SDK_INFO["name"] @@ -990,16 +992,13 @@ def _capture_metric(self, metric): if release is not None and "sentry.release" not in metric["attributes"]: metric["attributes"]["sentry.release"] = release - span = sentry_sdk.get_current_span() - metric["trace_id"] = "00000000-0000-0000-0000-000000000000" + trace_context = current_scope.get_trace_context() + trace_id = trace_context.get("trace_id") + span_id = trace_context.get("span_id") - if span: - metric["trace_id"] = span.trace_id - metric["span_id"] = span.span_id - else: - propagation_context = isolation_scope.get_active_propagation_context() - if propagation_context and propagation_context.trace_id: - metric["trace_id"] = propagation_context.trace_id + metric["trace_id"] = trace_id or "00000000-0000-0000-0000-000000000000" + if span_id is not None: + metric["span_id"] = span_id if isolation_scope._user is not None: for metric_attribute, user_attribute in ( diff --git a/sentry_sdk/scope.py b/sentry_sdk/scope.py index ecb8f370c5..8e55add770 100644 --- a/sentry_sdk/scope.py +++ b/sentry_sdk/scope.py @@ -107,6 +107,10 @@ global_event_processors = [] # type: List[EventProcessor] +# A function returning a (trace_id, span_id) tuple +# from an external tracing source (such as otel) +_external_propagation_context_fn = None # type: Optional[Callable[[], Optional[Tuple[str, str]]]] + class ScopeType(Enum): CURRENT = "current" @@ -142,6 +146,25 @@ def add_global_event_processor(processor): global_event_processors.append(processor) +def register_external_propagation_context(fn): + # type: (Callable[[], Optional[Tuple[str, str]]]) -> None + global _external_propagation_context_fn + _external_propagation_context_fn = fn + + +def remove_external_propagation_context(): + # type: () -> None + global _external_propagation_context_fn + _external_propagation_context_fn = None + + +def get_external_propagation_context(): + # type: () -> Optional[Tuple[str, str]] + return ( + _external_propagation_context_fn() if _external_propagation_context_fn else None + ) + + def _attr_setter(fn): # type: (Any) -> Any return property(fset=fn, doc=fn.__doc__) @@ -562,21 +585,29 @@ def get_baggage(self, *args, **kwargs): return self.get_isolation_scope().get_baggage() def get_trace_context(self): - # type: () -> Any + # type: () -> Dict[str, Any] """ Returns the Sentry "trace" context from the Propagation Context. """ - if self._propagation_context is None: - return None + if has_tracing_enabled(self.get_client().options) and self._span is not None: + return self._span.get_trace_context() - trace_context = { - "trace_id": self._propagation_context.trace_id, - "span_id": self._propagation_context.span_id, - "parent_span_id": self._propagation_context.parent_span_id, - "dynamic_sampling_context": self.get_dynamic_sampling_context(), - } # type: Dict[str, Any] + # if we are tracing externally (otel), those values take precedence + external_propagation_context = get_external_propagation_context() + if external_propagation_context: + trace_id, span_id = external_propagation_context + return {"trace_id": trace_id, "span_id": span_id} - return trace_context + propagation_context = self.get_active_propagation_context() + if propagation_context is None: + return {} + + return { + "trace_id": propagation_context.trace_id, + "span_id": propagation_context.span_id, + "parent_span_id": propagation_context.parent_span_id, + "dynamic_sampling_context": self.get_dynamic_sampling_context(), + } def trace_propagation_meta(self, *args, **kwargs): # type: (*Any, **Any) -> str @@ -1438,10 +1469,7 @@ def _apply_contexts_to_event(self, event, hint, options): # Add "trace" context if contexts.get("trace") is None: - if has_tracing_enabled(options) and self._span is not None: - contexts["trace"] = self._span.get_trace_context() - else: - contexts["trace"] = self.get_trace_context() + contexts["trace"] = self.get_trace_context() def _apply_flags_to_event(self, event, hint, options): # type: (Event, Hint, Optional[Dict[str, Any]]) -> None diff --git a/tests/integrations/logging/test_logging.py b/tests/integrations/logging/test_logging.py index 7a00ceadd2..121025bbb6 100644 --- a/tests/integrations/logging/test_logging.py +++ b/tests/integrations/logging/test_logging.py @@ -478,6 +478,10 @@ def test_logger_with_all_attributes(sentry_init, capture_envelopes): assert attributes.pop("sentry.sdk.name").startswith("sentry.python") + assert "sentry.trace.parent_span_id" in attributes + assert isinstance(attributes["sentry.trace.parent_span_id"], str) + del attributes["sentry.trace.parent_span_id"] + # Assert on the remaining non-dynamic attributes. assert attributes == { "foo": "bar", diff --git a/tests/integrations/loguru/test_loguru.py b/tests/integrations/loguru/test_loguru.py index 3d04d7d1ea..2a307a50cb 100644 --- a/tests/integrations/loguru/test_loguru.py +++ b/tests/integrations/loguru/test_loguru.py @@ -458,6 +458,10 @@ def test_logger_with_all_attributes( assert attributes.pop("sentry.sdk.name").startswith("sentry.python") + assert "sentry.trace.parent_span_id" in attributes + assert isinstance(attributes["sentry.trace.parent_span_id"], str) + del attributes["sentry.trace.parent_span_id"] + # Assert on the remaining non-dynamic attributes. assert attributes == { "logger.name": "tests.integrations.loguru.test_loguru", diff --git a/tests/test_logs.py b/tests/test_logs.py index 7494aa01c8..da7d0d5f03 100644 --- a/tests/test_logs.py +++ b/tests/test_logs.py @@ -310,7 +310,7 @@ def test_logs_tied_to_transactions(sentry_init, capture_envelopes): """ Log messages are also tied to transactions. """ - sentry_init(enable_logs=True) + sentry_init(enable_logs=True, traces_sample_rate=1.0) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction") as trx: @@ -326,7 +326,7 @@ def test_logs_tied_to_spans(sentry_init, capture_envelopes): """ Log messages are also tied to spans. """ - sentry_init(enable_logs=True) + sentry_init(enable_logs=True, traces_sample_rate=1.0) envelopes = capture_envelopes() with sentry_sdk.start_transaction(name="test-transaction"): diff --git a/tests/test_metrics.py b/tests/test_metrics.py index b0e92c9320..8026e17bb6 100644 --- a/tests/test_metrics.py +++ b/tests/test_metrics.py @@ -122,7 +122,7 @@ def test_metrics_with_span(sentry_init, capture_envelopes): sentry_init(traces_sample_rate=1.0) envelopes = capture_envelopes() - with sentry_sdk.start_transaction(op="test", name="test-span"): + with sentry_sdk.start_transaction(op="test", name="test-span") as transaction: sentry_sdk.metrics.count("test.span.counter", 1) get_client().flush() @@ -131,24 +131,26 @@ def test_metrics_with_span(sentry_init, capture_envelopes): assert len(metrics) == 1 assert metrics[0]["trace_id"] is not None - assert metrics[0]["trace_id"] != "00000000-0000-0000-0000-000000000000" - assert metrics[0]["span_id"] is not None + assert metrics[0]["trace_id"] == transaction.trace_id + assert metrics[0]["span_id"] == transaction.span_id def test_metrics_tracing_without_performance(sentry_init, capture_envelopes): sentry_init() envelopes = capture_envelopes() - sentry_sdk.metrics.count("test.span.counter", 1) + with sentry_sdk.isolation_scope() as isolation_scope: + sentry_sdk.metrics.count("test.span.counter", 1) get_client().flush() metrics = envelopes_to_metrics(envelopes) assert len(metrics) == 1 - assert metrics[0]["trace_id"] is not None - assert metrics[0]["trace_id"] != "00000000-0000-0000-0000-000000000000" - assert metrics[0]["span_id"] is None + propagation_context = isolation_scope._propagation_context + assert propagation_context is not None + assert metrics[0]["trace_id"] == propagation_context.trace_id + assert metrics[0]["span_id"] == propagation_context.span_id def test_metrics_before_send(sentry_init, capture_envelopes): diff --git a/tests/test_scope.py b/tests/test_scope.py index 68c93f3036..1ace1cc73c 100644 --- a/tests/test_scope.py +++ b/tests/test_scope.py @@ -16,6 +16,8 @@ use_isolation_scope, use_scope, should_send_default_pii, + register_external_propagation_context, + remove_external_propagation_context, ) @@ -971,3 +973,47 @@ def test_handle_error_on_token_reset_isolation_scope(error_cls, scope_manager): mock_capture.assert_called_once() mock_current_scope.reset.assert_called_once_with(mock_current_token) + + +def test_trace_context_tracing(sentry_init): + sentry_init(traces_sample_rate=1.0) + + with sentry_sdk.start_transaction(name="trx") as transaction: + with sentry_sdk.start_span(op="span1"): + with sentry_sdk.start_span(op="span2") as span: + trace_context = sentry_sdk.get_current_scope().get_trace_context() + + assert trace_context["trace_id"] == transaction.trace_id + assert trace_context["span_id"] == span.span_id + assert trace_context["parent_span_id"] == span.parent_span_id + assert "dynamic_sampling_context" in trace_context + + +def test_trace_context_external_tracing(sentry_init): + sentry_init() + + def external_propagation_context(): + return ("trace_id_foo", "span_id_bar") + + register_external_propagation_context(external_propagation_context) + + trace_context = sentry_sdk.get_current_scope().get_trace_context() + + assert trace_context["trace_id"] == "trace_id_foo" + assert trace_context["span_id"] == "span_id_bar" + + remove_external_propagation_context() + + +def test_trace_context_without_performance(sentry_init): + sentry_init() + + with sentry_sdk.isolation_scope() as isolation_scope: + trace_context = sentry_sdk.get_current_scope().get_trace_context() + + propagation_context = isolation_scope._propagation_context + assert propagation_context is not None + assert trace_context["trace_id"] == propagation_context.trace_id + assert trace_context["span_id"] == propagation_context.span_id + assert trace_context["parent_span_id"] == propagation_context.parent_span_id + assert "dynamic_sampling_context" in trace_context From f2e14a64a2a9235056901ca26e7cd8077928e2c9 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 3 Nov 2025 12:52:29 +0100 Subject: [PATCH 589/591] =?UTF-8?q?ci:=20=F0=9F=A4=96=20Update=20test=20ma?= =?UTF-8?q?trix=20with=20new=20releases=20(11/03)=20(#5054)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update our test matrix with new releases of integrated frameworks and libraries. ## How it works - Scan PyPI for all supported releases of all frameworks we have a dedicated test suite for. - Pick a representative sample of releases to run our test suite against. We always test the latest and oldest supported version. - Update [tox.ini](https://github.com/getsentry/sentry-python/blob/master/tox.ini) with the new releases. ## Action required - If CI passes on this PR, it's safe to approve and merge. It means our integrations can handle new versions of frameworks that got pulled in. - If CI doesn't pass on this PR, this points to an incompatibility of either our integration or our test setup with a new version of a framework. - Check what the failures look like and either fix them, or update the [test config](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/config.py) and rerun [scripts/generate-test-files.sh](https://github.com/getsentry/sentry-python/blob/master/scripts/generate-test-files.sh). See [scripts/populate_tox/README.md](https://github.com/getsentry/sentry-python/blob/master/scripts/populate_tox/README.md) for what configuration options are available. _____________________ _🤖 This PR was automatically created using [a GitHub action](https://github.com/getsentry/sentry-python/blob/master/.github/workflows/update-tox.yml)._ --------- Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Alexander Alderman Webb --- .../populate_tox/package_dependencies.jsonl | 14 ++-- scripts/populate_tox/releases.jsonl | 32 ++++---- tox.ini | 76 +++++++++---------- 3 files changed, 61 insertions(+), 61 deletions(-) diff --git a/scripts/populate_tox/package_dependencies.jsonl b/scripts/populate_tox/package_dependencies.jsonl index 8217222282..d467fc125c 100644 --- a/scripts/populate_tox/package_dependencies.jsonl +++ b/scripts/populate_tox/package_dependencies.jsonl @@ -1,18 +1,18 @@ -{"name": "boto3", "version": "1.40.61", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/61/24/3bf865b07d15fea85b63504856e137029b6acbc73762496064219cdb265d/boto3-1.40.61-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/38/c5/f6ce561004db45f0b847c2cd9b19c67c6bf348a82018a48cb718be6b58b0/botocore-1.40.61-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/f0/ae7ca09223a81a1d890b2557186ea015f6e0502e9b8cb8e1813f1d8cfa4e/s3transfer-0.14.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}]} +{"name": "boto3", "version": "1.40.64", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/65/c2/27da558ceb90d17b1e4c0cca5dab29f8aea7f63242a1005a8f54230ce5e6/boto3-1.40.64-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8f/c5/70bec18aef3fe9af63847d8766f81864b20daacd1dc7bf0c1d1ad90c7e98/botocore-1.40.64-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/f0/ae7ca09223a81a1d890b2557186ea015f6e0502e9b8cb8e1813f1d8cfa4e/s3transfer-0.14.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}]} {"name": "django", "version": "6.0b1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/0e/1a/306fda7e62e27ccbcb92d97f67f1094352a9f22c62f3c2b238fa50eb82d7/django-6.0b1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/17/9c/fc2331f538fbf7eedba64b2052e99ccf9ba9d6888e2f41441ee28847004b/asgiref-3.10.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a9/5c/bfd6bd0bf979426d405cc6e71eceb8701b148b16c21d2dc3c261efc61c7b/sqlparse-0.5.3-py3-none-any.whl"}}]} -{"name": "fastapi", "version": "0.120.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/7e/bb/1a74dbe87e9a595bf63052c886dfef965dc5b91d149456a8301eb3d41ce2/fastapi-0.120.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/51/da/545b75d420bb23b5d494b0517757b351963e974e79933f01e05c929f20a6/starlette-0.49.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/02/b7/cf592cb5de5cb3bade3357f8d2cf42bf103bbe39f459824b4939fd212911/annotated_doc-0.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "fastapi", "version": "0.120.4", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/ed/47/14a76b926edc3957c8a8258423db789d3fa925d2fed800102fce58959413/fastapi-0.120.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a3/e0/021c772d6a662f43b63044ab481dc6ac7592447605b5b35a957785363122/starlette-0.49.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/02/b7/cf592cb5de5cb3bade3357f8d2cf42bf103bbe39f459824b4939fd212911/annotated_doc-0.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} {"name": "flask", "version": "2.3.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/fd/56/26f0be8adc2b4257df20c1c4260ddd0aa396cf8e75d90ab2f7ff99bc34f9/flask-2.3.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} {"name": "flask", "version": "3.1.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} -{"name": "google-genai", "version": "1.46.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/db/79/8993ec6cbf56e5c8f88c165380e55de34ec74f7b928bc302ff5c370f9c4e/google_genai-1.46.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/87/24/ec82aee6ba1a076288818fe5cc5125f4d93fffdc68bb7b381c68286c8aaa/google_auth-2.42.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/96/c5/1e741d26306c42e2bf6ab740b2202872727e0f606033c9dd713f8b93f5a8/cachetools-6.2.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/8d/0133e4eb4beed9e425d9a98ed6e081a55d195481b7632472be1af08d2f6b/rsa-4.9.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fa/a8/5b41e0da817d64113292ab1f8247140aac61cbf6cfd085d6a0fa77f4984f/websockets-15.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c8/f1/d6a797abb14f6283c0ddff96bbdd46937f64122b8c925cab503dd37f8214/pyasn1-0.6.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/47/8d/d529b5d697919ba8c11ad626e835d4039be708a35b0d22de83a269a6682c/pyasn1_modules-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} -{"name": "langchain", "version": "1.0.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/8c/06/0e03587da37173c29a58bf17312793c2453df9ca2912e9adfe869c120437/langchain-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c0/54/3cdbe9d151d06cd689b5aa937ac11403b64bbfe76486fda6431a24061721/langchain_core-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b1/3c/acc0956a0da96b25a2c5c1a85168eacf1253639a04ed391d7a7bcaae5d6c/langgraph-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/68/47/9ffd10882403020ea866e381de7f8e504a78f606a914af7f8244456c7783/langgraph_prebuilt-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b4/2b/7e0248f65e35800ea8e4e3dbb3bcc36c61b81f5b8abeddaceec8320ab491/langsmith-0.4.38-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "google-genai", "version": "1.47.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/89/ef/e080e8d67c270ea320956bb911a9359664fc46d3b87d1f029decd33e5c4c/google_genai-1.47.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/92/05/adeb6c495aec4f9d93f9e2fc29eeef6e14d452bba11d15bdb874ce1d5b10/google_auth-2.42.1-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/96/c5/1e741d26306c42e2bf6ab740b2202872727e0f606033c9dd713f8b93f5a8/cachetools-6.2.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/8d/0133e4eb4beed9e425d9a98ed6e081a55d195481b7632472be1af08d2f6b/rsa-4.9.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fa/a8/5b41e0da817d64113292ab1f8247140aac61cbf6cfd085d6a0fa77f4984f/websockets-15.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c8/f1/d6a797abb14f6283c0ddff96bbdd46937f64122b8c925cab503dd37f8214/pyasn1-0.6.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/47/8d/d529b5d697919ba8c11ad626e835d4039be708a35b0d22de83a269a6682c/pyasn1_modules-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "langchain", "version": "1.0.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/68/c8/b5dcfdde8b96369e5445f0fbac52fe8495bbd11b23ca83691d90d464eb15/langchain-1.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/54/3aed89938a42cf7115575c333647551e35adc380feed651105d2d86c22f5/langchain_core-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d7/b1/9f4912e13d4ed691f2685c8a4b764b5a9237a30cca0c5782bc213d9f0a9a/langgraph-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/27/2f/9a7d00d4afa036e65294059c7c912002fb72ba5dbbd5c2a871ca06360278/langgraph_prebuilt-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1f/38/9a97f650b8cdb2ba0356d65aef9239f4a30db69ae44c30daa2cf8dd3f350/langsmith-0.4.39-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b1/5e/f77dc6b9036943e285ba76b49e118d9ea929885becb0a29ba8a7c75e29fe/pyyaml-6.0.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/97/86/e5b50247a61caec5718122feb2719ea9d451d30ac0516c288c1dbc6408e8/ormsgpack-1.11.0-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8c/63/8ffc2cc97e811c0ca5d00ab36604b3ea6f4254f20b7bc658ca825ce6c954/xxhash-3.6.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} {"name": "langgraph", "version": "0.6.11", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/df/94/430f0341c5c2fe3e3b9f5ab2622f35e2bda12c4a7d655c519468e853d1b0/langgraph-0.6.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8e/d1/e4727f4822943befc3b7046f79049b1086c9493a34b4d44a1adf78577693/langgraph_prebuilt-0.6.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c0/54/3cdbe9d151d06cd689b5aa937ac11403b64bbfe76486fda6431a24061721/langchain_core-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b4/2b/7e0248f65e35800ea8e4e3dbb3bcc36c61b81f5b8abeddaceec8320ab491/langsmith-0.4.38-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} -{"name": "launchdarkly-server-sdk", "version": "9.12.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/d8/63/30b2a1ec9994d04d82355663022734919903b54c7497b302206f67316237/launchdarkly_server_sdk-9.12.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/95/c20ba9cd28951f360b222da09138daf9169ad951be2a0d500eb0f0c84614/launchdarkly_eventsource-1.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/cb/84/a04c59324445f4bcc98dc05b39a1cd07c242dde643c1a3c21e4f7beaf2f2/expiringdict-1.2.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/34/90/0200184d2124484f918054751ef997ed6409cb05b7e8dcbf5a22da4c4748/pyrfc3339-2.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl"}}]} +{"name": "launchdarkly-server-sdk", "version": "9.12.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/5a/c0/8c0bc0584ac8c01288c09f14eb8002a2ebe433d6901f0d978c605c51ca8d/launchdarkly_server_sdk-9.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/95/c20ba9cd28951f360b222da09138daf9169ad951be2a0d500eb0f0c84614/launchdarkly_eventsource-1.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/cb/84/a04c59324445f4bcc98dc05b39a1cd07c242dde643c1a3c21e4f7beaf2f2/expiringdict-1.2.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/34/90/0200184d2124484f918054751ef997ed6409cb05b7e8dcbf5a22da4c4748/pyrfc3339-2.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl"}}]} {"name": "openfeature-sdk", "version": "0.7.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9b/20/cb043f54b11505d993e4dd84652cfc44c1260dc94b7f41aa35489af58277/openfeature_sdk-0.7.5-py3-none-any.whl"}}]} {"name": "openfeature-sdk", "version": "0.8.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/f0/f5/707a5b144115de1a49bf5761a63af2545fef0a1824f72db39ddea0a3438f/openfeature_sdk-0.8.3-py3-none-any.whl"}}]} {"name": "quart", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/7e/e9/cc28f21f52913adf333f653b9e0a3bf9cb223f5083a26422968ba73edd8d/quart-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/3b/dfa13a8d96aa24e40ea74a975a9906cfdc2ab2f4e3b498862a57052f04eb/hypercorn-0.17.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/69/b2/119f6e6dcbd96f9069ce9a2665e0146588dc9f88f29549711853645e736a/h2-4.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/c6/80c95b1b2b94682a72cbdbfb85b81ae2daffa4291fbfa1b1464502ede10d/hpack-4.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/30/47d0bf6072f7252e6521f3447ccfa40b421b6824517f82854703d0f5a98b/hyperframe-6.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/58/e860788190eba3bcce367f74d29c4675466ce8dddfba85f7827588416f01/wsproto-1.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/bc/8a/340a1555ae33d7354dbca4faa54948d76d89a27ceef032c8c3bc661d003e/aiofiles-25.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/5e/5f/82c8074f7e84978129347c2c6ec8b6c59f3584ff1a20bc3c940a3e061790/priority-2.0.0-py3-none-any.whl"}}]} {"name": "requests", "version": "2.32.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}]} -{"name": "starlette", "version": "0.49.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/51/da/545b75d420bb23b5d494b0517757b351963e974e79933f01e05c929f20a6/starlette-0.49.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "starlette", "version": "0.50.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/d9/52/1064f510b141bd54025f9b55105e26d1fa970b9be67ad766380a3c9b74b0/starlette-0.50.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} {"name": "statsig", "version": "0.55.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9d/17/de62fdea8aab8aa7c4a833378e0e39054b728dfd45ef279e975ed5ef4e86/statsig-0.55.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} -{"name": "statsig", "version": "0.66.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/dd/a3/d448734d1921c08b1b00fd69ef6e383b1a747ce9a32fc8e65aa3700ccb8c/statsig-0.66.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2f/c2/f9e977608bdf958650638c3f1e28f85a1b075f075ebbe77db8555463787b/Brotli-1.1.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} +{"name": "statsig", "version": "0.66.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/75/cf/06d818a72e489c4d5aec4399ef4ee69777ba2cb73ad9a64fdeed19b149a1/statsig-0.66.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2f/c2/f9e977608bdf958650638c3f1e28f85a1b075f075ebbe77db8555463787b/Brotli-1.1.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/78/63a0bcc0707037df4e22bb836451279d850592258c859685a402c27f5d6d/ijson-3.4.0.post0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} {"name": "strawberry-graphql", "version": "0.284.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/eb/9d/03a4225721f82dc8d25de894ae0b29ede0c5710de88a1710a8c081c88a4c/strawberry_graphql-0.284.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ae/4f/7297663840621022bc73c22d7d9d80dbc78b4db6297f764b545cd5dd462d/graphql_core-3.2.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/00/f2/c68a97c727c795119f1056ad2b7e716c23f26f004292517c435accf90b5c/lia_web-0.2.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} {"name": "typer", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/78/64/7713ffe4b5983314e9d436a90d5bd4f63b6054e2aca783a3cfc44cb95bbf/typer-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/25/7a/b0178788f8dc6cafce37a212c99565fa1fe7872c70c6c9c1e1a372d9d88f/rich-14.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/54/e7d793b573f298e1c9013b8c4dade17d481164aa517d1d7148619c2cedbf/markdown_it_py-4.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index fd475ed1ef..9cfa4137b7 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -47,12 +47,12 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.12.49-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.12.49.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.20.54-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.20.54.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.28.85-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.28.85.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.61", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.40.61-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.40.61.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.64", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.40.64-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.40.64.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.12.25-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.12.25.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.13.4-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.13.4.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-4.4.7-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-4.4.7.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.8", "version": "5.5.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-5.5.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-5.5.3.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.9", "version": "5.6.0b2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-5.6.0b2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-5.6.0b2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Celery", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=3.9", "version": "5.6.0rc1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-5.6.0rc1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-5.6.0rc1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "chalice", "requires_python": "", "version": "1.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "chalice-1.16.0-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "chalice-1.16.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "chalice", "requires_python": null, "version": "1.32.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "chalice-1.32.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "chalice-1.32.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: SQL", "Topic :: Database", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "clickhouse-driver", "requires_python": "<4,>=3.7", "version": "0.2.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-musllinux_1_1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp310-pypy310_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp37-pypy37_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp38-pypy38_pp73-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "clickhouse_driver-0.2.9-pp39-pypy39_pp73-win_amd64.whl"}, {"packagetype": "sdist", "filename": "clickhouse-driver-0.2.9.tar.gz"}]} @@ -67,13 +67,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "falcon-3.1.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "falcon-4.1.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.107.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.107.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.107.0.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.120.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.120.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.120.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.120.4.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.79.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.79.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.93.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.93.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.93.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.29.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.29.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.35.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.35.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.41.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.41.0.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.46.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.46.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.46.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.47.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.47.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.47.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-3.4.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-3.4.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.0.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.0.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.2.0b0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.2.0b0.tar.gz"}]} @@ -100,23 +100,23 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Scientific/Engineering :: Artificial Intelligence"], "name": "huggingface-hub", "requires_python": ">=3.9.0", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "huggingface_hub-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "huggingface_hub-1.0.1.tar.gz"}]} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9"], "name": "langchain", "requires_python": "<4.0,>=3.8.1", "version": "0.1.20", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-0.1.20-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-0.1.20.tar.gz"}]} {"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0,>=3.9", "version": "0.3.27", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-0.3.27-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-0.3.27.tar.gz"}]} -{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0.0,>=3.10.0", "version": "1.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-1.0.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-1.0.2.tar.gz"}]} +{"info": {"classifiers": [], "name": "langchain", "requires_python": "<4.0.0,>=3.10.0", "version": "1.0.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langchain-1.0.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langchain-1.0.3.tar.gz"}]} {"info": {"classifiers": [], "name": "langgraph", "requires_python": ">=3.9", "version": "0.6.11", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langgraph-0.6.11-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langgraph-0.6.11.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langgraph-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langgraph-1.0.1.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "launchdarkly_server_sdk-9.12.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "launchdarkly_server_sdk-9.12.2.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "langgraph", "requires_python": ">=3.10", "version": "1.0.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "langgraph-1.0.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "langgraph-1.0.2.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.9", "version": "9.12.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "launchdarkly_server_sdk-9.12.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "launchdarkly_server_sdk-9.12.3.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries"], "name": "launchdarkly-server-sdk", "requires_python": ">=3.8", "version": "9.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "launchdarkly_server_sdk-9.8.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "launchdarkly_server_sdk-9.8.1.tar.gz"}]} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.77.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.77.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.77.7.tar.gz"}]} {"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.78.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.78.7-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.78.7.tar.gz"}]} -{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.79.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.79.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.79.0.tar.gz"}]} +{"info": {"classifiers": ["License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "litellm", "requires_python": "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8", "version": "1.79.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litellm-1.79.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litellm-1.79.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": ">=3.8,<4.0", "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.0.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.12.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.12.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.12.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.18.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.18.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.18.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "litestar", "requires_python": "<4.0,>=3.8", "version": "2.6.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "litestar-2.6.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "litestar-2.6.4.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: System :: Logging"], "name": "loguru", "requires_python": "<4.0,>=3.5", "version": "0.7.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "loguru-0.7.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "loguru-0.7.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.15.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.15.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.15.0.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.16.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.16.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.17.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.17.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.17.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.19.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.19.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.19.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.20.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.0.1.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.103.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.103.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.109.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.109.1.tar.gz"}]} @@ -135,9 +135,9 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pure_eval-0.0.3.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pure_eval-0.2.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pure_eval-0.2.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.0.18-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.0.18.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.2.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.2.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.2.1.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.4.0.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.7.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.7.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.7.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.3.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.3.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.3.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.6.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.6.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.6.0.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.9.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.9.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.9.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}, "urls": [{"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.5-macosx-10.5-i386.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.6-macosx-10.5-i386.egg"}, {"packagetype": "sdist", "filename": "pymongo-0.6.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-cp32m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.8-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.9-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-2.8.1.tar.gz"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.5.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.4.exe"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.13.0-py2.7-macosx-10.14-intel.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.13.0.tar.gz"}]} @@ -161,7 +161,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.6", "version": "3.1.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-3.1.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.8", "version": "3.5.7", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-3.5.7.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Typing :: Typed"], "name": "pyspark", "requires_python": ">=3.9", "version": "4.0.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pyspark-4.0.1.tar.gz"}]} -{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.51.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp313-cp313-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.0-cp39-cp39-win_amd64.whl"}]} +{"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": ">=3.9", "version": "2.51.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp310-cp310-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp311-cp311-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp311-cp311-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp311-cp311-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp312-cp312-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp312-cp312-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp312-cp312-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp313-cp313-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp313-cp313-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp313-cp313-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp39-cp39-macosx_12_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.51.1-cp39-cp39-win_amd64.whl"}]} {"info": {"classifiers": ["Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "ray", "requires_python": "", "version": "2.7.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-macosx_10_15_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "ray-2.7.2-cp39-cp39-win_amd64.whl"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python"], "name": "redis", "requires_python": null, "version": "0.6.1", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "redis-0.6.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6"], "name": "redis", "requires_python": "", "version": "2.10.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "redis-2.10.6-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "redis-2.10.6.tar.gz"}]} @@ -203,11 +203,11 @@ {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.6", "version": "0.16.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.16.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.16.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.7", "version": "0.27.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.27.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.27.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.8", "version": "0.38.6", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.38.6-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.38.6.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.9", "version": "0.49.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.49.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.49.1.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: AnyIO", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Internet :: WWW/HTTP"], "name": "starlette", "requires_python": ">=3.10", "version": "0.50.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlette-0.50.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlette-0.50.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": ">=3.8,<4.0", "version": "1.48.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlite-1.48.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlite-1.48.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Typing :: Typed"], "name": "starlite", "requires_python": "<4.0,>=3.8", "version": "1.51.16", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "starlite-1.51.16-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "starlite-1.51.16.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.55.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "statsig-0.55.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "statsig-0.55.3.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.66.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "statsig-0.66.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "statsig-0.66.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries"], "name": "statsig", "requires_python": ">=3.7", "version": "0.66.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "statsig-0.66.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "statsig-0.66.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": ">=3.8,<4.0", "version": "0.209.8", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "strawberry_graphql-0.209.8-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "strawberry_graphql-0.209.8.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "strawberry-graphql", "requires_python": "<4.0,>=3.10", "version": "0.284.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "strawberry_graphql-0.284.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "strawberry_graphql-0.284.1.tar.gz"}]} {"info": {"classifiers": ["License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy"], "name": "tornado", "requires_python": ">= 3.5", "version": "6.0.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "tornado-6.0.4-cp38-cp38-win_amd64.whl"}, {"packagetype": "sdist", "filename": "tornado-6.0.4.tar.gz"}]} @@ -221,6 +221,6 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "5.8.16", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-5.8.16-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-5.8.16.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.6", "version": "6.2.14", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-6.2.14-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-6.2.14.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.8", "version": "6.8.17", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-6.8.17-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-6.8.17.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-7.6.9-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-7.6.9.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: No Input/Output (Daemon)", "Framework :: Tryton", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Natural Language :: Bulgarian", "Natural Language :: Catalan", "Natural Language :: Chinese (Simplified)", "Natural Language :: Czech", "Natural Language :: Dutch", "Natural Language :: English", "Natural Language :: Finnish", "Natural Language :: French", "Natural Language :: German", "Natural Language :: Hungarian", "Natural Language :: Indonesian", "Natural Language :: Italian", "Natural Language :: Persian", "Natural Language :: Polish", "Natural Language :: Portuguese (Brazilian)", "Natural Language :: Romanian", "Natural Language :: Russian", "Natural Language :: Slovenian", "Natural Language :: Spanish", "Natural Language :: Turkish", "Natural Language :: Ukrainian", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "trytond", "requires_python": ">=3.9", "version": "7.6.10", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "trytond-7.6.10-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "trytond-7.6.10.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.7", "version": "0.15.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "typer-0.15.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "typer-0.15.4.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "typer", "requires_python": ">=3.8", "version": "0.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "typer-0.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "typer-0.20.0.tar.gz"}]} diff --git a/tox.ini b/tox.ini index 9a663a12c7..b0779ce57b 100644 --- a/tox.ini +++ b/tox.ini @@ -60,7 +60,7 @@ envlist = {py3.9,py3.12,py3.13}-google_genai-v1.29.0 {py3.9,py3.12,py3.13}-google_genai-v1.35.0 {py3.9,py3.12,py3.13}-google_genai-v1.41.0 - {py3.9,py3.13,py3.14,py3.14t}-google_genai-v1.46.0 + {py3.9,py3.13,py3.14,py3.14t}-google_genai-v1.47.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.36.0 @@ -68,23 +68,23 @@ envlist = {py3.9,py3.11,py3.12}-langchain-base-v0.1.20 {py3.9,py3.12,py3.13}-langchain-base-v0.3.27 - {py3.10,py3.13,py3.14}-langchain-base-v1.0.2 + {py3.10,py3.13,py3.14}-langchain-base-v1.0.3 {py3.9,py3.11,py3.12}-langchain-notiktoken-v0.1.20 {py3.9,py3.12,py3.13}-langchain-notiktoken-v0.3.27 - {py3.10,py3.13,py3.14}-langchain-notiktoken-v1.0.2 + {py3.10,py3.13,py3.14}-langchain-notiktoken-v1.0.3 {py3.9,py3.13,py3.14}-langgraph-v0.6.11 - {py3.10,py3.12,py3.13}-langgraph-v1.0.1 + {py3.10,py3.12,py3.13}-langgraph-v1.0.2 {py3.9,py3.12,py3.13}-litellm-v1.77.7 {py3.9,py3.12,py3.13}-litellm-v1.78.7 - {py3.9,py3.12,py3.13}-litellm-v1.79.0 + {py3.9,py3.12,py3.13}-litellm-v1.79.1 {py3.10,py3.12,py3.13}-mcp-v1.15.0 - {py3.10,py3.12,py3.13}-mcp-v1.16.0 {py3.10,py3.12,py3.13}-mcp-v1.17.0 {py3.10,py3.12,py3.13}-mcp-v1.19.0 + {py3.10,py3.12,py3.13}-mcp-v1.20.0 {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 @@ -100,16 +100,16 @@ envlist = {py3.10,py3.12,py3.13}-openai_agents-v0.4.2 {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.18 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.2.1 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.4.0 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.7.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.3.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.6.0 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.9.1 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.13,py3.14,py3.14t}-boto3-v1.40.61 + {py3.9,py3.13,py3.14,py3.14t}-boto3-v1.40.64 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -144,13 +144,13 @@ envlist = # ~~~ Flags ~~~ {py3.8,py3.12,py3.13}-launchdarkly-v9.8.1 - {py3.9,py3.13,py3.14,py3.14t}-launchdarkly-v9.12.2 + {py3.9,py3.13,py3.14,py3.14t}-launchdarkly-v9.12.3 {py3.8,py3.13,py3.14,py3.14t}-openfeature-v0.7.5 {py3.9,py3.13,py3.14,py3.14t}-openfeature-v0.8.3 {py3.7,py3.13,py3.14}-statsig-v0.55.3 - {py3.7,py3.13,py3.14}-statsig-v0.66.0 + {py3.7,py3.13,py3.14}-statsig-v0.66.1 {py3.8,py3.12,py3.13}-unleash-v6.0.1 {py3.8,py3.12,py3.13}-unleash-v6.3.0 @@ -195,7 +195,7 @@ envlist = {py3.6,py3.7,py3.8}-celery-v4.4.7 {py3.8,py3.12,py3.13}-celery-v5.5.3 - {py3.9,py3.12,py3.13}-celery-v5.6.0b2 + {py3.9,py3.12,py3.13}-celery-v5.6.0rc1 {py3.6,py3.7}-dramatiq-v1.9.0 {py3.9,py3.12,py3.13}-dramatiq-v1.18.0 @@ -204,7 +204,7 @@ envlist = {py3.6,py3.11,py3.12}-huey-v2.5.4 {py3.9,py3.10}-ray-v2.7.2 - {py3.9,py3.12,py3.13}-ray-v2.51.0 + {py3.9,py3.12,py3.13}-ray-v2.51.1 {py3.6}-rq-v0.8.2 {py3.6,py3.7}-rq-v0.13.0 @@ -231,12 +231,12 @@ envlist = {py3.6,py3.9,py3.10}-starlette-v0.16.0 {py3.7,py3.10,py3.11}-starlette-v0.27.0 {py3.8,py3.12,py3.13}-starlette-v0.38.6 - {py3.9,py3.13,py3.14,py3.14t}-starlette-v0.49.1 + {py3.10,py3.13,py3.14,py3.14t}-starlette-v0.50.0 {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.93.0 {py3.8,py3.10,py3.11}-fastapi-v0.107.0 - {py3.8,py3.13,py3.14,py3.14t}-fastapi-v0.120.1 + {py3.8,py3.13,py3.14,py3.14t}-fastapi-v0.120.4 # ~~~ Web 2 ~~~ @@ -287,7 +287,7 @@ envlist = {py3.6}-trytond-v4.8.18 {py3.6,py3.7,py3.8}-trytond-v5.8.16 {py3.8,py3.10,py3.11}-trytond-v6.8.17 - {py3.9,py3.12,py3.13}-trytond-v7.6.9 + {py3.9,py3.12,py3.13}-trytond-v7.6.10 {py3.7,py3.12,py3.13}-typer-v0.15.4 {py3.8,py3.13,py3.14,py3.14t}-typer-v0.20.0 @@ -365,7 +365,7 @@ deps = google_genai-v1.29.0: google-genai==1.29.0 google_genai-v1.35.0: google-genai==1.35.0 google_genai-v1.41.0: google-genai==1.41.0 - google_genai-v1.46.0: google-genai==1.46.0 + google_genai-v1.47.0: google-genai==1.47.0 google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 @@ -376,34 +376,34 @@ deps = langchain-base-v0.1.20: langchain==0.1.20 langchain-base-v0.3.27: langchain==0.3.27 - langchain-base-v1.0.2: langchain==1.0.2 + langchain-base-v1.0.3: langchain==1.0.3 langchain-base: openai langchain-base: tiktoken langchain-base: langchain-openai langchain-base-v0.3.27: langchain-community - langchain-base-v1.0.2: langchain-community - langchain-base-v1.0.2: langchain-classic + langchain-base-v1.0.3: langchain-community + langchain-base-v1.0.3: langchain-classic langchain-notiktoken-v0.1.20: langchain==0.1.20 langchain-notiktoken-v0.3.27: langchain==0.3.27 - langchain-notiktoken-v1.0.2: langchain==1.0.2 + langchain-notiktoken-v1.0.3: langchain==1.0.3 langchain-notiktoken: openai langchain-notiktoken: langchain-openai langchain-notiktoken-v0.3.27: langchain-community - langchain-notiktoken-v1.0.2: langchain-community - langchain-notiktoken-v1.0.2: langchain-classic + langchain-notiktoken-v1.0.3: langchain-community + langchain-notiktoken-v1.0.3: langchain-classic langgraph-v0.6.11: langgraph==0.6.11 - langgraph-v1.0.1: langgraph==1.0.1 + langgraph-v1.0.2: langgraph==1.0.2 litellm-v1.77.7: litellm==1.77.7 litellm-v1.78.7: litellm==1.78.7 - litellm-v1.79.0: litellm==1.79.0 + litellm-v1.79.1: litellm==1.79.1 mcp-v1.15.0: mcp==1.15.0 - mcp-v1.16.0: mcp==1.16.0 mcp-v1.17.0: mcp==1.17.0 mcp-v1.19.0: mcp==1.19.0 + mcp-v1.20.0: mcp==1.20.0 mcp: pytest-asyncio openai-base-v1.0.1: openai==1.0.1 @@ -426,9 +426,9 @@ deps = openai_agents: pytest-asyncio pydantic_ai-v1.0.18: pydantic-ai==1.0.18 - pydantic_ai-v1.2.1: pydantic-ai==1.2.1 - pydantic_ai-v1.4.0: pydantic-ai==1.4.0 - pydantic_ai-v1.7.0: pydantic-ai==1.7.0 + pydantic_ai-v1.3.0: pydantic-ai==1.3.0 + pydantic_ai-v1.6.0: pydantic-ai==1.6.0 + pydantic_ai-v1.9.1: pydantic-ai==1.9.1 pydantic_ai: pytest-asyncio @@ -436,7 +436,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.61: boto3==1.40.61 + boto3-v1.40.64: boto3==1.40.64 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -480,13 +480,13 @@ deps = # ~~~ Flags ~~~ launchdarkly-v9.8.1: launchdarkly-server-sdk==9.8.1 - launchdarkly-v9.12.2: launchdarkly-server-sdk==9.12.2 + launchdarkly-v9.12.3: launchdarkly-server-sdk==9.12.3 openfeature-v0.7.5: openfeature-sdk==0.7.5 openfeature-v0.8.3: openfeature-sdk==0.8.3 statsig-v0.55.3: statsig==0.55.3 - statsig-v0.66.0: statsig==0.66.0 + statsig-v0.66.1: statsig==0.66.1 statsig: typing_extensions unleash-v6.0.1: UnleashClient==6.0.1 @@ -556,7 +556,7 @@ deps = celery-v4.4.7: celery==4.4.7 celery-v5.5.3: celery==5.5.3 - celery-v5.6.0b2: celery==5.6.0b2 + celery-v5.6.0rc1: celery==5.6.0rc1 celery: newrelic<10.17.0 celery: redis {py3.7}-celery: importlib-metadata<5.0 @@ -568,7 +568,7 @@ deps = huey-v2.5.4: huey==2.5.4 ray-v2.7.2: ray==2.7.2 - ray-v2.51.0: ray==2.51.0 + ray-v2.51.1: ray==2.51.1 rq-v0.8.2: rq==0.8.2 rq-v0.13.0: rq==0.13.0 @@ -626,7 +626,7 @@ deps = starlette-v0.16.0: starlette==0.16.0 starlette-v0.27.0: starlette==0.27.0 starlette-v0.38.6: starlette==0.38.6 - starlette-v0.49.1: starlette==0.49.1 + starlette-v0.50.0: starlette==0.50.0 starlette: pytest-asyncio starlette: python-multipart starlette: requests @@ -640,7 +640,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.93.0: fastapi==0.93.0 fastapi-v0.107.0: fastapi==0.107.0 - fastapi-v0.120.1: fastapi==0.120.1 + fastapi-v0.120.4: fastapi==0.120.4 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -735,7 +735,7 @@ deps = trytond-v4.8.18: trytond==4.8.18 trytond-v5.8.16: trytond==5.8.16 trytond-v6.8.17: trytond==6.8.17 - trytond-v7.6.9: trytond==7.6.9 + trytond-v7.6.10: trytond==7.6.10 trytond: werkzeug trytond-v4.6.22: werkzeug<1.0 trytond-v4.8.18: werkzeug<1.0 From 7264a9f7881a75e876bdac9fb0c632b83732ef57 Mon Sep 17 00:00:00 2001 From: Neel Shah Date: Mon, 3 Nov 2025 14:26:14 +0100 Subject: [PATCH 590/591] Fix NOT_GIVEN check in anthropic (#5058) same as #4926 #### Issues * resolves: #5057 * resolves: PY-1959 --- sentry_sdk/integrations/anthropic.py | 26 ++++++++++++++++++++++---- 1 file changed, 22 insertions(+), 4 deletions(-) diff --git a/sentry_sdk/integrations/anthropic.py b/sentry_sdk/integrations/anthropic.py index e61a3556e1..e252ab2424 100644 --- a/sentry_sdk/integrations/anthropic.py +++ b/sentry_sdk/integrations/anthropic.py @@ -22,9 +22,14 @@ try: try: - from anthropic import NOT_GIVEN + from anthropic import NotGiven except ImportError: - NOT_GIVEN = None + NotGiven = None + + try: + from anthropic import Omit + except ImportError: + Omit = None from anthropic.resources import AsyncMessages, Messages @@ -168,12 +173,13 @@ def _set_input_data(span, kwargs, integration): } for key, attribute in kwargs_keys_to_attributes.items(): value = kwargs.get(key) - if value is not NOT_GIVEN and value is not None: + + if value is not None and _is_given(value): set_data_normalized(span, attribute, value) # Input attributes: Tools tools = kwargs.get("tools") - if tools is not NOT_GIVEN and tools is not None and len(tools) > 0: + if tools is not None and _is_given(tools) and len(tools) > 0: set_data_normalized( span, SPANDATA.GEN_AI_REQUEST_AVAILABLE_TOOLS, safe_serialize(tools) ) @@ -419,3 +425,15 @@ async def _sentry_patched_create_async(*args, **kwargs): span.__exit__(None, None, None) return _sentry_patched_create_async + + +def _is_given(obj): + # type: (Any) -> bool + """ + Check for givenness safely across different anthropic versions. + """ + if NotGiven is not None and isinstance(obj, NotGiven): + return False + if Omit is not None and isinstance(obj, Omit): + return False + return True From e910302c5d4d9f10a31f0cd32739655379c08d7d Mon Sep 17 00:00:00 2001 From: Simon Hellmayr Date: Tue, 4 Nov 2025 11:13:24 +0100 Subject: [PATCH 591/591] feat(integrations): add ability to auto-deactivate lower-level integrations based on map (#5052) - Add a map that contains integrations that should be deactivated when others are active to prevent duplicate spans, etc. - Allow positive and negative overrides by users - Add the existing langchain deactivations for anthropic and openai - Simplifies setup for users, because they don't have to add the lower level integrations to the disabled integrations array in the init --------- Co-authored-by: Alexander Alderman Webb --- .github/workflows/test-integrations-misc.yml | 4 + .../populate_tox/package_dependencies.jsonl | 16 +- scripts/populate_tox/populate_tox.py | 1 + scripts/populate_tox/releases.jsonl | 18 +- scripts/populate_tox/tox.jinja | 9 + .../split_tox_gh_actions.py | 1 + sentry_sdk/integrations/__init__.py | 37 +++ tests/test_ai_integration_deactivation.py | 222 ++++++++++++++++++ tox.ini | 33 ++- 9 files changed, 312 insertions(+), 29 deletions(-) create mode 100644 tests/test_ai_integration_deactivation.py diff --git a/.github/workflows/test-integrations-misc.yml b/.github/workflows/test-integrations-misc.yml index 5e10dc8b6c..10cccaebac 100644 --- a/.github/workflows/test-integrations-misc.yml +++ b/.github/workflows/test-integrations-misc.yml @@ -74,6 +74,10 @@ jobs: run: | set -x # print commands that are executed ./scripts/runtox.sh "py${{ matrix.python-version }}-typer" + - name: Test integration_deactivation + run: | + set -x # print commands that are executed + ./scripts/runtox.sh "py${{ matrix.python-version }}-integration_deactivation" - name: Generate coverage XML (Python 3.6) if: ${{ !cancelled() && matrix.python-version == '3.6' }} run: | diff --git a/scripts/populate_tox/package_dependencies.jsonl b/scripts/populate_tox/package_dependencies.jsonl index d467fc125c..254f42bbe1 100644 --- a/scripts/populate_tox/package_dependencies.jsonl +++ b/scripts/populate_tox/package_dependencies.jsonl @@ -1,18 +1,18 @@ -{"name": "boto3", "version": "1.40.64", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/65/c2/27da558ceb90d17b1e4c0cca5dab29f8aea7f63242a1005a8f54230ce5e6/boto3-1.40.64-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8f/c5/70bec18aef3fe9af63847d8766f81864b20daacd1dc7bf0c1d1ad90c7e98/botocore-1.40.64-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/f0/ae7ca09223a81a1d890b2557186ea015f6e0502e9b8cb8e1813f1d8cfa4e/s3transfer-0.14.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}]} +{"name": "boto3", "version": "1.40.65", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/d5/a2/973a1346bb8d23ddda241342e126a26407e08648a556fb5022e0db457860/boto3-1.40.65-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/00/ccd1612ee99865435ad04ef477168af9d3a8d2ec6a7a694a37af47143543/botocore-1.40.65-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/31/b4/b9b800c45527aadd64d5b442f9b932b00648617eb5d63d2c7a6587b7cafc/jmespath-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/f0/ae7ca09223a81a1d890b2557186ea015f6e0502e9b8cb8e1813f1d8cfa4e/s3transfer-0.14.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}]} {"name": "django", "version": "6.0b1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/0e/1a/306fda7e62e27ccbcb92d97f67f1094352a9f22c62f3c2b238fa50eb82d7/django-6.0b1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/17/9c/fc2331f538fbf7eedba64b2052e99ccf9ba9d6888e2f41441ee28847004b/asgiref-3.10.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a9/5c/bfd6bd0bf979426d405cc6e71eceb8701b148b16c21d2dc3c261efc61c7b/sqlparse-0.5.3-py3-none-any.whl"}}]} -{"name": "fastapi", "version": "0.120.4", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/ed/47/14a76b926edc3957c8a8258423db789d3fa925d2fed800102fce58959413/fastapi-0.120.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a3/e0/021c772d6a662f43b63044ab481dc6ac7592447605b5b35a957785363122/starlette-0.49.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/02/b7/cf592cb5de5cb3bade3357f8d2cf42bf103bbe39f459824b4939fd212911/annotated_doc-0.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "fastapi", "version": "0.121.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/dd/2c/42277afc1ba1a18f8358561eee40785d27becab8f80a1f945c0a3051c6eb/fastapi-0.121.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a3/e0/021c772d6a662f43b63044ab481dc6ac7592447605b5b35a957785363122/starlette-0.49.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/02/b7/cf592cb5de5cb3bade3357f8d2cf42bf103bbe39f459824b4939fd212911/annotated_doc-0.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} {"name": "flask", "version": "2.3.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/fd/56/26f0be8adc2b4257df20c1c4260ddd0aa396cf8e75d90ab2f7ff99bc34f9/flask-2.3.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} {"name": "flask", "version": "3.1.2", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}]} -{"name": "google-genai", "version": "1.47.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/89/ef/e080e8d67c270ea320956bb911a9359664fc46d3b87d1f029decd33e5c4c/google_genai-1.47.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/92/05/adeb6c495aec4f9d93f9e2fc29eeef6e14d452bba11d15bdb874ce1d5b10/google_auth-2.42.1-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/96/c5/1e741d26306c42e2bf6ab740b2202872727e0f606033c9dd713f8b93f5a8/cachetools-6.2.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/8d/0133e4eb4beed9e425d9a98ed6e081a55d195481b7632472be1af08d2f6b/rsa-4.9.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fa/a8/5b41e0da817d64113292ab1f8247140aac61cbf6cfd085d6a0fa77f4984f/websockets-15.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c8/f1/d6a797abb14f6283c0ddff96bbdd46937f64122b8c925cab503dd37f8214/pyasn1-0.6.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/47/8d/d529b5d697919ba8c11ad626e835d4039be708a35b0d22de83a269a6682c/pyasn1_modules-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} -{"name": "langchain", "version": "1.0.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/68/c8/b5dcfdde8b96369e5445f0fbac52fe8495bbd11b23ca83691d90d464eb15/langchain-1.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/54/3aed89938a42cf7115575c333647551e35adc380feed651105d2d86c22f5/langchain_core-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d7/b1/9f4912e13d4ed691f2685c8a4b764b5a9237a30cca0c5782bc213d9f0a9a/langgraph-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/27/2f/9a7d00d4afa036e65294059c7c912002fb72ba5dbbd5c2a871ca06360278/langgraph_prebuilt-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1f/38/9a97f650b8cdb2ba0356d65aef9239f4a30db69ae44c30daa2cf8dd3f350/langsmith-0.4.39-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/29/b53a9ca6cd366bfc928823679c6a76c7a4c69f8201c0ba7903ad18ebae2f/pydantic_core-2.41.4-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b1/5e/f77dc6b9036943e285ba76b49e118d9ea929885becb0a29ba8a7c75e29fe/pyyaml-6.0.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/97/86/e5b50247a61caec5718122feb2719ea9d451d30ac0516c288c1dbc6408e8/ormsgpack-1.11.0-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8c/63/8ffc2cc97e811c0ca5d00ab36604b3ea6f4254f20b7bc658ca825ce6c954/xxhash-3.6.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} -{"name": "langgraph", "version": "0.6.11", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/df/94/430f0341c5c2fe3e3b9f5ab2622f35e2bda12c4a7d655c519468e853d1b0/langgraph-0.6.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8e/d1/e4727f4822943befc3b7046f79049b1086c9493a34b4d44a1adf78577693/langgraph_prebuilt-0.6.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c0/54/3cdbe9d151d06cd689b5aa937ac11403b64bbfe76486fda6431a24061721/langchain_core-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b4/2b/7e0248f65e35800ea8e4e3dbb3bcc36c61b81f5b8abeddaceec8320ab491/langsmith-0.4.38-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} -{"name": "launchdarkly-server-sdk", "version": "9.12.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/5a/c0/8c0bc0584ac8c01288c09f14eb8002a2ebe433d6901f0d978c605c51ca8d/launchdarkly_server_sdk-9.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/95/c20ba9cd28951f360b222da09138daf9169ad951be2a0d500eb0f0c84614/launchdarkly_eventsource-1.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/cb/84/a04c59324445f4bcc98dc05b39a1cd07c242dde643c1a3c21e4f7beaf2f2/expiringdict-1.2.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/34/90/0200184d2124484f918054751ef997ed6409cb05b7e8dcbf5a22da4c4748/pyrfc3339-2.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl"}}]} +{"name": "google-genai", "version": "1.48.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/25/7d/a5c02159099546ec01131059294d1f0174eee33a872fc888c6c99e8cd6d9/google_genai-1.48.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/92/05/adeb6c495aec4f9d93f9e2fc29eeef6e14d452bba11d15bdb874ce1d5b10/google_auth-2.42.1-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/96/c5/1e741d26306c42e2bf6ab740b2202872727e0f606033c9dd713f8b93f5a8/cachetools-6.2.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/8d/0133e4eb4beed9e425d9a98ed6e081a55d195481b7632472be1af08d2f6b/rsa-4.9.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fa/a8/5b41e0da817d64113292ab1f8247140aac61cbf6cfd085d6a0fa77f4984f/websockets-15.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c8/f1/d6a797abb14f6283c0ddff96bbdd46937f64122b8c925cab503dd37f8214/pyasn1-0.6.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/47/8d/d529b5d697919ba8c11ad626e835d4039be708a35b0d22de83a269a6682c/pyasn1_modules-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}]} +{"name": "langchain", "version": "1.0.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/68/c8/b5dcfdde8b96369e5445f0fbac52fe8495bbd11b23ca83691d90d464eb15/langchain-1.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/f2/1b/b0a37674bdcbd2931944e12ea742fd167098de5212ee2391e91dce631162/langchain_core-1.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d7/b1/9f4912e13d4ed691f2685c8a4b764b5a9237a30cca0c5782bc213d9f0a9a/langgraph-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/27/2f/9a7d00d4afa036e65294059c7c912002fb72ba5dbbd5c2a871ca06360278/langgraph_prebuilt-1.0.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b3/8a/de38127895f185f677778930478c6884b087c2cab5744dd7e27c0188fbec/langsmith-0.4.40-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "langgraph", "version": "0.6.11", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/df/94/430f0341c5c2fe3e3b9f5ab2622f35e2bda12c4a7d655c519468e853d1b0/langgraph-0.6.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/85/2a/2efe0b5a72c41e3a936c81c5f5d8693987a1b260287ff1bbebaae1b7b888/langgraph_checkpoint-3.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/8e/d1/e4727f4822943befc3b7046f79049b1086c9493a34b4d44a1adf78577693/langgraph_prebuilt-0.6.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/66/05/b2d34e16638241e6f27a6946d28160d4b8b641383787646d41a3727e0896/langgraph_sdk-0.2.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/f2/1b/b0a37674bdcbd2931944e12ea742fd167098de5212ee2391e91dce631162/langchain_core-1.0.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/73/07/02e16ed01e04a374e644b575638ec7987ae846d25ad97bcc9945a3ee4b0e/jsonpatch-1.33-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b3/8a/de38127895f185f677778930478c6884b087c2cab5744dd7e27c0188fbec/langsmith-0.4.40-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a1/6b/83661fa77dcefa195ad5f8cd9af3d1a7450fd57cc883ad04d65446ac2029/pydantic-2.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0d/c2/472f2e31b95eff099961fa050c376ab7156a81da194f9edb9f710f68787b/pydantic_core-2.41.4-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e5/30/643397144bfbfec6f6ef821f36f33e57d35946c44a2352d3c9f0ae847619/tenacity-9.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c6/fe/ed708782d6709cc60eb4c2d8a361a440661f74134675c72990f2c48c785f/orjson-3.11.4.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/19/2b/776d1b411d2be50f77a6e6e94a25825cca55dcacfe7415fd691a144db71b/ormsgpack-1.11.0-cp314-cp314t-macosx_10_12_x86_64.macosx_11_0_arm64.macosx_10_12_universal2.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/3f/51/d4db610ef29373b879047326cbf6fa98b6c1969d6f6dc423279de2b1be2c/requests_toolbelt-1.0.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/fd/aa/3e0508d5a5dd96529cdc5a97011299056e14c6505b678fd58938792794b1/zstandard-0.25.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} +{"name": "launchdarkly-server-sdk", "version": "9.12.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/5a/c0/8c0bc0584ac8c01288c09f14eb8002a2ebe433d6901f0d978c605c51ca8d/launchdarkly_server_sdk-9.12.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/71/51/b1af4ab7302dbdf1eb13e5cfb2f3dce776d786fb93e91de5de56f60ca814/launchdarkly_eventsource-1.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/cb/84/a04c59324445f4bcc98dc05b39a1cd07c242dde643c1a3c21e4f7beaf2f2/expiringdict-1.2.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/34/90/0200184d2124484f918054751ef997ed6409cb05b7e8dcbf5a22da4c4748/pyrfc3339-2.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl"}}]} {"name": "openfeature-sdk", "version": "0.7.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9b/20/cb043f54b11505d993e4dd84652cfc44c1260dc94b7f41aa35489af58277/openfeature_sdk-0.7.5-py3-none-any.whl"}}]} {"name": "openfeature-sdk", "version": "0.8.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/f0/f5/707a5b144115de1a49bf5761a63af2545fef0a1824f72db39ddea0a3438f/openfeature_sdk-0.8.3-py3-none-any.whl"}}]} {"name": "quart", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/7e/e9/cc28f21f52913adf333f653b9e0a3bf9cb223f5083a26422968ba73edd8d/quart-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/10/cb/f2ad4230dc2eb1a74edf38f1a38b9b52277f75bef262d8908e60d957e13c/blinker-1.9.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/f9/7f9263c5695f4bd0023734af91bedb2ff8209e8de6ead162f35d8dc762fd/flask-3.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/3b/dfa13a8d96aa24e40ea74a975a9906cfdc2ab2f4e3b498862a57052f04eb/hypercorn-0.17.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/69/b2/119f6e6dcbd96f9069ce9a2665e0146588dc9f88f29549711853645e736a/h2-4.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/c6/80c95b1b2b94682a72cbdbfb85b81ae2daffa4291fbfa1b1464502ede10d/hpack-4.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/48/30/47d0bf6072f7252e6521f3447ccfa40b421b6824517f82854703d0f5a98b/hyperframe-6.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/96/92447566d16df59b2a776c0fb82dbc4d9e07cd95062562af01e408583fc4/itsdangerous-2.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/78/58/e860788190eba3bcce367f74d29c4675466ce8dddfba85f7827588416f01/wsproto-1.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/bc/8a/340a1555ae33d7354dbca4faa54948d76d89a27ceef032c8c3bc661d003e/aiofiles-25.1.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/5e/5f/82c8074f7e84978129347c2c6ec8b6c59f3584ff1a20bc3c940a3e061790/priority-2.0.0-py3-none-any.whl"}}]} {"name": "requests", "version": "2.32.5", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}]} {"name": "starlette", "version": "0.50.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/d9/52/1064f510b141bd54025f9b55105e26d1fa970b9be67ad766380a3c9b74b0/starlette-0.50.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/15/b3/9b1a8074496371342ec1e796a96f99c82c945a339cd81a8e73de28b4cf9e/anyio-4.11.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl"}}]} {"name": "statsig", "version": "0.55.3", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/9d/17/de62fdea8aab8aa7c4a833378e0e39054b728dfd45ef279e975ed5ef4e86/statsig-0.55.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} -{"name": "statsig", "version": "0.66.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/75/cf/06d818a72e489c4d5aec4399ef4ee69777ba2cb73ad9a64fdeed19b149a1/statsig-0.66.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2f/c2/f9e977608bdf958650638c3f1e28f85a1b075f075ebbe77db8555463787b/Brotli-1.1.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/64/78/63a0bcc0707037df4e22bb836451279d850592258c859685a402c27f5d6d/ijson-3.4.0.post0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} -{"name": "strawberry-graphql", "version": "0.284.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/eb/9d/03a4225721f82dc8d25de894ae0b29ede0c5710de88a1710a8c081c88a4c/strawberry_graphql-0.284.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ae/4f/7297663840621022bc73c22d7d9d80dbc78b4db6297f764b545cd5dd462d/graphql_core-3.2.6-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/00/f2/c68a97c727c795119f1056ad2b7e716c23f26f004292517c435accf90b5c/lia_web-0.2.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} +{"name": "statsig", "version": "0.66.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/75/cf/06d818a72e489c4d5aec4399ef4ee69777ba2cb73ad9a64fdeed19b149a1/statsig-0.66.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2f/c2/f9e977608bdf958650638c3f1e28f85a1b075f075ebbe77db8555463787b/Brotli-1.1.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/2a/97/e88295f9456ba939d90d4603af28fcabda3b443ef55e709e9381df3daa58/ijson-3.4.0.post0-cp314-cp314t-macosx_11_0_arm64.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/d6/b7/48a7f1ab9eee62f1113114207df7e7e6bc29227389d554f42cc11bc98108/ip3country-0.4.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/07/d1/0a28c21707807c6aacd5dc9c3704b2aa1effbf37adebd8caeaf68b17a636/protobuf-6.33.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e4/37/af0d2ef3967ac0d6113837b44a4f0bfe1328c2b9763bd5b1744520e5cfed/certifi-2025.10.5-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/37/be6dfbfa45719aa82c008fb4772cfe5c46db765a2ca4b6f524a1fdfee4d7/ua_parser-1.0.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/6f/d3/13adff37f15489c784cc7669c35a6c3bf94b87540229eedf52ef2a1d0175/ua_parser_builtins-0.18.0.post1-py3-none-any.whl"}}]} +{"name": "strawberry-graphql", "version": "0.284.1", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/eb/9d/03a4225721f82dc8d25de894ae0b29ede0c5710de88a1710a8c081c88a4c/strawberry_graphql-0.284.1-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/0a/14/933037032608787fb92e365883ad6a741c235e0ff992865ec5d904a38f1e/graphql_core-3.2.7-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/00/f2/c68a97c727c795119f1056ad2b7e716c23f26f004292517c435accf90b5c/lia_web-0.2.3-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} {"name": "typer", "version": "0.20.0", "dependencies": [{"download_info": {"url": "https://files.pythonhosted.org/packages/78/64/7713ffe4b5983314e9d436a90d5bd4f63b6054e2aca783a3cfc44cb95bbf/typer-0.20.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/db/d3/9dcc0f5797f070ec8edf30fbadfb200e71d9db6b84d211e3b2085a7589a0/click-8.3.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/25/7a/b0178788f8dc6cafce37a212c99565fa1fe7872c70c6c9c1e1a372d9d88f/rich-14.2.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/94/54/e7d793b573f298e1c9013b8c4dade17d481164aa517d1d7148619c2cedbf/markdown_it_py-4.0.0-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl"}}, {"download_info": {"url": "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl"}}]} diff --git a/scripts/populate_tox/populate_tox.py b/scripts/populate_tox/populate_tox.py index 0700a82637..2d81a85ea2 100644 --- a/scripts/populate_tox/populate_tox.py +++ b/scripts/populate_tox/populate_tox.py @@ -64,6 +64,7 @@ "aws_lambda", "cloud_resource_context", "common", + "integration_deactivation", "gcp", "gevent", "opentelemetry", diff --git a/scripts/populate_tox/releases.jsonl b/scripts/populate_tox/releases.jsonl index 9cfa4137b7..8d57c183bf 100644 --- a/scripts/populate_tox/releases.jsonl +++ b/scripts/populate_tox/releases.jsonl @@ -47,7 +47,7 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7"], "name": "boto3", "requires_python": "", "version": "1.12.49", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.12.49-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.12.49.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.6", "version": "1.20.54", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.20.54-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.20.54.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">= 3.7", "version": "1.28.85", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.28.85-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.28.85.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.64", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.40.64-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.40.64.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9"], "name": "boto3", "requires_python": ">=3.9", "version": "1.40.65", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "boto3-1.40.65-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "boto3-1.40.65.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": "", "version": "0.12.25", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.12.25-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.12.25.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet :: WWW/HTTP :: Dynamic Content :: CGI Tools/Libraries", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Internet :: WWW/HTTP :: WSGI :: Application", "Topic :: Internet :: WWW/HTTP :: WSGI :: Middleware", "Topic :: Internet :: WWW/HTTP :: WSGI :: Server", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "bottle", "requires_python": null, "version": "0.13.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "bottle-0.13.4-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "bottle-0.13.4.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Object Brokering", "Topic :: System :: Distributed Computing"], "name": "celery", "requires_python": ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*", "version": "4.4.7", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "celery-4.4.7-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "celery-4.4.7.tar.gz"}]} @@ -67,13 +67,13 @@ {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks"], "name": "falcon", "requires_python": ">=3.5", "version": "3.1.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-macosx_10_14_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-macosx_11_0_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-3.1.3-cp39-cp39-win_amd64.whl"}, {"packagetype": "sdist", "filename": "falcon-3.1.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Cython", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Internet :: WWW/HTTP :: WSGI", "Topic :: Software Development :: Libraries :: Application Frameworks", "Typing :: Typed"], "name": "falcon", "requires_python": ">=3.8", "version": "4.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp312-cp312-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp313-cp313-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_10_13_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-macosx_11_0_arm64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-musllinux_1_2_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-cp314-cp314-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "falcon-4.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "falcon-4.1.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.107.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.107.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.107.0.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.120.4", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.120.4-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.120.4.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Framework :: Pydantic", "Framework :: Pydantic :: 1", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.8", "version": "0.121.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.121.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.121.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.6.1", "version": "0.79.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.79.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.79.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: AsyncIO", "Framework :: FastAPI", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "fastapi", "requires_python": ">=3.7", "version": "0.93.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "fastapi-0.93.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "fastapi-0.93.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.29.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.29.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.29.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.35.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.35.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.35.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.41.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.41.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.41.0.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.9", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.9", "version": "1.47.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.47.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.47.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Internet", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "google-genai", "requires_python": ">=3.10", "version": "1.48.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "google_genai-1.48.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "google_genai-1.48.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": "", "version": "3.4.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-3.4.1-py2.py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-3.4.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.0.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.0.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.0.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Software Development :: Libraries"], "name": "gql", "requires_python": ">=3.8.1", "version": "4.2.0b0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "gql-4.2.0b0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "gql-4.2.0b0.tar.gz"}]} @@ -118,13 +118,13 @@ {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.19.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.19.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.19.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13"], "name": "mcp", "requires_python": ">=3.10", "version": "1.20.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "mcp-1.20.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "mcp-1.20.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.7.1", "version": "1.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.0.1.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.103.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.103.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.103.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.104.2", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.104.2-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.104.2.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.109.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.109.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.109.1.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.59.9", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.59.9-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.59.9.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.88.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.88.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.88.0.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.0.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.0.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.0.1.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.4.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.4.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.4.0.tar.gz"}]} -{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.6.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.6.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.6.1.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "1.89.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-1.89.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-1.89.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.1.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.5.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.5.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.5.0.tar.gz"}]} +{"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai", "requires_python": ">=3.8", "version": "2.7.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai-2.7.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai-2.7.1.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.0.19", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.0.19-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.0.19.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.1.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.1.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.1.0.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.9", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed"], "name": "openai-agents", "requires_python": ">=3.9", "version": "0.2.11", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "openai_agents-0.2.11-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "openai_agents-0.2.11.tar.gz"}]} @@ -135,9 +135,9 @@ {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8"], "name": "pure-eval", "requires_python": "", "version": "0.0.3", "yanked": false}, "urls": [{"packagetype": "sdist", "filename": "pure_eval-0.0.3.tar.gz"}]} {"info": {"classifiers": ["Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9"], "name": "pure-eval", "requires_python": null, "version": "0.2.3", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pure_eval-0.2.3-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pure_eval-0.2.3.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.0.18", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.0.18-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.0.18.tar.gz"}]} +{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.10.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.10.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.10.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.3.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.3.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.3.0.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.6.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.6.0-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.6.0.tar.gz"}]} -{"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Framework :: Pydantic", "Framework :: Pydantic :: 2", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Internet", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules"], "name": "pydantic-ai", "requires_python": ">=3.10", "version": "1.9.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pydantic_ai-1.9.1-py3-none-any.whl"}, {"packagetype": "sdist", "filename": "pydantic_ai-1.9.1.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "0.6", "yanked": false}, "urls": [{"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.5-macosx-10.5-i386.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-0.6-py2.6-macosx-10.5-i386.egg"}, {"packagetype": "sdist", "filename": "pymongo-0.6.tar.gz"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.1", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: Jython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": null, "version": "2.8.1", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp26-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_10_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_8_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-macosx_10_9_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp27-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-cp32m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp32-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-cp33m-macosx_10_6_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp33-none-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-cp34m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-2.8.1-cp34-none-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.8-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-macosx-10.9-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.5-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.6-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.10-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.8-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-macosx-10.9-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py2.7-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.2-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-macosx-10.6-x86_64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.3-win-amd64.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-macosx-10.6-intel.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win32.egg"}, {"packagetype": "bdist_egg", "filename": "pymongo-2.8.1-py3.4-win-amd64.egg"}, {"packagetype": "sdist", "filename": "pymongo-2.8.1.tar.gz"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.5.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win32-py3.4.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.6.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py2.7.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.2.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.3.exe"}, {"packagetype": "bdist_wininst", "filename": "pymongo-2.8.1.win-amd64-py3.4.exe"}]} {"info": {"classifiers": ["Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Database"], "name": "pymongo", "requires_python": "", "version": "3.13.0", "yanked": false}, "urls": [{"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp27-cp27m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp310-cp310-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp311-cp311-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp35-cp35m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp36-cp36m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp37-cp37m-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp38-cp38-win_amd64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win32.whl"}, {"packagetype": "bdist_wheel", "filename": "pymongo-3.13.0-cp39-cp39-win_amd64.whl"}, {"packagetype": "bdist_egg", "filename": "pymongo-3.13.0-py2.7-macosx-10.14-intel.egg"}, {"packagetype": "sdist", "filename": "pymongo-3.13.0.tar.gz"}]} diff --git a/scripts/populate_tox/tox.jinja b/scripts/populate_tox/tox.jinja index 4de4d94b5f..a04087ddfd 100755 --- a/scripts/populate_tox/tox.jinja +++ b/scripts/populate_tox/tox.jinja @@ -23,6 +23,9 @@ envlist = # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent + # === Integration Deactivation === + {py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-integration_deactivation + # === Integrations === # Asgi @@ -88,6 +91,11 @@ deps = {py3.10,py3.11}-gevent: zope.event<5.0.0 {py3.10,py3.11}-gevent: zope.interface<8.0 + # === Integration Deactivation === + integration_deactivation: openai + integration_deactivation: anthropic + integration_deactivation: langchain + # === Integrations === # Asgi @@ -144,6 +152,7 @@ setenv = # TESTPATH definitions for test suites not managed by toxgen common: TESTPATH=tests gevent: TESTPATH=tests + integration_deactivation: TESTPATH=tests/test_ai_integration_deactivation.py asgi: TESTPATH=tests/integrations/asgi aws_lambda: TESTPATH=tests/integrations/aws_lambda cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context diff --git a/scripts/split_tox_gh_actions/split_tox_gh_actions.py b/scripts/split_tox_gh_actions/split_tox_gh_actions.py index 9f9369d4ab..59c3473d8c 100755 --- a/scripts/split_tox_gh_actions/split_tox_gh_actions.py +++ b/scripts/split_tox_gh_actions/split_tox_gh_actions.py @@ -154,6 +154,7 @@ "pure_eval", "trytond", "typer", + "integration_deactivation", ], } diff --git a/sentry_sdk/integrations/__init__.py b/sentry_sdk/integrations/__init__.py index af4206e208..1de0b83c06 100644 --- a/sentry_sdk/integrations/__init__.py +++ b/sentry_sdk/integrations/__init__.py @@ -171,6 +171,11 @@ def iter_default_integrations(with_auto_enabling_integrations): } +_INTEGRATION_DEACTIVATES = { + "langchain": {"openai", "anthropic"}, +} + + def setup_integrations( integrations, # type: Sequence[Integration] with_defaults=True, # type: bool @@ -187,6 +192,15 @@ def setup_integrations( `disabled_integrations` takes precedence over `with_defaults` and `with_auto_enabling_integrations`. + + Some integrations are designed to automatically deactivate other integrations + in order to avoid conflicts and prevent duplicate telemetry from being collected. + For example, enabling the `langchain` integration will auto-deactivate both the + `openai` and `anthropic` integrations. + + Users can override this behavior by: + - Explicitly providing an integration in the `integrations=[]` list, or + - Disabling the higher-level integration via the `disabled_integrations` option. """ integrations = dict( (integration.identifier, integration) for integration in integrations or () @@ -194,6 +208,8 @@ def setup_integrations( logger.debug("Setting up integrations (with default = %s)", with_defaults) + user_provided_integrations = set(integrations.keys()) + # Integrations that will not be enabled disabled_integrations = [ integration if isinstance(integration, type) else type(integration) @@ -212,6 +228,27 @@ def setup_integrations( integrations[instance.identifier] = instance used_as_default_integration.add(instance.identifier) + disabled_integration_identifiers = { + integration.identifier for integration in disabled_integrations + } + + for integration, targets_to_deactivate in _INTEGRATION_DEACTIVATES.items(): + if ( + integration in integrations + and integration not in disabled_integration_identifiers + ): + for target in targets_to_deactivate: + if target not in user_provided_integrations: + for cls in iter_default_integrations(True): + if cls.identifier == target: + if cls not in disabled_integrations: + disabled_integrations.append(cls) + logger.debug( + "Auto-deactivating %s integration because %s integration is active", + target, + integration, + ) + for identifier, integration in integrations.items(): with _installer_lock: if identifier not in _processed_integrations: diff --git a/tests/test_ai_integration_deactivation.py b/tests/test_ai_integration_deactivation.py new file mode 100644 index 0000000000..71ac4f70a9 --- /dev/null +++ b/tests/test_ai_integration_deactivation.py @@ -0,0 +1,222 @@ +import pytest + +from sentry_sdk import get_client +from sentry_sdk.integrations import _INTEGRATION_DEACTIVATES + + +try: + from sentry_sdk.integrations.langchain import LangchainIntegration + + has_langchain = True +except Exception: + has_langchain = False + +try: + from sentry_sdk.integrations.openai import OpenAIIntegration + + has_openai = True +except Exception: + has_openai = False + +try: + from sentry_sdk.integrations.anthropic import AnthropicIntegration + + has_anthropic = True +except Exception: + has_anthropic = False + + +pytestmark = pytest.mark.skipif( + not (has_langchain and has_openai and has_anthropic), + reason="Requires langchain, openai, and anthropic packages to be installed", +) + + +def test_integration_deactivates_map_exists(): + assert "langchain" in _INTEGRATION_DEACTIVATES + assert "openai" in _INTEGRATION_DEACTIVATES["langchain"] + assert "anthropic" in _INTEGRATION_DEACTIVATES["langchain"] + + +def test_langchain_auto_deactivates_openai_and_anthropic( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + if LangchainIntegration in integration_types: + assert OpenAIIntegration not in integration_types + assert AnthropicIntegration not in integration_types + + +def test_user_can_override_with_explicit_openai(sentry_init, reset_integrations): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + integrations=[OpenAIIntegration()], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert OpenAIIntegration in integration_types + + +def test_user_can_override_with_explicit_anthropic(sentry_init, reset_integrations): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + integrations=[AnthropicIntegration()], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert AnthropicIntegration in integration_types + + +def test_user_can_override_with_both_explicit_integrations( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + integrations=[OpenAIIntegration(), AnthropicIntegration()], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert OpenAIIntegration in integration_types + assert AnthropicIntegration in integration_types + + +def test_disabling_langchain_allows_openai_and_anthropic( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + disabled_integrations=[LangchainIntegration], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert LangchainIntegration not in integration_types + + +def test_explicit_langchain_still_deactivates_others(sentry_init, reset_integrations): + sentry_init( + default_integrations=False, + auto_enabling_integrations=False, + integrations=[LangchainIntegration()], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + if LangchainIntegration in integration_types: + assert OpenAIIntegration not in integration_types + assert AnthropicIntegration not in integration_types + + +def test_langchain_and_openai_both_explicit_both_active( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=False, + auto_enabling_integrations=False, + integrations=[LangchainIntegration(), OpenAIIntegration()], + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert LangchainIntegration in integration_types + assert OpenAIIntegration in integration_types + + +def test_no_langchain_means_openai_and_anthropic_can_auto_enable( + sentry_init, reset_integrations, monkeypatch +): + import sys + import sentry_sdk.integrations + + old_iter = sentry_sdk.integrations.iter_default_integrations + + def filtered_iter(with_auto_enabling): + for cls in old_iter(with_auto_enabling): + if cls.identifier != "langchain": + yield cls + + monkeypatch.setattr( + sentry_sdk.integrations, "iter_default_integrations", filtered_iter + ) + + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + assert LangchainIntegration not in integration_types + + +def test_deactivation_with_default_integrations_enabled( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=True, + auto_enabling_integrations=True, + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + if LangchainIntegration in integration_types: + assert OpenAIIntegration not in integration_types + assert AnthropicIntegration not in integration_types + + +def test_only_auto_enabling_integrations_without_defaults( + sentry_init, reset_integrations +): + sentry_init( + default_integrations=False, + auto_enabling_integrations=True, + ) + + client = get_client() + integration_types = { + type(integration) for integration in client.integrations.values() + } + + if LangchainIntegration in integration_types: + assert OpenAIIntegration not in integration_types + assert AnthropicIntegration not in integration_types diff --git a/tox.ini b/tox.ini index b0779ce57b..7a1fd39f03 100644 --- a/tox.ini +++ b/tox.ini @@ -23,6 +23,9 @@ envlist = # === Gevent === {py3.6,py3.8,py3.10,py3.11,py3.12}-gevent + # === Integration Deactivation === + {py3.9,py3.10,py3.11,py3.12,py3.13,py3.14}-integration_deactivation + # === Integrations === # Asgi @@ -60,7 +63,7 @@ envlist = {py3.9,py3.12,py3.13}-google_genai-v1.29.0 {py3.9,py3.12,py3.13}-google_genai-v1.35.0 {py3.9,py3.12,py3.13}-google_genai-v1.41.0 - {py3.9,py3.13,py3.14,py3.14t}-google_genai-v1.47.0 + {py3.10,py3.13,py3.14,py3.14t}-google_genai-v1.48.0 {py3.8,py3.10,py3.11}-huggingface_hub-v0.24.7 {py3.8,py3.12,py3.13}-huggingface_hub-v0.36.0 @@ -88,11 +91,11 @@ envlist = {py3.8,py3.11,py3.12}-openai-base-v1.0.1 {py3.8,py3.12,py3.13}-openai-base-v1.109.1 - {py3.9,py3.12,py3.13}-openai-base-v2.6.1 + {py3.9,py3.12,py3.13}-openai-base-v2.7.1 {py3.8,py3.11,py3.12}-openai-notiktoken-v1.0.1 {py3.8,py3.12,py3.13}-openai-notiktoken-v1.109.1 - {py3.9,py3.12,py3.13}-openai-notiktoken-v2.6.1 + {py3.9,py3.12,py3.13}-openai-notiktoken-v2.7.1 {py3.10,py3.11,py3.12}-openai_agents-v0.0.19 {py3.10,py3.12,py3.13}-openai_agents-v0.1.0 @@ -102,14 +105,14 @@ envlist = {py3.10,py3.12,py3.13}-pydantic_ai-v1.0.18 {py3.10,py3.12,py3.13}-pydantic_ai-v1.3.0 {py3.10,py3.12,py3.13}-pydantic_ai-v1.6.0 - {py3.10,py3.12,py3.13}-pydantic_ai-v1.9.1 + {py3.10,py3.12,py3.13}-pydantic_ai-v1.10.0 # ~~~ Cloud ~~~ {py3.6,py3.7}-boto3-v1.12.49 {py3.6,py3.9,py3.10}-boto3-v1.20.54 {py3.7,py3.11,py3.12}-boto3-v1.28.85 - {py3.9,py3.13,py3.14,py3.14t}-boto3-v1.40.64 + {py3.9,py3.13,py3.14,py3.14t}-boto3-v1.40.65 {py3.6,py3.7,py3.8}-chalice-v1.16.0 {py3.9,py3.12,py3.13}-chalice-v1.32.0 @@ -236,7 +239,7 @@ envlist = {py3.6,py3.9,py3.10}-fastapi-v0.79.1 {py3.7,py3.10,py3.11}-fastapi-v0.93.0 {py3.8,py3.10,py3.11}-fastapi-v0.107.0 - {py3.8,py3.13,py3.14,py3.14t}-fastapi-v0.120.4 + {py3.8,py3.13,py3.14,py3.14t}-fastapi-v0.121.0 # ~~~ Web 2 ~~~ @@ -325,6 +328,11 @@ deps = {py3.10,py3.11}-gevent: zope.event<5.0.0 {py3.10,py3.11}-gevent: zope.interface<8.0 + # === Integration Deactivation === + integration_deactivation: openai + integration_deactivation: anthropic + integration_deactivation: langchain + # === Integrations === # Asgi @@ -365,7 +373,7 @@ deps = google_genai-v1.29.0: google-genai==1.29.0 google_genai-v1.35.0: google-genai==1.35.0 google_genai-v1.41.0: google-genai==1.41.0 - google_genai-v1.47.0: google-genai==1.47.0 + google_genai-v1.48.0: google-genai==1.48.0 google_genai: pytest-asyncio huggingface_hub-v0.24.7: huggingface_hub==0.24.7 @@ -408,14 +416,14 @@ deps = openai-base-v1.0.1: openai==1.0.1 openai-base-v1.109.1: openai==1.109.1 - openai-base-v2.6.1: openai==2.6.1 + openai-base-v2.7.1: openai==2.7.1 openai-base: pytest-asyncio openai-base: tiktoken openai-base-v1.0.1: httpx<0.28 openai-notiktoken-v1.0.1: openai==1.0.1 openai-notiktoken-v1.109.1: openai==1.109.1 - openai-notiktoken-v2.6.1: openai==2.6.1 + openai-notiktoken-v2.7.1: openai==2.7.1 openai-notiktoken: pytest-asyncio openai-notiktoken-v1.0.1: httpx<0.28 @@ -428,7 +436,7 @@ deps = pydantic_ai-v1.0.18: pydantic-ai==1.0.18 pydantic_ai-v1.3.0: pydantic-ai==1.3.0 pydantic_ai-v1.6.0: pydantic-ai==1.6.0 - pydantic_ai-v1.9.1: pydantic-ai==1.9.1 + pydantic_ai-v1.10.0: pydantic-ai==1.10.0 pydantic_ai: pytest-asyncio @@ -436,7 +444,7 @@ deps = boto3-v1.12.49: boto3==1.12.49 boto3-v1.20.54: boto3==1.20.54 boto3-v1.28.85: boto3==1.28.85 - boto3-v1.40.64: boto3==1.40.64 + boto3-v1.40.65: boto3==1.40.65 {py3.7,py3.8}-boto3: urllib3<2.0.0 chalice-v1.16.0: chalice==1.16.0 @@ -640,7 +648,7 @@ deps = fastapi-v0.79.1: fastapi==0.79.1 fastapi-v0.93.0: fastapi==0.93.0 fastapi-v0.107.0: fastapi==0.107.0 - fastapi-v0.120.4: fastapi==0.120.4 + fastapi-v0.121.0: fastapi==0.121.0 fastapi: httpx fastapi: pytest-asyncio fastapi: python-multipart @@ -760,6 +768,7 @@ setenv = # TESTPATH definitions for test suites not managed by toxgen common: TESTPATH=tests gevent: TESTPATH=tests + integration_deactivation: TESTPATH=tests/test_ai_integration_deactivation.py asgi: TESTPATH=tests/integrations/asgi aws_lambda: TESTPATH=tests/integrations/aws_lambda cloud_resource_context: TESTPATH=tests/integrations/cloud_resource_context