commit:     77e68fac13e82b9e434144c8fb427c6acca9ef9a
Author:     Huang Rui <vowstar <AT> gmail <DOT> com>
AuthorDate: Mon Apr  7 13:42:51 2025 +0000
Commit:     Rui Huang <vowstar <AT> gmail <DOT> com>
CommitDate: Mon Apr  7 14:26:54 2025 +0000
URL:        https://gitweb.gentoo.org/repo/proj/guru.git/commit/?id=77e68fac

dev-python/scrapy: drop 2.11.0-r1

Signed-off-by: Huang Rui <vowstar <AT> gmail.com>

 dev-python/scrapy/Manifest                         |  1 -
 .../scrapy-2.11.0-lift-twisted-restriction.patch   | 99 ----------------------
 dev-python/scrapy/scrapy-2.11.0-r1.ebuild          | 61 -------------
 3 files changed, 161 deletions(-)

diff --git a/dev-python/scrapy/Manifest b/dev-python/scrapy/Manifest
index f5bd91867..37c0254c0 100644
--- a/dev-python/scrapy/Manifest
+++ b/dev-python/scrapy/Manifest
@@ -1,2 +1 @@
-DIST scrapy-2.11.0.tar.gz 1373184 BLAKE2B 
7fc758a729d9e9f88c89e30c503ff8bf4c2e111bc00cbd003d5f8785dc818e5ef69ae3774013c4fe27db623c1246a57d57db097cf5220b257a9632bdd29b7023
 SHA512 
08da9684f4312e4149929f3785cc516207070fdd0eb99d8b7ed3276095378cc71d18534d0535fd1107af442aed42d6c91792ad7749d631b5f54522c3912ce073
 DIST scrapy-2.12.0.tar.gz 1210998 BLAKE2B 
f0bd637ac4182de4573d37fb6db8ee26aac20eeb9ded6d255a9abe246014a100e1ba567e80f64279d382954e18bab7af165a69f2a7732619b2e4369693ea8d1d
 SHA512 
3d733d2cb858e0f6201d08e5e6388667155a3e9746cca6887652b1ce514d803da8c6523d1bba27d451a5e4337fc191983de46c2ad8e6a6b0479f38ebfa1ad3b1

diff --git 
a/dev-python/scrapy/files/scrapy-2.11.0-lift-twisted-restriction.patch 
b/dev-python/scrapy/files/scrapy-2.11.0-lift-twisted-restriction.patch
deleted file mode 100644
index c3ef23287..000000000
--- a/dev-python/scrapy/files/scrapy-2.11.0-lift-twisted-restriction.patch
+++ /dev/null
@@ -1,99 +0,0 @@
-Backport commits 
720f351a3eea5e5bfa83a6eaf50210cd1fa43992^..0630e4aaa10c3fb8c79c2542a229f5c0632cddde.
-At the time of writing, ::gentoo no longer has <dev-python/twisted-23.8.0.
-These commits refactor the signal handling logic to support newer versions of 
Twisted.
-
-Author: Lucio Sauer <[email protected]>
---- a/scrapy/crawler.py
-+++ b/scrapy/crawler.py
-@@ -404,8 +404,8 @@ class CrawlerProcess(CrawlerRunner):
-         :param bool stop_after_crawl: stop or not the reactor when all
-             crawlers have finished
- 
--        :param bool install_signal_handlers: whether to install the shutdown
--            handlers (default: True)
-+        :param bool install_signal_handlers: whether to install the OS signal
-+            handlers from Twisted and Scrapy (default: True)
-         """
-         from twisted.internet import reactor
- 
-@@ -416,15 +416,17 @@ class CrawlerProcess(CrawlerRunner):
-                 return
-             d.addBoth(self._stop_reactor)
- 
--        if install_signal_handlers:
--            install_shutdown_handlers(self._signal_shutdown)
-         resolver_class = load_object(self.settings["DNS_RESOLVER"])
-         resolver = create_instance(resolver_class, self.settings, self, 
reactor=reactor)
-         resolver.install_on_reactor()
-         tp = reactor.getThreadPool()
-         
tp.adjustPoolsize(maxthreads=self.settings.getint("REACTOR_THREADPOOL_MAXSIZE"))
-         reactor.addSystemEventTrigger("before", "shutdown", self.stop)
--        reactor.run(installSignalHandlers=False)  # blocking call
-+        if install_signal_handlers:
-+            reactor.addSystemEventTrigger(
-+                "after", "startup", install_shutdown_handlers, 
self._signal_shutdown
-+            )
-+        reactor.run(installSignalHandlers=install_signal_handlers)  # 
blocking call
- 
-     def _graceful_stop_reactor(self) -> Deferred:
-         d = self.stop()
---- a/scrapy/utils/ossignal.py
-+++ b/scrapy/utils/ossignal.py
-@@ -19,13 +19,10 @@ def install_shutdown_handlers(
-     function: SignalHandlerT, override_sigint: bool = True
- ) -> None:
-     """Install the given function as a signal handler for all common shutdown
--    signals (such as SIGINT, SIGTERM, etc). If override_sigint is ``False`` 
the
--    SIGINT handler won't be install if there is already a handler in place
--    (e.g.  Pdb)
-+    signals (such as SIGINT, SIGTERM, etc). If ``override_sigint`` is 
``False`` the
-+    SIGINT handler won't be installed if there is already a handler in place
-+    (e.g. Pdb)
-     """
--    from twisted.internet import reactor
--
--    reactor._handleSignals()
-     signal.signal(signal.SIGTERM, function)
-     if signal.getsignal(signal.SIGINT) == signal.default_int_handler or 
override_sigint:
-         signal.signal(signal.SIGINT, function)
---- a/scrapy/utils/testproc.py
-+++ b/scrapy/utils/testproc.py
-@@ -2,7 +2,7 @@ from __future__ import annotations
- 
- import os
- import sys
--from typing import Iterable, Optional, Tuple, cast
-+from typing import Iterable, List, Optional, Tuple, cast
- 
- from twisted.internet.defer import Deferred
- from twisted.internet.error import ProcessTerminated
-@@ -26,14 +26,15 @@ class ProcessTest:
-         env = os.environ.copy()
-         if settings is not None:
-             env["SCRAPY_SETTINGS_MODULE"] = settings
-+        assert self.command
-         cmd = self.prefix + [self.command] + list(args)
-         pp = TestProcessProtocol()
--        pp.deferred.addBoth(self._process_finished, cmd, check_code)
-+        pp.deferred.addCallback(self._process_finished, cmd, check_code)
-         reactor.spawnProcess(pp, cmd[0], cmd, env=env, path=self.cwd)
-         return pp.deferred
- 
-     def _process_finished(
--        self, pp: TestProcessProtocol, cmd: str, check_code: bool
-+        self, pp: TestProcessProtocol, cmd: List[str], check_code: bool
-     ) -> Tuple[int, bytes, bytes]:
-         if pp.exitcode and check_code:
-             msg = f"process {cmd} exit with code {pp.exitcode}"
---- a/setup.py
-+++ b/setup.py
-@@ -6,8 +6,7 @@ version = (Path(__file__).parent / 
"scrapy/VERSION").read_text("ascii").strip()
- 
- 
- install_requires = [
--    # 23.8.0 incompatibility: https://github.com/scrapy/scrapy/issues/6024
--    "Twisted>=18.9.0,<23.8.0",
-+    "Twisted>=18.9.0",
-     "cryptography>=36.0.0",
-     "cssselect>=0.9.1",
-     "itemloaders>=1.0.1",

diff --git a/dev-python/scrapy/scrapy-2.11.0-r1.ebuild 
b/dev-python/scrapy/scrapy-2.11.0-r1.ebuild
deleted file mode 100644
index 999638aa3..000000000
--- a/dev-python/scrapy/scrapy-2.11.0-r1.ebuild
+++ /dev/null
@@ -1,61 +0,0 @@
-# Copyright 1999-2023 Gentoo Authors
-# Distributed under the terms of the GNU General Public License v2
-
-EAPI=8
-
-PYTHON_COMPAT=( python3_{11..12} )
-DISTUTILS_USE_PEP517=setuptools
-inherit distutils-r1
-
-DESCRIPTION="A high-level Web Crawling and Web Scraping framework"
-HOMEPAGE="https://scrapy.org/";
-SRC_URI="https://github.com/scrapy/scrapy/archive/refs/tags/${PV}.tar.gz -> 
${P}.tar.gz"
-
-LICENSE="BSD"
-SLOT=0
-KEYWORDS="~amd64"
-IUSE="test"
-RESTRICT="!test? ( test )"
-
-# The 'PyDispatcher>=2.0.5' distribution was not found and is required by 
Scrapy
-# https://bugs.gentoo.org/684734
-RDEPEND="dev-python/cssselect[${PYTHON_USEDEP}]
-       dev-python/cryptography[${PYTHON_USEDEP}]
-       <dev-python/priority-2.0.0[${PYTHON_USEDEP}]
-       dev-python/h2[${PYTHON_USEDEP}]
-       dev-python/itemadapter[${PYTHON_USEDEP}]
-       dev-python/itemloaders[${PYTHON_USEDEP}]
-       dev-python/lxml[${PYTHON_USEDEP}]
-       dev-python/parsel[${PYTHON_USEDEP}]
-       dev-python/protego[${PYTHON_USEDEP}]
-       >=dev-python/pydispatcher-2.0.5[${PYTHON_USEDEP}]
-       dev-python/pyopenssl[${PYTHON_USEDEP}]
-       dev-python/queuelib[${PYTHON_USEDEP}]
-       dev-python/service-identity[${PYTHON_USEDEP}]
-       dev-python/six[${PYTHON_USEDEP}]
-       dev-python/tldextract[${PYTHON_USEDEP}]
-       >=dev-python/twisted-18.9.0[${PYTHON_USEDEP}]
-       dev-python/w3lib[${PYTHON_USEDEP}]
-       dev-python/zope-interface[${PYTHON_USEDEP}]
-"
-BDEPEND="
-       test? (
-               ${RDEPEND}
-               dev-python/testfixtures[${PYTHON_USEDEP}]
-               dev-python/uvloop[${PYTHON_USEDEP}]
-       )
-"
-
-PATCHES="${FILESDIR}"/${P}-lift-twisted-restriction.patch
-
-distutils_enable_tests pytest
-
-EPYTEST_DESELECT=(
-       # these require (local) network access
-       tests/test_command_check.py
-       tests/test_feedexport.py
-       tests/test_pipeline_files.py::TestFTPFileStore::test_persist
-       # Flaky test: https://github.com/scrapy/scrapy/issues/6193
-       tests/test_crawl.py::CrawlTestCase::test_start_requests_laziness
-       )
-EPYTEST_IGNORE=( docs )

Reply via email to