BDEPEND=test? ( dev-python/six[python_targets_python3_10(-)?,python_targets_python3_11(-)?] dev-python/unittest-or-fail[python_targets_python3_10(-)?,python_targets_python3_11(-)?] ) python_targets_python3_10? ( >=dev-lang/python-3.10.9-r1:3.10 ) python_targets_python3_11? ( >=dev-lang/python-3.11.1-r1:3.11 ) >=dev-python/setuptools-65.7.0[python_targets_python3_10(-)?,python_targets_python3_11(-)?] DEFINED_PHASES=compile configure install prepare test DESCRIPTION=Module to detect if a given HTTP User Agent is a web crawler EAPI=8 HOMEPAGE=https://github.com/rory/robot-detection INHERIT=distutils-r1 IUSE=test python_targets_python3_10 python_targets_python3_11 KEYWORDS=~amd64 LICENSE=GPL-3 RDEPEND=dev-python/six[python_targets_python3_10(-)?,python_targets_python3_11(-)?] python_targets_python3_10? ( >=dev-lang/python-3.10.9-r1:3.10 ) python_targets_python3_11? ( >=dev-lang/python-3.11.1-r1:3.11 ) REQUIRED_USE=|| ( python_targets_python3_10 python_targets_python3_11 ) RESTRICT=!test? ( test ) SLOT=0 SRC_URI=https://github.com/rory/robot-detection/archive/v0.4.0.tar.gz -> robot-detection-0.4.0.tar.gz _eclasses_=out-of-source-utils 1a9007554652a6e627edbccb3c25a439 multibuild bddcb51b74f4a76724ff7cf8e7388869 multilib d1408425c7c4a7669b9b17735404b693 toolchain-funcs 14a8ae365191b518fad51caad7a08f3e multiprocessing b4e253ab22cef7b1085e9b67c7a3b730 ninja-utils 76050953ad5b70d7e09a6ca55558db92 python-utils-r1 4544ab96b58e177dc52d75e42f5ee307 python-r1 3c6cd0f418ba702c186a9865b85e704d distutils-r1 c377e6aa4e8a33d679b05d0441586b86 _md5_=b8e0e183779254e693c4874c529fbd5e