head	1.1;
access;
symbols
	pkgsrc-2026Q1:1.1.0.24
	pkgsrc-2026Q1-base:1.1
	pkgsrc-2025Q4:1.1.0.22
	pkgsrc-2025Q4-base:1.1
	pkgsrc-2025Q3:1.1.0.20
	pkgsrc-2025Q3-base:1.1
	pkgsrc-2025Q2:1.1.0.18
	pkgsrc-2025Q2-base:1.1
	pkgsrc-2025Q1:1.1.0.16
	pkgsrc-2025Q1-base:1.1
	pkgsrc-2024Q4:1.1.0.14
	pkgsrc-2024Q4-base:1.1
	pkgsrc-2024Q3:1.1.0.12
	pkgsrc-2024Q3-base:1.1
	pkgsrc-2024Q2:1.1.0.10
	pkgsrc-2024Q2-base:1.1
	pkgsrc-2024Q1:1.1.0.8
	pkgsrc-2024Q1-base:1.1
	pkgsrc-2023Q4:1.1.0.6
	pkgsrc-2023Q4-base:1.1
	pkgsrc-2023Q3:1.1.0.4
	pkgsrc-2023Q3-base:1.1
	pkgsrc-2023Q2:1.1.0.2
	pkgsrc-2023Q2-base:1.1;
locks; strict;
comment	@# @;


1.1
date	2023.05.10.21.59.57;	author markd;	state Exp;
branches;
next	;
commitid	mpnXhGCCRcl2VsoE;


desc
@@


1.1
log
@py-robot-detection: add version 0.4

robot_detection is a python module to detect if a given HTTP User Agent
is a web crawler. It uses the list of registered robots from
http://www.robotstxt.org: (Robots Database)
@
text
@# $NetBSD$

DISTNAME=	robot-detection-0.4
PKGNAME=	${PYPKGPREFIX}-${DISTNAME}
CATEGORIES=	www python
MASTER_SITES=	${MASTER_SITE_PYPI:=r/robot-detection/}

MAINTAINER=	pkgsrc-users@@NetBSD.org
HOMEPAGE=	https://github.com/rory/robot-detection/
COMMENT=	Detect web crawlers using HTTP User Agent
LICENSE=	gnu-gpl-v3

DEPENDS+=	${PYPKGPREFIX}-six-[0-9]*:../../lang/py-six

USE_LANGUAGES=	# none

PYTHON_VERSIONS_INCOMPATIBLE=	27

.include "../../lang/python/egg.mk"
.include "../../mk/bsd.pkg.mk"
@
