head 1.3; access; symbols pkgsrc-2026Q1:1.3.0.8 pkgsrc-2026Q1-base:1.3 pkgsrc-2025Q4:1.3.0.6 pkgsrc-2025Q4-base:1.3 pkgsrc-2025Q3:1.3.0.4 pkgsrc-2025Q3-base:1.3 pkgsrc-2025Q2:1.3.0.2 pkgsrc-2025Q2-base:1.3 pkgsrc-2025Q1:1.2.0.16 pkgsrc-2025Q1-base:1.2 pkgsrc-2024Q4:1.2.0.14 pkgsrc-2024Q4-base:1.2 pkgsrc-2024Q3:1.2.0.12 pkgsrc-2024Q3-base:1.2 pkgsrc-2024Q2:1.2.0.10 pkgsrc-2024Q2-base:1.2 pkgsrc-2024Q1:1.2.0.8 pkgsrc-2024Q1-base:1.2 pkgsrc-2023Q4:1.2.0.6 pkgsrc-2023Q4-base:1.2 pkgsrc-2023Q3:1.2.0.4 pkgsrc-2023Q3-base:1.2 pkgsrc-2023Q2:1.2.0.2 pkgsrc-2023Q2-base:1.2 pkgsrc-2023Q1:1.1.0.2 pkgsrc-2023Q1-base:1.1; locks; strict; comment @# @; 1.3 date 2025.06.06.07.51.53; author wiz; state Exp; branches; next 1.2; commitid uuU2IdUq244yRNXF; 1.2 date 2023.04.25.14.55.28; author wiz; state Exp; branches; next 1.1; commitid f59XpOPTnf7r3vmE; 1.1 date 2023.03.13.14.18.27; author wiz; state Exp; branches; next ; commitid po2eibnr1pNreYgE; desc @@ 1.3 log @{py-,}sentencepiece: update to 0.2.0 @ text @# $NetBSD: Makefile,v 1.2 2023/04/25 14:55:28 wiz Exp $ PKGNAME= ${PYPKGPREFIX}-${DISTNAME} GITHUB_PROJECT= sentencepiece WRKSRC= ${WRKDIR}/${DISTNAME}/python USE_TOOLS+= cmake pkg-config TOOL_DEPENDS+= ${PYPKGPREFIX}-setuptools-[0-9]*:../../devel/py-setuptools pre-configure: ${ECHO} '#!/bin/sh' > ${BUILDLINK_DIR}/bin/nproc ${ECHO} 'echo 1' > ${BUILDLINK_DIR}/bin/nproc ${CHMOD} 555 ${BUILDLINK_DIR}/bin/nproc .include "../../textproc/sentencepiece/Makefile.common" .include "../../textproc/sentencepiece/buildlink3.mk" .include "../../lang/python/wheel.mk" .include "../../mk/bsd.pkg.mk" @ 1.2 log @py-sentencepiece: not for python 2 @ text @d1 1 a1 1 # $NetBSD: Makefile,v 1.1 2023/03/13 14:18:27 wiz Exp $ d4 1 d8 1 a8 1 USE_TOOLS+= pkg-config d10 6 a15 1 PYTHON_VERSIONS_INCOMPATIBLE= 27 d19 1 a19 1 .include "../../lang/python/egg.mk" @ 1.1 log @textproc/py-sentencepiece: import py-sentencepiece-0.1.97 SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabulary size is predetermined prior to the neural model training. SentencePiece implements subword units (e.g., byte-pair-encoding (BPE)) and unigram language model with the extension of direct training from raw sentences. SentencePiece allows us to make a purely end-to-end system that does not depend on language-specific pre/postprocessing. This package contains the Python module. @ text @d1 1 a1 1 # $NetBSD$ d9 2 @