packages: perl-WWW-RobotRules/perl-WWW-RobotRules.spec (NEW) - new
qboosh
qboosh at pld-linux.org
Sun Mar 13 08:11:53 CET 2011
Author: qboosh Date: Sun Mar 13 07:11:53 2011 GMT
Module: packages Tag: HEAD
---- Log message:
- new
---- Files affected:
packages/perl-WWW-RobotRules:
perl-WWW-RobotRules.spec (NONE -> 1.1) (NEW)
---- Diffs:
================================================================
Index: packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec
diff -u /dev/null packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec:1.1
--- /dev/null Sun Mar 13 08:11:53 2011
+++ packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec Sun Mar 13 08:11:48 2011
@@ -0,0 +1,85 @@
+# $Revision$, $Date$
+#
+# Conditional build:
+%bcond_without tests # do not perform "make test"
+#
+%define pdir WWW
+%define pnam RobotRules
+%include /usr/lib/rpm/macros.perl
+Summary: WWW::RobotRules - database of robots.txt-derived permissions
+Summary(pl.UTF-8): WWW::RobotRules - baza danych uprawnień z robots.txt
+Name: perl-WWW-RobotRules
+Version: 6.00
+Release: 1
+# same as perl
+License: GPL v1+ or Artistic
+Group: Development/Languages/Perl
+Source0: http://www.cpan.org/modules/by-module/WWW/%{pdir}-%{pnam}-%{version}.tar.gz
+# Source0-md5: 4b5f4838bf1de44057b0bd6240b05dd7
+URL: http://search.cpan.org/dist/WWW-RobotRules/
+BuildRequires: perl-devel >= 1:5.8.8
+BuildRequires: rpm-perlprov >= 4.1-13
+%if %{with tests}
+BuildRequires: perl-URI >= 1.10
+%endif
+Conflicts: perl-libwww < 6
+BuildArch: noarch
+BuildRoot: %{tmpdir}/%{name}-%{version}-root-%(id -u -n)
+
+%description
+This module parses /robots.txt files as specified in "A Standard for
+Robot Exclusion", at <http://www.robotstxt.org/wc/norobots.html>.
+Webmasters can use the /robots.txt file to forbid conforming robots
+from accessing parts of their web site.
+
+The parsed files are kept in a WWW::RobotRules object, and this object
+provides methods to check if access to a given URL is prohibited. The
+same WWW::RobotRules object can be used for one or more parsed
+/robots.txt files on any number of hosts.
+
+%description -l pl.UTF-8
+Ten moduł analizuje pliki /robots.txt opisane w dokumencie "A Standard
+for Robot Exclusion" (<http://www.robotstxt.org/wc/norobots.html>).
+Webmasterzy mogą używać pliku /robots.txt, aby zabronić automatom
+zgodnym z tym standardem dostępu do pewnych plików na stronie WWW.
+
+Przeanalizowane pliki są trzymane w obiekcie WWW::RobotRules, a obiekt
+ten udostępnia metody do sprawdzania, czy dostęp do danego URL-a jest
+zabroniony. Ten sam obiekt WWW::RobotRules może być używany dla
+jednego lub większej liczby przeanalizowanych plików /robots.txt z
+różnych hostów.
+
+%prep
+%setup -q -n %{pdir}-%{pnam}-%{version}
+
+%build
+%{__perl} Makefile.PL \
+ INSTALLDIRS=vendor
+%{__make}
+
+%{?with_tests:%{__make} test}
+
+%install
+rm -rf $RPM_BUILD_ROOT
+
+%{__make} pure_install \
+ DESTDIR=$RPM_BUILD_ROOT
+
+%clean
+rm -rf $RPM_BUILD_ROOT
+
+%files
+%defattr(644,root,root,755)
+%doc Changes README
+%{perl_vendorlib}/WWW/RobotRules.pm
+%{perl_vendorlib}/WWW/RobotRules
+%{_mandir}/man3/WWW::RobotRules*.3pm*
+
+%define date %(echo `LC_ALL="C" date +"%a %b %d %Y"`)
+%changelog
+* %{date} PLD Team <feedback at pld-linux.org>
+All persons listed below can be reached at <cvs_login>@pld-linux.org
+
+$Log$
+Revision 1.1 2011/03/13 07:11:48 qboosh
+- new
================================================================
More information about the pld-cvs-commit
mailing list