./net/wget, Retrieve files from the net via HTTP and FTP

[ CVSweb ] [ Homepage ] [ RSS ] [ Required by ] [ Add to tracker ]

Branch: CURRENT, Version: 1.19.1nb3, Package name: wget-1.19.1nb3, Maintainer: pkgsrc-users

Wget [formerly known as Geturl] is a freely available network utility
to retrieve files from the World Wide Web using HTTP and FTP, the two
most widely used Internet protocols. It works non-interactively, thus
enabling work in the background, after having logged off.

The recursive retrieval of HTML pages, as well as FTP sites is
supported -- you can use Wget to make mirrors of archives and home
pages, or traverse the web like a WWW robot (Wget understands

Required to run:

Required to build:

Package options: idn, inet6, openssl

Master sites: (Expand)

SHA1: cde25e99c144191644406793cbd1c69c102c6970
RMD160: 158d759b81c0893cc9c83e4beabb104f4987f6dd
Filesize: 2062.262 KB

Version history: (Expand)

CVS history: (Expand)

   2017-11-14 10:51:13 by Leonardo Taccari | Files touched by this commit (2) | Package updated
Log message:
wget: Use devel/libidn2 and adjust `idn' option logic in options.mk

Since wget-1.19, libidn2 is needed for the IDN/IRIs support. Adjust
the `idn' package option logic to reflect that and explicitly ask
for it via CONFIGURE_ARGS. This should also fix the build without
the `idn' option selected pointed out by john heasley via PR pkg/52726.

   2017-10-26 17:01:39 by Tim Zingelman | Files touched by this commit (4)
Log message:
wget: patches for VE-2017-13089 and CVE-2017-13090
   2017-05-15 07:10:09 by Kimmo Suominen | Files touched by this commit (2)
Log message:
Add a patch for CVE-2017-6508 from upstream.
   2017-02-20 13:25:48 by Ryo ONODERA | Files touched by this commit (4) | Package updated
Log message:
Update to 1.19.1

* Changes in Wget 1.19.1

* Fix bugs, a regression, portability/build issues

* Add new option --retry-on-http-error

* Changes in Wget 1.19

* New option --use-askpass=COMMAND. Fetch user/password by calling
  an external program.

* Use IDNA2008 (+ TR46 if available) through libidn2

* When processing a Metalink header, --metalink-index=<number> allows
  to process the header's application/metalink4+xml files.

* When processing a Metalink file, --trust-server-names enables the
  use of the destination file names specified in the Metalink file,
  otherwise a safe destination file name is computed.

* When processing a Metalink file, enforce a safe destination path.
  Remove any drive letter prefix under w32, i.e. 'C:D:file'.  Call
  libmetalink's metalink_check_safe_path() to prevent absolute,
  relative, or home paths:

* When processing a Metalink file, --directory-prefix=<prefix> sets
  the top of the retrieval tree to prefix for Metalink downloads.

* When processing a Metalink file, reject downloaded files which don't
  agree with their own metalink:size value:

* When processing a Metalink file, with --continue resume partially
  downloaded files and keep fully downloaded files even if they fail
  the verification.

* When processing a Metalink file, create the parent directories of a
  "path/file" destination file name:

* On a recursive download, append a .tmp suffix to temporary files
  that will be deleted after being parsed, and create them
  readable/writable only by the owner.

* New make target 'check-valgrind'

* Fix several bugs

* Fix compatibility issues
   2016-10-30 21:55:39 by S.P.Zeidler | Files touched by this commit (3)
Log message:
add a patch for CVE-2016-7098 from upstream
   2016-09-19 15:04:29 by Thomas Klausner | Files touched by this commit (147)
Log message:
Recursive PKGREVISION bump for gnutls shlib major bump.
   2016-07-09 08:39:18 by Thomas Klausner | Files touched by this commit (1068) | Package updated
Log message:
Bump PKGREVISION for perl-5.24.0 for everything mentioning perl.
   2016-06-11 20:33:22 by Thomas Klausner | Files touched by this commit (3) | Package updated
Log message:
Updated wget to 1.18.

* Changes in Wget 1.18

* By default, on server redirects to a FTP resource, use the original
  URL to get the local file name. Close CVE-2016-4971.  This
  introduces a backward-incompatibility for HTTP->FTP redirects and
  any script that relies on the old  behaviour must use

* Check the HSTS file is not world-writable before using it.

* Parse <img srcset> attributes on a recursive download.

* Fix problem with SNI server names having trailing dot(s)

* New options --bind-dns-address and --dns-servers.

* When Wget is built with libiconv, it now converts non-ASCII URIs to
  the locale's codeset when it creates files.  The encoding of the
  remote files and URIs is taken from --remote-encoding, defaulting to
  UTF-8.  The result is that non-ASCII URIs and files downloaded via
  HTTP/HTTPS and FTP will have names on the local filesystem that
  correspond to their remote names.