[tlu] Strange error about wget

Bruno Voisin bvoisin at icloud.com
Fri May 14 11:30:01 CEST 2021

Looking closer, wget's --no-check-certificate is called in TLConfig.pm line 134. The version on my setup is dated May 7, part of texlive.infra revision 59154 (says texlive.tlpdb), updated this morning.


> Le 14 mai 2021 à 10:58, Bruno Voisin via tlu <tlu at tug.org> a écrit :
> Today, with TLU 1.50b1, I got this strange error when applying updates:
> 2021-05-14 08:31:18 +0000 Notice tlu_ipctask[31541]	Mise à jour en cours xint
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	/usr/local/texlive/2021/tlpkg/installer/wget/wget.universal-darwin: unrecognized option '--no-check-certificate'
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	Usage: wget.universal-darwin [OPTION]... [URL]...
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	Try `wget.universal-darwin --help' for more options.
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	TLPDB::_install_data: downloading did not succeed (download_file failed) for https://ctan.tetaneutral.net/systems/texlive/tlnet/archive/xint.doc.tar.xz
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	tlmgr: Installation of new version of xint failed, trying to unwind.
> 2021-05-14 08:32:37 +0000 Warning tlu_ipctask[31541]	tlmgr: Restoring old package state succeeded.
> To be sure, I quit TLU then tried again, applying the xint update alone, and got the same with a different mirror:
> 2021-05-14 08:51:21 +0000 Notice tlu_ipctask[32449]	Mise à jour en cours xint
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	/usr/local/texlive/2021/tlpkg/installer/wget/wget.universal-darwin: unrecognized option '--no-check-certificate'
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	Usage: wget.universal-darwin [OPTION]... [URL]...
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	Try `wget.universal-darwin --help' for more options.
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	TLPDB::_install_data: downloading did not succeed (download_file failed) for https://ctan.mines-albi.fr/systems/texlive/tlnet/archive/xint.tar.xz
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	tlmgr: Installation of new version of xint failed, trying to unwind.
> 2021-05-14 08:51:29 +0000 Warning tlu_ipctask[32449]	tlmgr: Restoring old package state succeeded.
> 2021-05-14 08:51:30 +0000 Warning tlu_ipctask[32449]	tlmgr: action update returned an error; continuing.
> It may be that the error has nothing to see with wget's --no-check-certificate, that it is caused by very poor internet connection (download speed about 2 Mb/s on Catalina, 0.1 Mb/s on Big Sur) at the place where I am for a few days.
> Below the details on wget (installed with MacTeX 2021).
> Bruno
> % /usr/local/texlive/2021/tlpkg/installer/wget/wget.universal-darwin --help
> GNU Wget 1.21, a non-interactive network retriever.
> Usage: wget.universal-darwin [OPTION]... [URL]...
> Mandatory arguments to long options are mandatory for short options too.
> Startup:
>  -V,  --version                   display the version of Wget and exit
>  -h,  --help                      print this help
>  -b,  --background                go to background after startup
>  -e,  --execute=COMMAND           execute a `.wgetrc'-style command
> Logging and input file:
>  -o,  --output-file=FILE          log messages to FILE
>  -a,  --append-output=FILE        append messages to FILE
>  -d,  --debug                     print lots of debugging information
>  -q,  --quiet                     quiet (no output)
>  -v,  --verbose                   be verbose (this is the default)
>  -nv, --no-verbose                turn off verboseness, without being quiet
>       --report-speed=TYPE         output bandwidth as TYPE.  TYPE can be bits
>  -i,  --input-file=FILE           download URLs found in local or external FILE
>  -F,  --force-html                treat input file as HTML
>  -B,  --base=URL                  resolves HTML input-file links (-i -F)
>                                     relative to URL
>       --config=FILE               specify config file to use
>       --no-config                 do not read any config file
>       --rejected-log=FILE         log reasons for URL rejection to FILE
> Download:
>  -t,  --tries=NUMBER              set number of retries to NUMBER (0 unlimits)
>       --retry-connrefused         retry even if connection is refused
>       --retry-on-http-error=ERRORS    comma-separated list of HTTP errors to retry
>  -O,  --output-document=FILE      write documents to FILE
>  -nc, --no-clobber                skip downloads that would download to
>                                     existing files (overwriting them)
>       --no-netrc                  don't try to obtain credentials from .netrc
>  -c,  --continue                  resume getting a partially-downloaded file
>       --start-pos=OFFSET          start downloading from zero-based position OFFSET
>       --progress=TYPE             select progress gauge type
>       --show-progress             display the progress bar in any verbosity mode
>  -N,  --timestamping              don't re-retrieve files unless newer than
>                                     local
>       --no-if-modified-since      don't use conditional if-modified-since get
>                                     requests in timestamping mode
>       --no-use-server-timestamps  don't set the local file's timestamp by
>                                     the one on the server
>  -S,  --server-response           print server response
>       --spider                    don't download anything
>  -T,  --timeout=SECONDS           set all timeout values to SECONDS
>       --dns-timeout=SECS          set the DNS lookup timeout to SECS
>       --connect-timeout=SECS      set the connect timeout to SECS
>       --read-timeout=SECS         set the read timeout to SECS
>  -w,  --wait=SECONDS              wait SECONDS between retrievals
>                                     (applies if more then 1 URL is to be retrieved)
>       --waitretry=SECONDS         wait 1..SECONDS between retries of a retrieval
>                                     (applies if more then 1 URL is to be retrieved)
>       --random-wait               wait from 0.5*WAIT...1.5*WAIT secs between retrievals
>                                     (applies if more then 1 URL is to be retrieved)
>       --no-proxy                  explicitly turn off proxy
>  -Q,  --quota=NUMBER              set retrieval quota to NUMBER
>       --bind-address=ADDRESS      bind to ADDRESS (hostname or IP) on local host
>       --limit-rate=RATE           limit download rate to RATE
>       --no-dns-cache              disable caching DNS lookups
>       --restrict-file-names=OS    restrict chars in file names to ones OS allows
>       --ignore-case               ignore case when matching files/directories
>  -4,  --inet4-only                connect only to IPv4 addresses
>  -6,  --inet6-only                connect only to IPv6 addresses
>       --prefer-family=FAMILY      connect first to addresses of specified family,
>                                     one of IPv6, IPv4, or none
>       --user=USER                 set both ftp and http user to USER
>       --password=PASS             set both ftp and http password to PASS
>       --ask-password              prompt for passwords
>       --use-askpass=COMMAND       specify credential handler for requesting 
>                                     username and password.  If no COMMAND is 
>                                     specified the WGET_ASKPASS or the SSH_ASKPASS 
>                                     environment variable is used.
>       --no-iri                    turn off IRI support
>       --local-encoding=ENC        use ENC as the local encoding for IRIs
>       --remote-encoding=ENC       use ENC as the default remote encoding
>       --unlink                    remove file before clobber
>       --xattr                     turn on storage of metadata in extended file attributes
> Directories:
>  -nd, --no-directories            don't create directories
>  -x,  --force-directories         force creation of directories
>  -nH, --no-host-directories       don't create host directories
>       --protocol-directories      use protocol name in directories
>  -P,  --directory-prefix=PREFIX   save files to PREFIX/..
>       --cut-dirs=NUMBER           ignore NUMBER remote directory components
> HTTP options:
>       --http-user=USER            set http user to USER
>       --http-password=PASS        set http password to PASS
>       --no-cache                  disallow server-cached data
>       --default-page=NAME         change the default page name (normally
>                                     this is 'index.html'.)
>  -E,  --adjust-extension          save HTML/CSS documents with proper extensions
>       --ignore-length             ignore 'Content-Length' header field
>       --header=STRING             insert STRING among the headers
>       --max-redirect              maximum redirections allowed per page
>       --proxy-user=USER           set USER as proxy username
>       --proxy-password=PASS       set PASS as proxy password
>       --referer=URL               include 'Referer: URL' header in HTTP request
>       --save-headers              save the HTTP headers to file
>  -U,  --user-agent=AGENT          identify as AGENT instead of Wget/VERSION
>       --no-http-keep-alive        disable HTTP keep-alive (persistent connections)
>       --no-cookies                don't use cookies
>       --load-cookies=FILE         load cookies from FILE before session
>       --save-cookies=FILE         save cookies to FILE after session
>       --keep-session-cookies      load and save session (non-permanent) cookies
>       --post-data=STRING          use the POST method; send STRING as the data
>       --post-file=FILE            use the POST method; send contents of FILE
>       --method=HTTPMethod         use method "HTTPMethod" in the request
>       --body-data=STRING          send STRING as data. --method MUST be set
>       --body-file=FILE            send contents of FILE. --method MUST be set
>       --content-disposition       honor the Content-Disposition header when
>                                     choosing local file names (EXPERIMENTAL)
>       --content-on-error          output the received content on server errors
>       --auth-no-challenge         send Basic HTTP authentication information
>                                     without first waiting for the server's
>                                     challenge
> FTP options:
>       --ftp-user=USER             set ftp user to USER
>       --ftp-password=PASS         set ftp password to PASS
>       --no-remove-listing         don't remove '.listing' files
>       --no-glob                   turn off FTP file name globbing
>       --no-passive-ftp            disable the "passive" transfer mode
>       --preserve-permissions      preserve remote file permissions
>       --retr-symlinks             when recursing, get linked-to files (not dir)
> WARC options:
>       --warc-file=FILENAME        save request/response data to a .warc.gz file
>       --warc-header=STRING        insert STRING into the warcinfo record
>       --warc-max-size=NUMBER      set maximum size of WARC files to NUMBER
>       --warc-cdx                  write CDX index files
>       --warc-dedup=FILENAME       do not store records listed in this CDX file
>       --no-warc-digests           do not calculate SHA1 digests
>       --no-warc-keep-log          do not store the log file in a WARC record
>       --warc-tempdir=DIRECTORY    location for temporary files created by the
>                                     WARC writer
> Recursive download:
>  -r,  --recursive                 specify recursive download
>  -l,  --level=NUMBER              maximum recursion depth (inf or 0 for infinite)
>       --delete-after              delete files locally after downloading them
>  -k,  --convert-links             make links in downloaded HTML or CSS point to
>                                     local files
>       --convert-file-only         convert the file part of the URLs only (usually known as the basename)
>       --backups=N                 before writing file X, rotate up to N backup files
>  -K,  --backup-converted          before converting file X, back up as X.orig
>  -m,  --mirror                    shortcut for -N -r -l inf --no-remove-listing
>  -p,  --page-requisites           get all images, etc. needed to display HTML page
>       --strict-comments           turn on strict (SGML) handling of HTML comments
> Recursive accept/reject:
>  -A,  --accept=LIST               comma-separated list of accepted extensions
>  -R,  --reject=LIST               comma-separated list of rejected extensions
>       --accept-regex=REGEX        regex matching accepted URLs
>       --reject-regex=REGEX        regex matching rejected URLs
>       --regex-type=TYPE           regex type (posix)
>  -D,  --domains=LIST              comma-separated list of accepted domains
>       --exclude-domains=LIST      comma-separated list of rejected domains
>       --follow-ftp                follow FTP links from HTML documents
>       --follow-tags=LIST          comma-separated list of followed HTML tags
>       --ignore-tags=LIST          comma-separated list of ignored HTML tags
>  -H,  --span-hosts                go to foreign hosts when recursive
>  -L,  --relative                  follow relative links only
>  -I,  --include-directories=LIST  list of allowed directories
>       --trust-server-names        use the name specified by the redirection
>                                     URL's last component
>  -X,  --exclude-directories=LIST  list of excluded directories
>  -np, --no-parent                 don't ascend to the parent directory
> Email bug reports, questions, discussions to <bug-wget at gnu.org>
> and/or open issues at https://savannah.gnu.org/bugs/?func=additem&group=wget.

More information about the tlu mailing list.