;;; url-util.el --- Miscellaneous helper routines for URL library -*- lexical-binding: t -*- ;; Copyright (C) 1996-1999, 2001, 2004-2021 Free Software Foundation, ;; Inc. ;; Author: Bill Perry ;; Maintainer: emacs-devel@gnu.org ;; Keywords: comm, data, processes ;; This file is part of GNU Emacs. ;; ;; GNU Emacs is free software: you can redistribute it and/or modify ;; it under the terms of the GNU General Public License as published by ;; the Free Software Foundation, either version 3 of the License, or ;; (at your option) any later version. ;; GNU Emacs is distributed in the hope that it will be useful, ;; but WITHOUT ANY WARRANTY; without even the implied warranty of ;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ;; GNU General Public License for more details. ;; You should have received a copy of the GNU General Public License ;; along with GNU Emacs. If not, see . ;;; Commentary: ;;; Code: (require 'url-parse) (require 'url-vars) (autoload 'timezone-parse-date "timezone") (autoload 'timezone-make-date-arpa-standard "timezone") (autoload 'mail-header-extract "mailheader") (defvar url-parse-args-syntax-table (copy-syntax-table emacs-lisp-mode-syntax-table) "A syntax table for parsing sgml attributes.") (modify-syntax-entry ?' "\"" url-parse-args-syntax-table) (modify-syntax-entry ?` "\"" url-parse-args-syntax-table) (modify-syntax-entry ?{ "(" url-parse-args-syntax-table) (modify-syntax-entry ?} ")" url-parse-args-syntax-table) ;;;###autoload (defcustom url-debug nil "What types of debug messages from the URL library to show. Debug messages are logged to the *URL-DEBUG* buffer. If t, all messages will be logged. If a number, all messages will be logged, as well shown via `message'. If a list, it is a list of the types of messages to be logged." :type '(choice (const :tag "none" nil) (const :tag "all" t) (checklist :tag "custom" (const :tag "HTTP" :value http) (const :tag "DAV" :value dav) (const :tag "General" :value retrieval) (const :tag "Filename handlers" :value handlers) (symbol :tag "Other"))) :group 'url-hairy) ;;;###autoload (defun url-debug (tag &rest args) (if (or (eq url-debug t) (numberp url-debug) (and (listp url-debug) (memq tag url-debug))) (with-current-buffer (get-buffer-create "*URL-DEBUG*") (goto-char (point-max)) (insert (symbol-name tag) " -> " (apply 'format args) "\n") (if (numberp url-debug) (apply 'message args))))) ;;;###autoload (defun url-parse-args (str &optional nodowncase) ;; Return an assoc list of attribute/value pairs from a string ;; that uses RFC 822 (or later) format. (let (name ; From name= value ; its value results ; Assoc list of results name-pos ; Start of XXXX= position val-pos) ; Start of value position (with-temp-buffer (insert str) (set-syntax-table url-parse-args-syntax-table) (goto-char (point-min)) (while (not (eobp)) (skip-chars-forward "; \n\t") (setq name-pos (point)) (skip-chars-forward "^ \n\t=;") (unless nodowncase (downcase-region name-pos (point))) (setq name (buffer-substring name-pos (point))) (skip-chars-forward " \t\n") (if (/= (or (char-after (point)) 0) ?=) ; There is no value (setq value nil) (skip-chars-forward " \t\n=") (setq val-pos (point) value (cond ((or (= (or (char-after val-pos) 0) ?\") (= (or (char-after val-pos) 0) ?')) (buffer-substring (1+ val-pos) (condition-case () (prog2 (forward-sexp 1) (1- (point)) (skip-chars-forward "\"")) (error (skip-chars-forward "^ \t\n") (point))))) (t (buffer-substring val-pos (progn (skip-chars-forward "^;") (skip-chars-backward " \t") (point))))))) (setq results (cons (cons name value) results)) (skip-chars-forward "; \n\t")) results))) ;;;###autoload (defun url-insert-entities-in-string (string) "Convert HTML markup-start characters to entity references in STRING. Also replaces the \" character, so that the result may be safely used as an attribute value in a tag. Returns a new string with the result of the conversion. Replaces these characters as follows: & ==> & < ==> < > ==> > \" ==> "" (if (string-match "[&<>\"]" string) (with-current-buffer (get-buffer-create " *entity*") (erase-buffer) (buffer-disable-undo (current-buffer)) (insert string) (goto-char (point-min)) (while (progn (skip-chars-forward "^&<>\"") (not (eobp))) (insert (cdr (assq (char-after (point)) '((?\" . """) (?& . "&") (?< . "<") (?> . ">"))))) (delete-char 1)) (buffer-string)) string)) ;;;###autoload (defun url-normalize-url (url) "Return a \"normalized\" version of URL. Strips out default port numbers, etc." (let (type data retval) (setq data (url-generic-parse-url url) type (url-type data)) (if (member type '("www" "about" "mailto" "info")) (setq retval url) ;; FIXME all this does, and all this function seems to do in ;; most cases, is remove any trailing "#anchor" part of a url. (setf (url-target data) nil) (setq retval (url-recreate-url data))) retval)) ;;;###autoload (defun url-lazy-message (&rest args) "Just like `message', but is a no-op if called more than once a second. Will not do anything if `url-show-status' is nil." (if (or (and url-current-object (url-silent url-current-object)) (null url-show-status) (active-minibuffer-window) (= url-lazy-message-time (setq url-lazy-message-time (time-convert nil 'integer)))) nil (apply 'message args))) ;;;###autoload (defun url-get-normalized-date (&optional specified-time) "Return a date string that most HTTP servers can understand." (let ((system-time-locale "C")) (format-time-string "%a, %d %b %Y %T GMT" specified-time t))) ;;;###autoload (defun url-eat-trailing-space (x) "Remove spaces/tabs at the end of a string." (let ((y (1- (length x))) (skip-chars (list ? ?\t ?\n))) (while (and (>= y 0) (memq (aref x y) skip-chars)) (setq y (1- y))) (substring x 0 (1+ y)))) ;;;###autoload (defun url-strip-leading-spaces (x) "Remove spaces at the front of a string." (let ((y (1- (length x))) (z 0) (skip-chars (list ? ?\t ?\n))) (while (and (<= z y) (memq (aref x z) skip-chars)) (setq z (1+ z))) (substring x z nil))) (define-obsolete-function-alias 'url-pretty-length 'file-size-human-readable "24.4") ;;;###autoload (defun url-display-percentage (fmt perc &rest args) (when (and url-show-status (or (null url-current-object) (not (url-silent url-current-object)))) (if (null fmt) (if (fboundp 'clear-progress-display) (clear-progress-display)) (if (and (fboundp 'progress-display) perc) (apply 'progress-display fmt perc args) (apply 'message fmt args))))) ;;;###autoload (defun url-percentage (x y) (if (fboundp 'float) (round (* 100 (/ x (float y)))) (/ (* x 100) y))) ;;;###autoload (defalias 'url-basepath 'url-file-directory) ;;;###autoload (defun url-file-directory (file) "Return the directory part of FILE, for a URL." (cond ((null file) "") ((string-match "\\?" file) (url-file-directory (substring file 0 (match-beginning 0)))) ((string-match "\\(.*\\(/\\|%2[fF]\\)\\)" file) (match-string 1 file)))) ;;;###autoload (defun url-file-nondirectory (file) "Return the nondirectory part of FILE, for a URL." (cond ((null file) "") ((string-match "\\?" file) (url-file-nondirectory (substring file 0 (match-beginning 0)))) ((string-match ".*\\(?:/\\|%2[fF]\\)\\(.*\\)" file) (match-string 1 file)) (t file))) ;;;###autoload (defun url-parse-query-string (query &optional downcase allow-newlines) (let (retval pairs cur key val) (setq pairs (split-string query "[;&]")) (while pairs (setq cur (car pairs) pairs (cdr pairs)) (unless (string-match "=" cur) (setq cur (concat cur "="))) (when (string-match "=" cur) (setq key (url-unhex-string (substring cur 0 (match-beginning 0)) allow-newlines)) (setq val (url-unhex-string (substring cur (match-end 0) nil) allow-newlines)) (if downcase (setq key (downcase key))) (setq cur (assoc key retval)) (if cur (setcdr cur (cons val (cdr cur))) (setq retval (cons (list key val) retval))))) retval)) ;;;###autoload (defun url-build-query-string (query &optional semicolons keep-empty) "Build a query-string. Given a QUERY in the form: ((key1 val1) (key2 val2) (key3 val1 val2) (key4) (key5 \"\")) \(This is the same format as produced by `url-parse-query-string') This will return a string \"key1=val1&key2=val2&key3=val1&key3=val2&key4&key5\". Keys may be strings or symbols; if they are symbols, the symbol name will be used. When SEMICOLONS is given, the separator will be \";\". When KEEP-EMPTY is given, empty values will show as \"key=\" instead of just \"key\" as in the example above." (mapconcat (lambda (key-vals) (let ((escaped (mapcar (lambda (sym) (url-hexify-string (format "%s" sym))) key-vals))) (mapconcat (lambda (val) (let ((vprint (format "%s" val)) (eprint (format "%s" (car escaped)))) (concat eprint (if (or keep-empty (and val (not (zerop (length vprint))))) "=" "") vprint))) (or (cdr escaped) '("")) (if semicolons ";" "&")))) query (if semicolons ";" "&"))) (defun url-unhex (x) (if (> x ?9) (if (>= x ?a) (+ 10 (- x ?a)) (+ 10 (- x ?A))) (- x ?0))) ;; Fixme: Is this definition better, and does it ever matter? ;; (defun url-unhex-string (str &optional allow-newlines) ;; "Remove %XX, embedded spaces, etc in a url. ;; If optional second argument ALLOW-NEWLINES is non-nil, then allow the ;; decoding of carriage returns and line feeds in the string, which is normally ;; forbidden in URL encoding." ;; (setq str (or str "")) ;; (setq str (replace-regexp-in-string "%[[:xdigit:]]\\{2\\}" ;; (lambda (match) ;; (string (string-to-number ;; (substring match 1) 16))) ;; str t t)) ;; (if allow-newlines ;; (replace-regexp-in-string "[\n\r]" (lambda (match) ;; (format "%%%.2X" (aref match 0))) ;; str t t) ;; str)) ;;;###autoload (defun url-unhex-string (str &optional allow-newlines) "Remove %XX embedded spaces, etc in a URL. If optional second argument ALLOW-NEWLINES is non-nil, then allow the decoding of carriage returns and line feeds in the string, which is normally forbidden in URL encoding." (setq str (or str "")) (let ((tmp "") (case-fold-search t)) (while (string-match "%[0-9a-f][0-9a-f]" str) (let* ((start (match-beginning 0)) (ch1 (url-unhex (elt str (+ start 1)))) (code (+ (* 16 ch1) (url-unhex (elt str (+ start 2)))))) (setq tmp (concat tmp (substring str 0 start) (cond (allow-newlines (byte-to-string code)) ((or (= code ?\n) (= code ?\r)) " ") (t (byte-to-string code)))) str (substring str (match-end 0))))) (concat tmp str))) (defconst url-unreserved-chars '(?a ?b ?c ?d ?e ?f ?g ?h ?i ?j ?k ?l ?m ?n ?o ?p ?q ?r ?s ?t ?u ?v ?w ?x ?y ?z ?A ?B ?C ?D ?E ?F ?G ?H ?I ?J ?K ?L ?M ?N ?O ?P ?Q ?R ?S ?T ?U ?V ?W ?X ?Y ?Z ?0 ?1 ?2 ?3 ?4 ?5 ?6 ?7 ?8 ?9 ?- ?_ ?. ?~) "List of characters that are unreserved in the URL spec. This is taken from RFC 3986 (section 2.3).") (defconst url-encoding-table (let ((vec (make-vector 256 nil))) (dotimes (byte 256) ;; RFC 3986 (Section 2.1): For consistency, URI producers and ;; normalizers should use uppercase hexadecimal digits for all ;; percent-encodings. (aset vec byte (format "%%%02X" byte))) vec) "Vector translating bytes to URI-encoded %-sequences.") (defun url--allowed-chars (char-list) "Return an \"allowed character\" mask (a 256-slot vector). The Nth element is non-nil if character N is in CHAR-LIST. The result can be passed as the second arg to `url-hexify-string'." (let ((vec (make-vector 256 nil))) (dolist (byte char-list) (ignore-errors (aset vec byte t))) vec)) ;;;###autoload (defun url-hexify-string (string &optional allowed-chars) "URI-encode STRING and return the result. If STRING is multibyte, it is first converted to a utf-8 byte string. Each byte corresponding to an allowed character is left as-is, while all other bytes are converted to a three-character string: \"%\" followed by two upper-case hex digits. The allowed characters are specified by ALLOWED-CHARS. If this argument is nil, the list `url-unreserved-chars' determines the allowed characters. Otherwise, ALLOWED-CHARS should be either a list of allowed chars, or a vector whose Nth element is non-nil if character N is allowed." (if allowed-chars (unless (vectorp allowed-chars) (setq allowed-chars (url--allowed-chars allowed-chars))) (setq allowed-chars (url--allowed-chars url-unreserved-chars))) (mapconcat (lambda (byte) (if (aref allowed-chars byte) (char-to-string byte) (aref url-encoding-table byte))) (if (multibyte-string-p string) (encode-coding-string string 'utf-8) string) "")) (defconst url-host-allowed-chars ;; Allow % to avoid re-encoding %-encoded sequences. (url--allowed-chars (append '(?% ?! ?$ ?& ?' ?\( ?\) ?* ?+ ?, ?\; ?=) url-unreserved-chars)) "Allowed-character byte mask for the host segment of a URI. These characters are specified in RFC 3986, Appendix A.") (defconst url-path-allowed-chars (let ((vec (copy-sequence url-host-allowed-chars))) (aset vec ?/ t) (aset vec ?: t) (aset vec ?@ t) vec) "Allowed-character byte mask for the path segment of a URI. These characters are specified in RFC 3986, Appendix A.") (defconst url-query-allowed-chars (let ((vec (copy-sequence url-path-allowed-chars))) (aset vec ?? t) vec) "Allowed-character byte mask for the query segment of a URI. These characters are specified in RFC 3986, Appendix A.") ;;;###autoload (defun url-encode-url (url) "Return a properly URI-encoded version of URL. This function also performs URI normalization, e.g. converting the scheme to lowercase if it is uppercase. Apart from normalization, if URL is already URI-encoded, this function should return it unchanged." (let* ((obj (url-generic-parse-url url)) (user (url-user obj)) (pass (url-password obj)) (path-and-query (url-path-and-query obj)) (path (car path-and-query)) (query (cdr path-and-query)) (frag (url-target obj))) (if user (setf (url-user obj) (url-hexify-string user))) (if pass (setf (url-password obj) (url-hexify-string pass))) (if path (setq path (url-hexify-string path url-path-allowed-chars))) (if query (setq query (url-hexify-string query url-query-allowed-chars))) (setf (url-filename obj) (if query (concat path "?" query) path)) (if frag (setf (url-target obj) (url-hexify-string frag url-query-allowed-chars))) (url-recreate-url obj))) ;;;###autoload (defun url-file-extension (fname &optional x) "Return the filename extension of FNAME. If optional argument X is t, then return the basename of the file with the extension stripped off." (if (and fname (setq fname (url-file-nondirectory fname)) (string-match "\\.[^./]+$" fname)) (if x (substring fname 0 (match-beginning 0)) (substring fname (match-beginning 0) nil)) ;; ;; If fname has no extension, and x then return fname itself instead of ;; nothing. When caching it allows the correct .hdr file to be produced ;; for filenames without extension. ;; (if x fname ""))) ;;;###autoload (defun url-truncate-url-for-viewing (url &optional width) "Return a shortened version of URL that is WIDTH characters wide or less. WIDTH defaults to the current frame width." (let* ((fr-width (or width (frame-width))) (str-width (length url)) (fname nil) (modified 0) (urlobj nil)) ;; The first thing that can go are the search strings (if (and (>= str-width fr-width) (string-match "\\?" url)) (setq url (concat (substring url 0 (match-beginning 0)) "?...") str-width (length url))) (if (< str-width fr-width) nil ; Hey, we are done! (setq urlobj (url-generic-parse-url url) fname (url-filename urlobj) fr-width (- fr-width 4)) (while (and (>= str-width fr-width) (string-match "/" fname)) (setq fname (substring fname (match-end 0) nil) modified (1+ modified)) (setf (url-filename urlobj) fname) (setq url (url-recreate-url urlobj) str-width (length url))) (if (> modified 1) (setq fname (concat "/.../" fname)) (setq fname (concat "/" fname))) (setf (url-filename urlobj) fname) (setq url (url-recreate-url urlobj))) url)) ;;;###autoload (defun url-view-url (&optional no-show) "View the current document's URL. Optional argument NO-SHOW means just return the URL, don't show it in the minibuffer. This uses `url-current-object', set locally to the buffer." (interactive) (if (not url-current-object) nil (if no-show (url-recreate-url url-current-object) (message "%s" (url-recreate-url url-current-object))))) (defvar url-get-url-filename-chars "-%.?@a-zA-Z0-9()_/:~=&" "Valid characters in a URL.") (defun url-get-url-at-point (&optional pt) "Get the URL closest to point, but don't change position. Has a preference for looking backward when not directly on a symbol." (declare (obsolete thing-at-point-url-at-point "27.1")) ;; Not at all perfect - point must be right in the name. (save-excursion (if pt (goto-char pt)) (let (start url) (save-excursion ;; first see if you're just past a filename (if (not (eobp)) (if (looking-at "[] \t\n[{}()]") ; whitespace or some parens (progn (skip-chars-backward " \n\t\r({[]})") (if (not (bobp)) (backward-char 1))))) (if (and (char-after (point)) (string-match (concat "[" url-get-url-filename-chars "]") (char-to-string (char-after (point))))) (progn (skip-chars-backward url-get-url-filename-chars) (setq start (point)) (skip-chars-forward url-get-url-filename-chars)) (setq start (point))) (setq url (buffer-substring-no-properties start (point)))) (if (and url (string-match "^(\\(.*\\))\\.?$" url)) (setq url (match-string 1 url))) (if (and url (string-match "^URL:" url)) (setq url (substring url 4 nil))) (if (and url (string-match "\\.$" url)) (setq url (substring url 0 -1))) (if (and url (string-match "^www\\." url)) (setq url (concat "http://" url))) (if (and url (not (string-match url-nonrelative-link url))) (setq url nil)) url))) (defun url-extract-mime-headers () "Set `url-current-mime-headers' in current buffer." (save-excursion (goto-char (point-min)) (unless url-current-mime-headers (setq-local url-current-mime-headers (mail-header-extract))))) (defun url-make-private-file (file) "Make FILE only readable and writable by the current user. Creates FILE and its parent directories if they do not exist." (let ((dir (file-name-directory file))) (when dir ;; For historical reasons. (make-directory dir t))) ;; Based on doc-view-make-safe-dir. (condition-case nil (with-file-modes #o0600 (with-temp-buffer (write-region (point-min) (point-max) file nil 'silent nil 'excl))) (file-already-exists (set-file-modes file #o0600 'nofollow)))) (autoload 'puny-encode-domain "puny") (autoload 'url-domsuf-cookie-allowed-p "url-domsuf") ;;;###autoload (defun url-domain (url) "Return the domain of the host of the URL. Return nil if this can't be determined. For instance, this function will return \"fsf.co.uk\" if the host in URL is \"www.fsf.co.uk\"." (let* ((host (puny-encode-domain (url-host url))) (parts (nreverse (split-string host "\\."))) (candidate (pop parts)) found) ;; IP addresses aren't domains. (when (string-match "\\`[0-9.]+\\'" host) (setq parts nil)) ;; We assume that the top-level domain is never an appropriate ;; thing as "the domain", so we start at the next one (eg. ;; "fsf.org"). (while (and parts (not (setq found (url-domsuf-cookie-allowed-p (setq candidate (concat (pop parts) "." candidate)))))) ) (and found candidate))) (provide 'url-util) ;;; url-util.el ends here