noble (7) urlwatch-cookbook.7.gz

Provided by: urlwatch_2.28-3_all bug

NAME

       urlwatch-cookbook - Advanced topics and recipes for urlwatch

ADDING URLS FROM THE COMMAND LINE

       Quickly adding new URLs to the job list from the command line:

          urlwatch --add url=http://example.org,name=Example

USING WORD-BASED DIFFERENCES

       You can also specify an external diff-style tool (a tool that takes two filenames (old, new) as parameter
       and returns on its standard output the difference of the files), for  example  to  use  wdiff(1)  to  get
       word-based  differences instead of line-based difference, or pandiff <https://github.com/davidar/pandiff>
       to get markdown differences:

          url: https://example.com/
          diff_tool: wdiff

       Note that diff_tool specifies an external command-line tool, so that tool must  be  installed  separately
       (e.g. apt  install  wdiff on Debian or brew install wdiff on macOS). Syntax highlighting is supported for
       wdiff-style output, but potentially not for other diff tools.

IGNORING WHITESPACE CHANGES

       If you would like to ignore whitespace changes so  that  you  don't  receive  notifications  for  trivial
       differences, you can use diff_tool for this.  For example:

          diff_tool: "diff --ignore-all-space --unified"

       When  using  another external diff-like tool, make sure it returns unified output format to retain syntax
       highlighting.

ONLY SHOW ADDED OR REMOVED LINES

       The diff_filter feature can be used to filter the diff output text with the same tools (see Filters) used
       for filtering web pages.

       In order to show only diff lines with added lines, use:

          url: http://example.com/things-get-added.html
          diff_filter:
            - grep: '^[@+]'

       This will only keep diff lines starting with @ or +. Similarly, to only keep removed lines:

          url: http://example.com/things-get-removed.html
          diff_filter:
            - grep: '^[@-]'

       More  sophisticated  diff  filtering  is  possibly by combining existing filters, writing a new filter or
       using shellpipe to delegate the filtering/processing of the diff output to an external tool.

       Read the next section if you want to disable empty notifications.

DISABLE EMPTY NOTIFICATIONS

       As an extension to the previous example, let's say you want to only get notified with  all  lines  added,
       but receive no notifications at all if lines are removed.

       A diff usually looks like this:

          --- @       Fri, 04 Mar 2022 19:58:14 +0100
          +++ @       Fri, 04 Mar 2022 19:58:22 +0100
          @@ -1,3 +1,3 @@
           someline
          -someotherlines
          +someotherline
           anotherline

       We  want  to  filter  all lines starting with "+" only, but because of the headers we also want to filter
       lines that start with "+++", which can be accomplished like so:

          url: http://example.com/only-added.html
          diff_filter:
            - grep: '^[+]'      # Include all lines starting with "+"
            - grepi: '^[+]{3}'  # Exclude the line starting with "+++"

       This deals with all diff lines now, but since urlwatch reports "changed" pages even when the  diff_filter
       returns  an  empty  string (which might be useful in some cases), you have to explicitly opt out by using
       urlwatch --edit-config and setting the empty-diff option to false in the display category:

          display:
            empty-diff: false

PASS DIFF OUTPUT TO A CUSTOM SCRIPT

       In some situations, it might be useful to run a script with the diff as input when changes were  detected
       (e.g.  to  start  an  update  or  process  something). This can be done by combining diff_filter with the
       shellpipe filter, which can be any custom script.

       The output of the custom script will then be the diff result as reported by urlwatch, so  if  it  outputs
       any status, the CHANGED notification that urlwatch does will contain the output of the custom script, not
       the original diff. This can even have a "normal" filter attached to only watch links (the css: a part  of
       the filter definitions):

          url: http://example.org/downloadlist.html
          filter:
            - css: a
          diff_filter:
            - shellpipe: /usr/local/bin/process_new_links.sh

COMPARING WEB PAGES VISUALLY

       To    compare    the    visual   contents   of   web   pages,   Nicolai   has   written   pyvisualcompare
       <https://github.com/nspo/pyvisualcompare> as a frontend (with GUI) to urlwatch. The tool can be  used  to
       select  a region of a web page. It then generates a configuration for urlwatch to run pyvisualcompare and
       generate a hash for the screen contents.

IGNORING CONNECTION ERRORS

       In some cases, it might be useful to ignore (temporary) network errors to avoid notifications being sent.
       While  there  is  a  display.error  config  option  (defaulting  to  true) to control reporting of errors
       globally, to ignore network errors for specific jobs only, you can use the  ignore_connection_errors  key
       in the job list configuration file:

          url: https://example.com/
          ignore_connection_errors: true

       Similarly, you might want to ignore some (temporary) HTTP errors on the server side:

          url: https://example.com/
          ignore_http_error_codes: 408, 429, 500, 502, 503, 504

       or ignore all HTTP errors if you like:

          url: https://example.com/
          ignore_http_error_codes: 4xx, 5xx

OVERRIDING THE CONTENT ENCODING

       For  web  pages with misconfigured HTTP headers or rare encodings, it may be useful to explicitly specify
       an encoding from  Python’s  Standard  Encodings  <https://docs.python.org/3/library/codecs.html#standard-
       encodings>.

          url: https://example.com/
          encoding: utf-8

CHANGING THE DEFAULT TIMEOUT

       By  default,  url  jobs timeout after 60 seconds. If you want a different timeout period, use the timeout
       key to specify it in number of seconds, or set it to 0 to never timeout.

          url: https://example.com/
          timeout: 300

       It is possible to add cookies to HTTP requests for pages that need it, the YAML syntax for this is:

          url: http://example.com/
          cookies:
              Key: ValueForKey
              OtherKey: OtherValue

COMPARING WITH SEVERAL LATEST SNAPSHOTS

       If a webpage frequently changes between several known stable states, it may be desirable to have  changes
       reported only if the webpage changes into a new unknown state. You can use compared_versions to do this.

          url: https://example.com/
          compared_versions: 3

       In  this  example,  changes  are  only  reported  if  the webpage becomes different from the latest three
       distinct states. The differences are shown relative to the closest match.

RECEIVING A REPORT EVERY TIME URLWATCH RUNS

       If you are watching pages that change seldomly, but you still want to be notified daily if urlwatch still
       works, you can watch the output of the date command, for example:

          name: "urlwatch watchdog"
          command: "date"

       Since  the  output  of date changes every second, this job should produce a report every time urlwatch is
       run.

USING REDIS AS A CACHE BACKEND

       If you want to use Redis as a cache backend over the default SQLite3 file:

          urlwatch --cache=redis://localhost:6379/

       There is no migration path from the SQLite3 format, the cache will be empty the first time Redis is used.

WATCHING CHANGES ON .ONION (TOR) PAGES

       Since pages on the Tor Network <https://www.torproject.org> are not accessible via public  DNS  and  TCP,
       you  need  to  either  configure  a Tor client as HTTP/HTTPS proxy or use the torify(1) tool from the tor
       package (apt install tor on Debian, brew install tor on macOS). Setting up Tor is out of scope  for  this
       document. On a properly set up Tor installation, one can just prefix the urlwatch command with the torify
       wrapper to access .onion pages:

          torify urlwatch

WATCHING FACEBOOK PAGE EVENTS

       If you want to be notified of new events on a public  Facebook  page,  you  can  use  the  following  job
       pattern,  replace  PAGE  with the name of the page (can be found by navigating to the events page on your
       browser):

          url: http://m.facebook.com/PAGE/pages/permalink/?view_type=tab_events
          filter:
            - css:
                selector: div#objects_container
                exclude: 'div.x, #m_more_friends_who_like_this, img'
            - re.sub:
                pattern: '(/events/\d*)[^"]*'
                repl: '\1'
            - html2text: pyhtml2text

SETTING THE CONTENT WIDTH FOR HTML2TEXT (LYNX METHOD)

       When using the lynx method in the html2text filter, it uses a default width that  will  cause  additional
       line breaks to be inserted.

       To set the lynx output width to 400 characters, use this filter setup:

          url: http://example.com/longlines.html
          filter:
            - html2text:
                method: lynx
                width: 400

CONFIGURING HOW LONG BROWSER JOBS WAIT FOR PAGES TO LOAD

       For  browser  jobs, you can configure how long the headless browser will wait before a page is considered
       loaded by using the wait_until option.

       It can take one  of  four  values  (see  wait_until  docs  <https://playwright.dev/python/docs/api/class-
       page#page-goto-option-wait-until> of Playwright):

          • load - consider operation to be finished when the load event is fired

          • domcontentloaded - consider operation to be finished when the DOMContentLoaded event is fired

          • networkidle  -  discouraged  consider operation to be finished when there are no network connections
            for at least 500 ms. Don't use this method for testing, rely on web assertions to  assess  readiness
            instead

          • commit  -  consider  operation  to  be  finished  when network response is received and the document
            started loading

TREATING NEW JOBS AS CHANGED

       In some cases (e.g. when the diff_tool or diff_filter executes some external command  as  a  side  effect
       that  should  also  run  for the initial page state), you can set the treat_new_as_changed to true, which
       will make the job report as CHANGED instead of NEW the first time it is retrieved (and the diff  will  be
       reported, too).

          url: http://example.com/initialpage.html
          treat_new_as_changed: true

       This  option  will  also  change the behavior of --test-diff-filter, and allow testing the diff filter if
       only a single version of the page has been retrieved.

MONITORING THE SAME URL IN MULTIPLE JOBS

       Because urlwatch uses the url/navigate (for URL/Browser jobs) and/or the command (for Shell jobs) key  as
       unique identifier, each URL can only appear in a single job. If you want to monitor the same URL multiple
       times, you can append #1, #2, ... (or anything that makes them unique) to the URLs, like this:

          name: "Looking for Thing A"
          url: http://example.com/#1
          filter:
            - grep: "Thing A"
          ---
          name: "Looking for Thing B"
          url: http://example.com/#2
          filter:
            - grep: "Thing B"

UPDATING A URL AND KEEPING PAST HISTORY

       Job history is stored based on the  value  of  the  url  parameter,  so  updating  a  job's  URL  in  the
       configuration  file  urls.yaml  will  create  a  new  job  with  no  history.   Retain  history  by using
       --change-location:

          urlwatch --change-location http://example.org#old http://example.org#new

       The command also works with Browser and Shell jobs, changing navigate and command respectively.

RUNNING A SUBSET OF JOBS

       To run one or more specific jobs instead of all known jobs, provide the job index numbers to the urlwatch
       command. For example, to run jobs with index 2, 4, and 7:

          urlwatch 2 4 7

SENDING HTML FORM DATA USING POST

       To  simulate submitting a HTML form using the POST method, you can pass the form fields in the data field
       of the job description:

          name: "My POST Job"
          url: http://example.com/foo
          data:
            username: "foo"
            password: "bar"
            submit: "Send query"

       By default,  the  request  will  use  the  HTTP  POST  method,  and  the  Content-type  will  be  set  to
       application/x-www-form-urlencoded.

SENDING ARBITRARY DATA USING HTTP PUT

       It  is  possible  to  customize  the  HTTP method and Content-type header, allowing you to send arbitrary
       requests to the server:

          name: "My PUT Request"
          url: http://example.com/item/new
          method: PUT
          headers:
            Content-type: application/json
          data: '{"foo": true}'

SEE ALSO

       urlwatch(1),    urlwatch-intro(7),     urlwatch-jobs(5),     urlwatch-filters(5),     urlwatch-config(5),
       urlwatch-reporters(5)

UTF-8 SUPPORT ON WINDOWS

       On  Windows, the default file encoding might be locale-specific and not work correctly if files are saved
       using the (recommended) UTF-8 encoding.

       If you are having problems loading UTF-8-encoded files on Windows,  you  might  see  an  issue  like  the
       following when urlwatch parses your config files:

          UnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 214: character maps to <undefined>

       To work around this issue, Python 3.7 and newer have a new UTF-8 Mode <https://peps.python.org/pep-0540/>
       that can be enabled by setting the environment variable PYTHONUTF8 to 1:

          set PYTHONUTF8=1
          urlwatch

       You can also add this environment variable to your user environment or system environment  to  apply  the
       UTF-8 Mode to all Python programs on your machine.

       2023 Thomas Perl

                                                  May 03, 2023                              URLWATCH-COOKBOOK(7)