Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix a few typos and double-spacing in the man page #560

Merged
merged 4 commits into from
Mar 6, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/cautions.rst
Original file line number Diff line number Diff line change
Expand Up @@ -298,7 +298,7 @@ If you use ``rmlint``'s ``sha1`` hash features, which features 160 bit output,
you need at least :math:`5.4\times10^{22}` files before you get a :math:`0.1\%`
probability of collision. ``rmlint``'s ``-p`` option uses ``SHA512``
(:math:`5.2\times10^{75}` files for :math:`0.1\%` risk), while ``rmlint``'s
``-pp`` option uses direct file comparison to eliminate the risk altogether.
``-p`` option uses direct file comparison to eliminate the risk altogether.
Refer to the :ref:`benchmark_ref` chapter for speed and memory overhead
implications.

Expand Down Expand Up @@ -347,7 +347,7 @@ The speed ratio gives an indication of how effectively the search algorithm mana
+----------------+----------------+---------------------+---------+
| ``rmlint`` | 38s | 106s | 2.8 |
+----------------+----------------+---------------------+---------+
| ``rmlint -pp`` | 40s | 139s | 3.5 |
| ``rmlint -p`` | 40s | 139s | 3.5 |
+----------------+----------------+---------------------+---------+

.. note::
Expand Down
2 changes: 1 addition & 1 deletion docs/faq.rst
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ by default. Suspicious options you should look into are:
* ``--merge-directories``: pulls in both ``--hidden`` and ``--hardlinked``.

If there's still a difference, check with another algorithm. In particular use
``-pp`` to enable paranoid mode. Also make sure to have ``-D``
``-p`` to enable paranoid mode. Also make sure to have ``-D``
(``--merge-directories``) disabled to see the raw number of duplicate files.

Still here? Maybe talk to us on the `issue tracker`_.
Expand Down
36 changes: 18 additions & 18 deletions docs/rmlint.1.rst
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ to build from scratch with standard tools.
In order to find the lint, ``rmlint`` is given one or more directories to traverse.
If no directories or files were given, the current working directory is assumed.
By default, ``rmlint`` will ignore hidden files and will not follow symlinks (see
`Traversal Options`_). ``rmlint`` will first find "other lint" and then search
`Traversal Options`_). ``rmlint`` will first find "other lint" and then search
the remaining files for duplicates.

``rmlint`` tries to be helpful by guessing what file of a group of duplicates
Expand All @@ -51,7 +51,7 @@ come from the same path, it will also apply different fallback sort strategies

This behaviour can be also overwritten if you know that a certain directory
contains duplicates and another one originals. In this case you write the
original directory after specifying a single ``//`` on the commandline.
original directory after specifying a single ``//`` on the commandline.
Everything that comes after is a preferred (or a "tagged") directory. If there
are duplicates from an unpreferred and from a preferred directory, the preferred
one will always count as original. Special options can also be used to always
Expand Down Expand Up @@ -108,14 +108,14 @@ General Options
:``-o --output=spec`` / ``-O --add-output=spec`` (**default\:** *-o sh\:rmlint.sh -o pretty\:stdout -o summary\:stdout -o json\:rmlint.json*):

Configure the way ``rmlint`` outputs its results. A ``spec`` is in the form
``format:file`` or just ``format``. A ``file`` might either be an
arbitrary path or ``stdout`` or ``stderr``. If file is omitted, ``stdout``
``format:file`` or just ``format``. A ``file`` might either be an
arbitrary path or ``stdout`` or ``stderr``. If file is omitted, ``stdout``
is assumed. ``format`` is the name of a formatter supported by this
program. For a list of formatters and their options, refer to the
**Formatters** section below.

If ``-o`` is specified, rmlint's default outputs are overwritten. With
``--O`` the defaults are preserved. Either ``-o`` or ``-O`` may be
If ``-o`` is specified, rmlint's default outputs are overwritten. With
``-O`` the defaults are preserved. Either ``-o`` or ``-O`` may be
specified multiple times to get multiple outputs, including multiple
outputs of the same format.

Expand Down Expand Up @@ -154,7 +154,7 @@ General Options

Choose the algorithm to use for finding duplicate files. The algorithm can be
either **paranoid** (byte-by-byte file comparison) or use one of several file hash
algorithms to identify duplicates. The following hash families are available (in
algorithms to identify duplicates. The following hash families are available (in
approximate descending order of cryptographic strength):

**sha3**, **blake**,
Expand Down Expand Up @@ -380,7 +380,7 @@ Traversal Options
:``-n --newer-than-stamp=<timestamp_filename>`` / ``-N --newer-than=<iso8601_timestamp_or_unix_timestamp>``:

Only consider files (and their size siblings for duplicates) newer than a
certain modification time (*mtime*). The age barrier may be given as
certain modification time (*mtime*). The age barrier may be given as
seconds since the epoch or as ISO8601-Timestamp like
*2014-09-08T00:12:32+0200*.

Expand All @@ -392,7 +392,7 @@ Traversal Options
``-N``, in contrast, takes the timestamp directly and will not write anything.

Note that ``rmlint`` will find duplicates newer than ``timestamp``, even if
the original is older. If you want only find duplicates where both
the original is older. If you want only find duplicates where both
original and duplicate are newer than ``timestamp`` you can use
``find(1)``:

Expand Down Expand Up @@ -662,7 +662,7 @@ FORMATTERS
``--reflink`` in ``man 1 cp``. Fails if the filesystem does not support
it.
* ``hardlink``: Replace the duplicate file with a hardlink to the original
file. The resulting files will have the same inode number. Fails if both
file. The resulting files will have the same inode number. Fails if both
files are not on the same partition. You can use ``ls -i`` to show the
inode number of a file and ``find -samefile <path>`` to find all
hardlinks for a certain file.
Expand Down Expand Up @@ -772,7 +772,7 @@ OTHER STAND-ALONE COMMANDS
This will only work when ``Shredder`` and its dependencies were installed.
See also: http://rmlint.readthedocs.org/en/latest/gui.html

The gui has its own set of options, see ``--gui --help`` for a list. These
The gui has its own set of options, see ``--gui --help`` for a list. These
should be placed at the end, ie ``rmlint --gui [options]`` when calling
it from commandline.

Expand All @@ -781,7 +781,7 @@ OTHER STAND-ALONE COMMANDS
Make ``rmlint`` work as a multi-threaded file hash utility, similar to the
popular ``md5sum`` or ``sha1sum`` utilities, but faster and with more algorithms.
A set of paths given on the commandline or from *stdin* is hashed using one
of the available hash algorithms. Use ``rmlint --hash -h`` to see options.
of the available hash algorithms. Use ``rmlint --hash -h`` to see options.

:``rmlint --equal [paths...]``:

Expand All @@ -792,7 +792,7 @@ OTHER STAND-ALONE COMMANDS
etc.) will be executed by default. At least two paths need to be passed.

Note: This even works for directories and also in combination with paranoid
mode (pass ``-pp`` for byte comparison); remember that rmlint does not care
mode (pass ``-p`` for byte comparison); remember that rmlint does not care
about the layout of the directory, but only about the content of the files
in it. At least two paths need to be given to the commandline.

Expand Down Expand Up @@ -859,7 +859,7 @@ This is a collection of common use cases and other tricks:

* Compare files byte-by-byte in current directory:

``$ rmlint -pp .``
``$ rmlint -p .``

* Find duplicates with same basename (excluding extension):

Expand Down Expand Up @@ -919,7 +919,7 @@ This is a collection of common use cases and other tricks:

* Produce a list of files that are unique, including original files ("one of each"):

``$ rmlint t -o json -o uniques:unique_files | jq -r '.[1:-1][] | select(.is_original) | .path' | sort > original_files``
``$ rmlint t -o json -o uniques:unique_files | jq -r '.[1:-1][] | select(.is_original) | .path' | sort > original_files``
``$ cat unique_files original_files``

* Sort files by a user-defined regular expression
Expand Down Expand Up @@ -950,13 +950,13 @@ PROBLEMS
The default hash function (blake2b) is very safe but in theory it is possible for
two files to have then same hash. If you had 10^73 different files, all the same
size, then the chance of a false positive is still less than 1 in a billion.
If you're concerned just use the ``--paranoid`` (``-pp``)
If you're concerned just use the ``--paranoid`` (``-p``)
option. This will compare all the files byte-by-byte and is not much slower than
blake2b (it may even be faster), although it is a lot more memory-hungry.

2. **File modification during or after rmlint run:** It is possible that a file
that ``rmlint`` recognized as duplicate is modified afterwards, resulting in
a different file. If you use the rmlint-generated shell script to delete
a different file. If you use the rmlint-generated shell script to delete
the duplicates, you can run it with the ``-p`` option to do a full re-check
of the duplicate against the original before it deletes the file. When using
``-c sh:hardlink`` or ``-c sh:symlink`` care should be taken that
Expand Down Expand Up @@ -1012,4 +1012,4 @@ PROGRAM AUTHORS
* Christopher <sahib> Pahl 2010-2017 (https://github.com/sahib)
* Daniel <SeeSpotRun> T. 2014-2017 (https://github.com/SeeSpotRun)

Also see the http://rmlint.rtfd.org for other people that helped us.
Also see the http://rmlint.rtfd.org for other people that helped us.