From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.2 (2018-09-13) on inbox.vuxu.org X-Spam-Level: X-Spam-Status: No, score=-1.1 required=5.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,FREEMAIL_FROM,MAILING_LIST_MULTI,RCVD_IN_DNSWL_NONE autolearn=ham autolearn_force=no version=3.4.2 Received: from primenet.com.au (ns1.primenet.com.au [203.24.36.2]) by inbox.vuxu.org (OpenSMTPD) with ESMTP id 7650dd85 for ; Mon, 8 Apr 2019 15:12:16 +0000 (UTC) Received: (qmail 20485 invoked by alias); 8 Apr 2019 15:12:04 -0000 Mailing-List: contact zsh-users-help@zsh.org; run by ezmlm Precedence: bulk X-No-Archive: yes List-Id: Zsh Users List List-Post: List-Help: List-Unsubscribe: X-Seq: 23908 Received: (qmail 16232 invoked by uid 1010); 8 Apr 2019 15:12:04 -0000 X-Qmail-Scanner-Diagnostics: from mail-lf1-f49.google.com by f.primenet.com.au (envelope-from , uid 7791) with qmail-scanner-2.11 (clamdscan: 0.101.1/25412. spamassassin: 3.4.2. Clear:RC:0(209.85.167.49):SA:0(-2.3/5.0):. Processed in 3.433574 secs); 08 Apr 2019 15:12:04 -0000 X-Envelope-From: kolombo.inc@gmail.com X-Qmail-Scanner-Mime-Attachments: | X-Qmail-Scanner-Zip-Files: | Received-SPF: pass (ns1.primenet.com.au: SPF record at _netblocks.google.com designates 209.85.167.49 as permitted sender) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=subject:to:references:from:message-id:date:user-agent:mime-version :in-reply-to:content-transfer-encoding:content-language; bh=JxdREKIk5wNFjlnk+VnoAkyQqFZitVn62h2HI8AHmj0=; b=ZzhM0V0tOgKOeQ2qoO1Xrcm5hpT7QQIFPwJ41JF8xuIjx5pxmBuWE/FAz+ViL8J+8L 2dk6m4vY71rCoEdJYfXuPaPpEeXB7avWgacdLnBJpBg33bP/owHWipOqH7IZsFN2N7kQ HeOJfJXF7UikirDCcpaNTvv+q2O3u07LOefsFSl59fYvMt+Ccls0P0oVA7CeBaPvr0YB ksGn56w/IUwZtVRGLRENmEU0xreUPRNrgzl6sPjTk0LWbrrL1GylthTN4bDfT9+SJ95I d9qegiXNw9oOQ7CQApu648TFURx2+mh8KDTZPZMIysQS2Nv945ZZso/9GIHasI3gcJYE vCkw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-transfer-encoding :content-language; bh=JxdREKIk5wNFjlnk+VnoAkyQqFZitVn62h2HI8AHmj0=; b=t3wmgT1qE23GRpOP3+K16nooM6pLE/yIALI1srcxMi8+xPFeDYqg5C+KqTqb30UCMO gEPDw4nJYfsETZUy8l9spYuf8fH4KE50S2n4M4VeUNNzL9aRkMeC85jATRgys/cwdY0s RO0N2oI7Z90ouFvdPN0QKjIKXjQFtqtM+SMzU9eJ5nRpMrQFoGrngNZYYDI2LgHuwd3Q YaN5vgHC7nhxNvXIP4h1ni3YFOV7UjoJIx+KrqzQNtCqBONX5lWExYpopFwEsky6rsva i9n8ngkYFNxmLm0LKpN67Q/3Tv1FbK4bbsp2eCEESR6lx7iCdcJ/h0+e7Rs5bJhynS45 tPSg== X-Gm-Message-State: APjAAAULH/RVdyW2JtxNsr5VpdPTxbqooXmTP0ckVESQasYZTIFuGg8z Ow/BF17cuWnda3PDe8NkI6oiIyX2 X-Google-Smtp-Source: APXvYqyJOgrPMiYePWdvU7xnFF3AuPBKSgpT6tmcFxbTX7AavgIGA4rlZjASYB7qL1ta5reDBPer9w== X-Received: by 2002:a19:f203:: with SMTP id q3mr15685808lfh.78.1554736286106; Mon, 08 Apr 2019 08:11:26 -0700 (PDT) Subject: Re: find duplicate files To: zsh-users@zsh.org References: <86v9zrbsic.fsf@zoho.eu> <20190406130242.GA29292@trot> <86tvfb9ore.fsf@zoho.eu> <86mul2apj8.fsf@zoho.eu> <20190408143748.GA21630@trot> <391277a7-d604-4c20-a666-a2886b1d2939@eastlink.ca> From: Volodymyr Khomchak Message-ID: <35f7f111-f6f4-8df3-b19f-5177ac48e128@gmail.com> Date: Mon, 8 Apr 2019 18:14:24 +0300 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.5.1 MIME-Version: 1.0 In-Reply-To: <391277a7-d604-4c20-a666-a2886b1d2939@eastlink.ca> Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Content-Language: en-US Why wouldn't simple find work for this purpose ? This example can do what you need |find -not -empty -type f -printf "%s\n" | sort -rn | uniq -d | xargs -I{} -n1 find -type f -size {}c -print0 | xargs -0 md5sum | sort | uniq -w32 --all-repeated=separate If you want something dedicated then look at *fdupes* utility | On 4/8/19 5:58 PM, Ray Andrews wrote: > Pardon a comment from the peanut gallery, but it seems strange to me that the issue of duplicate files would be > something that would not have been solved definitively  50 years ago.  I'd have thought that for donkey's years there > would be utilities that would let you find duplicates at various levels of rigour because it's so obviously an issue > that's bound to come up.  No? >