A reported Free Download Manager supply chain attack redirected Linux users to a malicious Debian package repository that installed information-stealing malware.

The malware used in this campaign establishes a reverse shell to a C2 server and installs a Bash stealer that collects user data and account credentials.

Kaspersky discovered the potential supply chain compromise case while investigating suspicious domains, finding that the campaign has been underway for over three years.

    • @[email protected]
      link
      fedilink
      439 months ago

      And via a website too. That’s like pushing a car. One of the main strengths of Linux are open repositories, maintained by reputable sources and checked by thousands of reputable people. Packages are checksummed and therefore unable to be switched by malicious parties. Even the AUR is arguably a safer and more regulated source. And it’s actually in there.

      • JWBananas
        link
        fedilink
        English
        519 months ago

        And via a website too

        Everyone knows real admins do curl https://raw.githubusercontent.com/something/or/other/install.sh | sudo bash

    • @[email protected]
      link
      fedilink
      119 months ago

      It’s still my favorite download manager on Windows. It often downloads file significantly faster than the download manager built into browsers. Luckily I never installed it on Linux, since I have a habit of only installing from package managers.

      Do you know of a good download manager for Linux?

      • @[email protected]
        link
        fedilink
        English
        149 months ago

        How much faster are we talking?

        I’ve honestly never looked at my downloads and though huh you should be quicker, well maybe in 90’s.

        • no banana
          link
          fedilink
          79 months ago

          Right? I’ve not thought about download speeds since the 2000’s.

        • @[email protected]
          link
          fedilink
          79 months ago

          FDM does some clever things to boost download speeds. It splits up a download into different chuncks, and somehow downloads them concurrently. It makes a big difference for large files (for example, Linux ISOs).

          • @[email protected]
            link
            fedilink
            English
            99 months ago

            Im curious as to how it would achieve that?

            It can’t split a file before it has the file. And all downloads are split up. They’re called packets.

            Not saying it doesn’t do it, just wondering how.

            • @[email protected]
              link
              fedilink
              139 months ago

              It could make multiple requests to the server, asking each request to resume starting at a certain byte.

              • @[email protected]
                link
                fedilink
                English
                59 months ago

                Interesting.

                I feel I’ll save this rabbit hole for weekend and go and have a look at what they do.

                • @[email protected]OP
                  link
                  fedilink
                  189 months ago

                  The key thing to know is that a client can do an HTTP HEAD request to get just the Content-Length of the file, and then perform GET requests with the Range request header to fetch a specific chunk of a file.

                  This mechanism was introduced in HTTP 1.1 (byte-serving).

        • arglebargle
          link
          fedilink
          English
          69 months ago

          just grabbed a gig file - it would take about 8 minutes with a standard download in Firefox. Use a manager or axel and it will be 30 seconds. Then again speed isnt everything, its also nice to be able to have auto retry and completion.

          • PenguinCoder
            link
            fedilink
            English
            19 months ago

            I was just going to recommend this too; Use axel, aria2 or even ancient hget.

      • @[email protected]
        link
        fedilink
        89 months ago

        JDownloader, XDM, FileCentipede (this one is the closest to IDM, although it uses closed source libraries), kGet, etc.

      • arglebargle
        link
        fedilink
        English
        49 months ago

        axel. use axel -n8 to make 8 connections/segments which it will assemble when it is done

      • @[email protected]
        link
        fedilink
        -19 months ago

        Xtreme Download Manager. Very similar to Internet Download Manager on Windows.

        Also, use either of those two. FDM is very meh.

    • u/lukmly013 (lemmy.sdf.org)
      link
      fedilink
      109 months ago

      Gotta admit, it was me. I’ve only used a computer for short time.
      I’ve got my first laptop 3 years ago, and that broke after just 2 months. And anyway, with AMD Athlon 64 it greatly struggled with a browser. So really I only started seriously using computer at the start of 2021, when I got another, usable laptop. And that’s when I downloaded freedownloadmanager.deb. Thankfully, I didn’t get that redirect, so it was a legitimate file.

    • @[email protected]
      link
      fedilink
      89 months ago

      Oh, I know someone who adds the word “free” to various search words like “free pdf reader” or “free flash player” (happened a very long time ago). He’s also the kind of person who I can imagine having a bunch of viruses and malware on his computer.

    • @[email protected]
      link
      fedilink
      59 months ago

      People not well versed in Linux.

      You know, the non-techies, which the Linux community claims should know such things but obviously does not.

    • @[email protected]
      link
      fedilink
      29 months ago

      I’ve installed and used it, and still do.

      My internet connection is not that reliable, and when I download big files that are not torrents (say >1000 MB) and the download is interrupted because of internet disconnect, Firefox often has trouble getting back to it while FDM doesn’t.

      FDM also lets me set download speed limits, which means I can still browse the internet while downloading.

      It’s not my main tool for downloading stuff, but it has its uses.

  • @[email protected]OP
    link
    fedilink
    569 months ago

    The article mentions how to check for infection:

    If you have installed the Linux version of the Free Download Manager between 2020 and 2022, you should check and see if the malicious version was installed.

    To do this, look for the following files dropped by the malware, and if found, delete them:

    /etc/cron.d/collect
    /var/tmp/crond
    /var/tmp/bs
    
  • Hairyblue
    link
    fedilink
    359 months ago

    What is a free download manager and why would someone need one?

    • lemmyvore
      link
      fedilink
      English
      239 months ago

      Back in the day when most stuff was on FTP and HTTP and your connection was crap and could drop at any time, you’d use a download manager to smooth things along. It could resume downloads when connection dropped, it could keep a download going for days on end and resume as needed, and it could abusing the bandwitdh limitations of the source site by using multiple parallel connections that pulled on different file chunks. In some ways it was very similar to how we use BT today.

      It was also useful to keep a history of stuff you’d downloaded in case you needed it again, manage the associated files etc.

      • @[email protected]OP
        link
        fedilink
        169 months ago

        and it could abusing the bandwitdh limitations of the source site by using multiple parallel connections that pulled on different file chunks

        Also for files which had multiple different mirror sites you could download chunks from multiple mirrors concurrently which would allow you to max out your bandwidth even if individual mirrors were limiting download speeds.

    • Dhs92
      link
      169 months ago

      It’s a download client that can pause/Resume downloads, as well as use multiple connections to download files

      • Hairyblue
        link
        fedilink
        49 months ago

        Like a BitTorrent?

        I guess I just don’t download that much stuff.

        • @[email protected]
          link
          fedilink
          English
          109 months ago

          Sucks having your connection drop and having to redlownload the entire thing again. Managers are a fix.

        • @[email protected]
          link
          fedilink
          39 months ago

          BitTorrent works in chunks basically, or can download it nonlinearly. Downloading from a site in a basic way gets the file from start to finish, the download manager can let you stop it and pick up where you left off, as long as the server you’re getting the file from is configured to allow it.

          https://github.com/agalwood/Motrix

          (Note: I don’t use that or any other download manager and haven’t since Windows 95, it’s linked as example only)

    • @[email protected]
      link
      fedilink
      6
      edit-2
      9 months ago

      Back in the 2000s, browsers were really bad at downloading big things over slow connections since they couldn’t resume, a brief disconnect could destroy hours of progress. But I don’t think you need this anymore.

  • @[email protected]
    link
    fedilink
    259 months ago

    How is it possible that users noticed strange behaviors (new Cron jobs) and they didn’t check the script launched by those jobs 😱

            • @[email protected]
              link
              fedilink
              English
              49 months ago

              No, thats a crown.

              A cron is a type of super virus that wants to destroy the entire net. An end to all things… Total crash. Only another virus superpowered by core energy can put a stop to it.

              • @[email protected]
                link
                fedilink
                2
                edit-2
                9 months ago

                No that is a Cron-virus of the show ReBoot that lasted from 1994-2001.

                Cron is just an other word for hag.

                • Zloubida
                  link
                  fedilink
                  English
                  19 months ago

                  No that’s a crone.

                  To cron is to sing romantic style of pop music.

        • @[email protected]M
          link
          fedilink
          5
          edit-2
          9 months ago

          They actually are, kind of. It’s called Tron: Ares and it’s been in production hell for some years, the most recent delay being due to the ongoing writer’s strike. Filming is expected to start after the strike is over, but personally my enthusiasm for the movie died after they announced Jared Leto as one of the cast.

      • @[email protected]
        link
        fedilink
        7
        edit-2
        9 months ago

        If they were complaining about cronjobs being created (like the post says), then they must have known what cron is.

  • @[email protected]
    link
    fedilink
    8
    edit-2
    9 months ago

    Mmmh. You kinda deserve being infected if you do things like this. Every beginner tutorial specifically tells you not to download random stuff from the internet and ‘sudo’ install it. Every Wiki with helpful information has these boxes that tell you not to do it. I’m okay if you do it anyways. But don’t blame anyone else for the consequences. And don’t tell me you haven’t been warned.

    Also I wonder about the impact this had. It went unnoticed for 3 years. So I can’t imagine it having affected many people. The text says it affected few people. And it didn’t have any real impact.

    But supply chain attacks are real. Don’t get fooled. And don’t install random stuff. Install the download manager from your package repository instead.

    • @[email protected]
      link
      fedilink
      89 months ago

      I kind of disagree. Applications often require root permissions to install themselves, since regular users can’t access certain folders like /opt, etc.

      Also, do you really think that people would actually read the source and then compile all their software themselves? Do you do the same?

      Generally though I do agree, you’re probably fine installing software from your distro’s repos but even that’s not bulletproof and also it’s not like third-party repos are uncommon either.

      • @[email protected]
        link
        fedilink
        1
        edit-2
        9 months ago

        Yes. I do it the correct way. I use my favourite distro’s package manager to install software. This way it’s tested, a few people had a look at the changes, and sometimes a CI script automatically determines if the installer affects other parts of the system. I go to great lengths to avoid doing it any other way. (I’ve been using some flatpaks in recent times, though. But sometimes I also install it only for a separate user account. Mainly when it’s proprietary or niche.)

        It is super rare that I install random stuff from the internet. Or ‘curl’ and then pipe the installer script into a root shell. And when I do, I put in some effort to see if it’s okay. I think i had a quick glance at most of the install .sh scripts before continuing. So yes, I kinda do my best. And I isolate that stuff and don’t put it on the same container that does my email.

        Most of the times you can avoid doing it the ‘stupid way’. And even the programming package managers like ‘npm’, ‘cargo’, … have started to take supply chain attacks seriously.

  • _cnt0
    link
    fedilink
    79 months ago

    malicious Debian package repository

    *laughs in RPM*

    This comment was presented by the fedora gang.

    • @[email protected]
      link
      fedilink
      139 months ago

      Right, but you could do the same with RPM. Not everyone is aware of this, but installing a package executes scripts with root access over your system.

  • @[email protected]
    link
    fedilink
    69 months ago

    I had to essentially read the same thing four times before there was any new information in this post. Not sure if that’s a Jerboa thing or what, but probably could have been avoided.

    • @[email protected]OP
      link
      fedilink
      89 months ago

      Yeah I agree, sorry about that. I thought that the body-text field was mandatory to fill in, so I used the introductory paragraph from the article so as not to editorialize.

    • PenguinCoder
      link
      fedilink
      English
      79 months ago

      curl --proto '=https' --tlsv1.2 -sSf https://download-more-ram.sh | sh

      PHEW thanks, I’m safe.