9 comments

  • 827a 21 minutes ago
    There is no legitimate reason why postinstall scripts need to exist. The npm team needs to grow up and declare "starting with npm version whatever, npm will only run postinstall scripts for versions of packages published before ${today}".
    • raggi 11 minutes ago
      install scripts are a distraction, just like package signatures are a distraction. adding/removing either feature has no significant impact on the wormability of this package ecosystem. installed npm code is run, with nearly zero exceptions.
    • nine_k 2 minutes ago
      ...and only if you invoke it with --dangerously-run-postinstall-scripts; otherwise it will report an error if a postinstall script is found.

      This is definitely going to affect any packages that need to link to native code and/or compile shims, but these are very few.

    • Rohansi 13 minutes ago
      This doesn't really fix the issue though because package code is also executed at build time and during testing. Just maybe restricts the scope a little bit.
  • btown 57 minutes ago
  • aselimov3 49 minutes ago
    What are the actual guarantees that go/Rust make that Python/npm don’t? It seems like it might just be that Python/npm are juicier targets? I’m starting to try and avoid all third party packages
    • nirvdrum 3 minutes ago
      Part of the point the article makes is that most other popular languages have a comprehensive standard library. JS has an astonishingly small on. Rather than have one vetted set of libraries that ship with the language, applications either need to roll it themselves or pull from a 3rd party package repository. We've drilled NIH into people, so they tend to reach for packages. That's not necessarily a bad thing, but it often means they're pulling in more code than they need. The JS ecosystem has also favored smaller modules, so you need many of them. And everyone builds on top of that, leading to massive growth in dependency graphs. It's a huge surface area for things to go wrong, intentionally or not.

      With many other languages, you have a lot of functionality out of the box. Certainly, there have been bugs and security issues, but they're a drop in the bucket compared to what you see in the JS ecosystem. With other languages, you have a much smaller external dependency graph and the core functionality is coming from a trusted 3rd party.

    • brunoborges 14 minutes ago
      It is 100% up to the package manager's steward to control how ownership of packages and namespaces are granted.

      Maven Central exists for decades the amount of incidents of people stealing namespaces is minimal.

      One can't simply publish a package under the groupId "com.ycombinator" without having some way to verify that they own the domain ycombinator.com. Then, once a package is published, it is 100% immutable, even if it has malicious code in it. Certainly, that library is flagged everywhere as vulnerable.

      It baffles me that NPM for so long couldn't replicate the same guardrails as Maven Central.

    • lostglass 23 minutes ago
      To be honest Rust has the exact same supply chain attack pattern - it's just newer and more maintained at the moment. Give it a decade.
      • nothinkjustai 21 minutes ago
        Rust doesn’t have post install scripts
        • est31 10 minutes ago
          There is build.rs, proc macros are unsandboxed, and lastly you install the binary so that you can run it. Even if the build and install were fully sandboxed, the binary could still do malicious stuff if ran.
        • fabrice_d 18 minutes ago
          It has build.rs that will run as soon as you compile the dependency. That's not the same thing but pretty close to a post install script: it's very likely to run.
        • tasn 18 minutes ago
          It has build.rs, which has essentially the same problems.
        • deeebug 18 minutes ago
    • jollyllama 25 minutes ago
      > It seems like it might just be that Python/npm are juicier targets?

      Attackers go where the victims are. Frontend is a monoculture with the vast majority using NPM; backend, less so. This isn't an excuse for NPM, but another strike against it.

      You could also argue that the attacks make a deeper point about frontend vs backend devs, but I won't go there.

    • panzi 25 minutes ago
      Last I checked npm had 2FA for publishing, but cargo didn't. I don't think cargo is any better than npm, just not that of an attractive target.
    • raggi 11 minutes ago
      none. they just have smaller target populations.
    • cookiengineer 41 minutes ago
      I suppose that go's go:generate workflow can also be abused to land a worm like the ones spreading via npm, as you can build programs that just scrape the whole hard drive for git projects and patch the go.mod dependencies there, and you could also just write this in go as a toolchain script, for example.

      NPM's achilles is the pre/postinstall step which can run arbitrary commands and shell scripts without the user having any way to intervene.

      Dependencies must be run in isolated chroot sandboxes or better, inside containers. That would be the only way to mitigate this problem, as the filesystem of the operating system must be separated from the filesystem of the development workflow.

      On top of that most host based firewalls are per-binary instead of per-cmdline. That leads to the warnings and rules relying on that e.g. "python" or "nodejs" getting network access allowlisted, instead of say "nodejs myworm.js". So firewalls in general are pretty useless against this type of malware.

      • yegle 25 minutes ago
        `go:generate` is for the package provider, the command never runs when someone `go install` or `go get` the package.
        • cookiengineer 16 minutes ago
          Note that the NPM worms are spreading because the package providers are developing on their libraries without them noticing a malicious dependency. It is not users/consumers spreading the worm, it is developers spreading it.

          Your mismatch is that you think in policies, not assessments here. Nothing in my normal go workflow will ask me if I want to run "curl download whatever from the internet" when I run go build.

          Though I agree with the difference in workflow, there is not a single mechanism in go catching this. go.mod files can be just patched by the worm, and/or hidden behind a /v123 folder or whatever to play shenanigans on API differences.

      • xena 25 minutes ago
        go:generate is done at dev time, not at build time.
        • cookiengineer 23 minutes ago
          Actually bindings are usually generated like that, at build time (though with a build cache that nobody knows how it corrupts all the time).

          Examples that come to mind: webview/webview, webkit, cilium/ebpf and most other CGo projects that I have seen.

    • jiggawatts 18 minutes ago
      Generally, other package managers aren't great either. Notably, crates.io / cargo has some of the same people behind it as NPM and the verbiage of their excuses is oddly similar.

      Something fascinating about the design and architecture of programming languages and their surrounding ecosystems is the enormous leverage that they provide to the "core team":

      For every 1 core language developer[1]...

      ... there may be 1,000 popular package developers...

      ... for which there may be 1,000,000 developers writing software...

      ... for over 1,000,000,000 users.

      This means that for every corner that is cut at the top of that pyramid, the harms are massively magnified at the lower tiers. A security vulnerability in a "top one thousand" package like log4j can cause billions of dollars in economic damage, man-centuries of remediation effort, etc.

      However, bizarrely, the funding at the top two levels is essentially a pittance! Most such projects are charities, begging for spare change with hat in hand on a street corner. Some of the most used libraries are often volunteer efforts! cough-OpenSSL-cough.

      The result is that the people most empowered to fix the issues are the least funded to do so.

      This is why NPM, Crates.io, etc... flatly refuse to do even the most basic security checks like adding namespaces and verifying the identity of major publishers like Google, Microsoft, and the like.

      That's a non-zero amount of effort, and no matter how trivial to implement technically and now cheap to police, it would likely blow their tiny budget of unreliable donations.

      The exceptions to this rule are package-managers with robust financial backing, such as NuGet, which gets reliable funding from Microsoft and supports their internal (for-profit!) workflows almost as much as it does external "free" users.

      "Free and open" is wonderful and all, but you get what you pay for.

      [1] Most of us can name them off the top of our heads: Guido van Rossum, Larry Wall, Kerningham & Richie, etc.

  • eulgro 5 minutes ago
    These satire articles on cybersecurity are really entertaining.

    The other one a few days ago was also good: https://nesbitt.io/2026/02/03/incident-report-cve-2024-yikes...

  • p-e-w 51 minutes ago
    With the recent high-profile attacks on PyPI packages, it’s no longer true that npm is the “only package manager where this regularly happens”.

    In fact, pip is much more dangerous than npm because it lacks a lockfile. uv fixes that, but adoption is proceeding at a snail’s pace.

    • godzillabrennus 32 minutes ago
      UV adoption is happening, though. NPM is still the only name in town.
      • manquer 24 minutes ago
        Huh ? uv is a package manager not a registry.

        In JS world there is plenty of competition for package managers pnpm/ yarn/ burn all viable alternatives to npm the package manager.

        Public registries for languages tend to coalesce around one service . Nobody wants to publish their library to 4 different registries .

    • esafak 18 minutes ago
    • fragmede 24 minutes ago
      I don't know about snails, but everything I'm in contact with has moved over to uv, and I can't imagine I'm the only one.
  • exabrial 47 minutes ago
    I really don't understand why the npm project cannot embrace PGP as an ambulatory 'good enough' solution.
    • loloquwowndueo 45 minutes ago
      The NIH mentality in the ecosystem would result in a JavaScript pgp library which itself would be an npm package and subject to supply chain attacks. lol.
      • panzi 28 minutes ago
        A good part of it is already implemented in web crypto, which is supported by browsers and node. There is a chance that npm could implement something there without extra dependencies. Maybe I'm too optimistic?
    • Gigachad 32 minutes ago
      Would that help? Most of these recent attacks, the attackers have gained access to the system that builds the packages. So it would have just signed the malicious build the same.
      • raggi 8 minutes ago
        nope, doesn't help. signatures and removal of script points have zero net effect on the value of the target that the ecosystem has, or how easy/hard it is to write a worm. the package code gets run, this is statistically true, and the exploited developers/environments will sign packages, this is also statistically true.
  • qrush 1 hour ago
    [flagged]
    • rileymat2 49 minutes ago
      I read it as a comparison of the attitude of helplessness around it, not the acts themselves. So it was a bit meta, but unremarkably inoffensive.
    • mikepurvis 48 minutes ago
      I don't think it's comparing them directly or arguing for equivalent seriousness. It is identifying a similarity of mindset where those who have their hands on the levers of power that could materially improve the situation act like there's nothing they can do.
    • mrandish 44 minutes ago
      But it's not comparing to school shootings, it's satirizing supposedly responsible parties who continue to deny responsibility despite repeated catastrophic failures which are their responsibility.
    • p-e-w 47 minutes ago
      You’re right. Major supply chain attacks affect far more people than school shootings do, and can potentially cost more lives through downstream effects.

      It’s 2026. Software is critical infrastructure for global civilization now. Lives and livelihoods depend on it working reliably. The “it’s just bits on a computer” quip has been outdated for 20 years now.

  • yegle 23 minutes ago
    Vendorizing using git submodule should be a robust mitigation for this problem.
    • raggi 7 minutes ago
      subtree is better for this case, you want to encourage actual reading before running. reading won't catch everything but it catches a lot, and the burden isn't as high as people always complain about before they try it.