tasiaiso.vulpecula.zone/docs/drafts/.curlpipebash.md
2024-05-07 13:02:54 +02:00

132 lines
6.4 KiB
Markdown

# DRAFT: curl | bash
In April of 2024 I wrote a post on Fediverse explaining that using `curl | bash` was not a security risk. A bit later, I debated on the same subject on a Matrix channel. The other parties involved caused me to do some further research on the subject and led me to review my opinion. As one could imagine, it turns out that the answer actually is, "it depends".
I based my original argument on the fact that you ultimately have to trust the person that provides you the code, which is true, but *incomplete*.
Here are the results of my reseach. At the end I will present a simple way we can use `curl | bash` safely.
TL;DR: If you're here because you just want to download software, go for it. You're *probably* going to be just fine. If you're interested in learning or want to implement a `curl | bash` script however, please read the rest.
## Terminology
Software artifact: Stuff that comes out of your repository: shell scripts, binaries, etc. In this article I will focus on the shell script that installs your binaries more than anything else.
## Surface attack
We can establish a simplified supply chain for a software artifact:
```
/----------\ /--------\ /--------\
| Artifact | ------>| Server | ------> | Client |
\----------/ | \--------/ | \--------/
(1) (2) (3) (4) (5)
```
An malicious actor could compromise the supply chain by attacking:
- (1): The machine the artifact is built on;
- (2): The connection beteen the artifact builder and the server;
- (3): The machine the server is served to client by;
- (4): The connection beteen the server and the client;
- (5): The client that requests the artifact.
For the purpose of this article however, the attack vectors (1), (2) and (5) are out of the scope of my research, which leaves us with only (3).
> There's not a lot that can be leveraged then ? So I'd imagine using `curl | bash` is safe *most of the time*.
Precisely.
## An example script
```bash
curl --proto '=https' --tlsv1.2 -sSf -L https://install.determinate.systems/nix | sh -s -- install
```
This script installs the Determinate Nix installer, a Nix-based package manager. We'll use this as an example for the rest of this article. Let's break it down a bit:
- `curl`: Call the cUrl commande line utility; This will create a request
- `--proto '=https'`:
- `--tlsv1.2`: Only connect to the server with a secure tunnel (TLS v1.2 or later);
- `-sSf -L`: Do not output progress updates, XXXX and folow redirections;
- `https://install.determinate.systems/nix`: The URL that points to an installation script;
- `|`: If `curl` gets the script successfully, pass it on to the next command;
- `sh`: Execute whatever `curl` gets from the server
- `-s`:
- `-- install`:
- ``:
We can see that the script explicitly requires `curl` to use a secure connection. At first glance, this seeems like a secure way to run the installer. However, if the server is compromised in some way, we could be downloading malware instead.
We can mitigate this risk by using a method used by most package managers, which is using 2 different servers with different functions: one that hosts the artifact's cryptographic hash or signature (here called *signing autohrity*), and another one that serves the artifact directly to us (here called *artifact provider*). This way, if either server is compromised, the software that's served to the client will not be verified and therefore not run.
We can reduce the risk of getting both machines compromised at once by:
- Having them be controlled by 2 different entities (companies and/or persons);
- Having them be managed by 2 different systems administrators;
- Using different data centers, network routes, domains and SSL certificates;
- Using different operating systems;
- Using different HTTP servers;
- Using different configurations;
This way, the only thing we have to trust is that the artifacts uploaded to the servers are healthy, and that **both** servers are not compromised at once (which should be overwhelmely unlikely if they are separate and different enough).
Now, our infrastructure looks like this:
```
/-----------\
| Signing |
/-> | authority | --\
/----------\ | \-----------/ | /--------\
| Artifact | ---- ----> | Client |
\----------/ | /-----------\ | \--------/
\-> | Artifact | --/
| provider |
\-----------/
```
There's still other parameters that I won't bother bringing into the picture right now, like the SSL certificates provider, and of course, the way the servers get the artifact in the first place (which depends on how your script is written and how your software is built).
An example infrastructure would look like this:
- Signing authority
- Managed by John Doe
- Hosted in DigitalOcean (Germany or Switzerland)
- NixOS
- HTTP server: Nginx
- Domain: `determinate.systems`
- Artifact provider
- Managed by Jane Poe
- Hosted by a worldwide CDN (Hetzner)
- RHEL
- HTTP server: Apache
- Domain: `install-determinate.systems` or `install.determinate.systems`
*This is not an endorsement for RHEL, Hetzner, Apache Web Server or even Determinate Systems; as of writing this, I've never tried them. I do very much endorse the use of NixOS however.*
Now, compromising this part of the supply chain has become extremely hard. The attacker will either:
- Need technical competency (TODO) in NixOS, RHEL, Nginx and Apache, as well as compromising an entire CDN (TODO);
- Compromise both of the sysadmin's machines through social engineering;
...
- Use several of the methods listed above.
Now, it would be a lot more feasible to attack another part of the supply chain, which is a subject for another article.
(more text)
## Implementing curl | bash safely
> You've spent so much time explaining that `curl | bash` is insecure, why would we bother making a secure version of it ?
Because the other way around this is to package your software for every distro and package manager under the sun, which is a task which simply imagining sends shivers down my spine.
Making a shell script that leverages this infrastructure isn't actually hard at all. Most of the work is around creating two resilient and independent servers. What we have to do is simply to check the artifact provider's response against a hash or a signature provided by the signing authority.
### Method 1: Hash
### Method 2: PGP signature