So Podman is an open source container engine like Docker—with "full"1 Docker compatibility. IMO Podman’s main benefit over Docker is security. But how is it more secure? Keep reading…
Docker traditionally runs a daemon as the root user, and you need to mount that daemon’s socket into various containers for them to work as intended (See: Traefik, Portainer, etc.) But if someone compromises such a container and therefore gains access to the Docker socket, it’s game over for your host. That Docker socket is the keys to the root kingdom, so to speak.
Podman doesn’t have a daemon by default, although you can run a very minimal one for Docker compatibility. And perhaps more importantly, Podman can run entirely as a non-root user.2 Non-root means if someone compromises a container and somehow manages to break out of it, they don’t get the keys to the kingdom. They only get access to your non-privileged Unix user. So like the keys to a little room that only contains the thing they already compromised.2.5 Pretty neat.
Okay, now for the annoying parts of Podman. In order to achieve this rootless, daemonless nirvana, you have to give up the convenience of Unix users in your containers being the same as the users on the host. (Or at least the same UIDs.) That’s because Podman typically3 runs as a non-root user, and most containers expect to either run as root or some other specific user.
The "solution"4 is user re-mapping. Meaning that you can configure your non-root user that Podman is running as to map into the container as the root user! Or as UID 1234. Or really any mapping you can imagine. If that makes your head spin, wait until you actually try to configure it. It’s actually not so bad on containers that expect to run as root. You just map your non-root user to the container UID 0 (root)… and Bob’s your uncle. But it can get more complicated and annoying when you have to do more involved UID and GID mappings—and then play the resultant permissions whack-a-mole on the host because your volumes are no longer accessed from a container running as host-root…
Still, it’s a pretty cool feeling the first time you run a “root” container in your completely unprivileged Unix user and everything just works. (After spending hours of swearing and Duck-Ducking to get it to that point.) At least, it was pretty cool for me. If it’s not when you do it, then Podman may not be for you.
The other big annoying thing about Podman is that because there’s no Big Bad Daemon managing everything, there are certain things you give up. Like containers actually starting on boot. You’d think that’d be a fundamental feature of a container engine in 2023, but you’d be wrong. Podman doesn’t do that. Podman adheres to the “Unix philosophy.” Meaning, briefly, if Podman doesn’t feel like doing something, then it doesn’t. And therefore expects you to use systemd for starting your containers on boot. Which is all good and well in theory, until you realize that means Podman wants you to manage your containers entirely with systemd. So… running each container with a systemd service, using those services to stop/start/manage your containers, etc.
Which, if you ask me, is totally bananasland. I don’t know about you, but I don’t want to individually manage my containers with systemd. I want to use my good old trusty Docker Compose. The good news is you can use good old trusty Docker Compose with Podman! Just run a compatibility daemon (tiny and minimal and rootless… don’t you worry) to present a Docker-like socket to Compose and boom everything works. Except your containers still don’t actually start on boot. You still need systemd for that. But if you make systemd run Docker Compose, problem solved!
This isn’t the “Podman Way” though, and any real Podman user will be happy to tell you that. The Podman Way is either the aforementioned systemd-running-the-show approach or something called Quadlet or even a Kubernetes compatibility feature. Briefly, about those: Quadlet is “just” a tighter integration between systemd and Podman so that you can declaratively define Podman containers and volumes directly in a sort of systemd service file. (Well, multiple.) It’s like Podman and Docker Compose and systemd and Windows 3.1 INI files all had a bastard love child—and it’s about as pretty as it sounds. IMO, you’d do well to stick with Docker Compose.
The Kubernetes compatibility feature lets you write Kubernetes-style configuration files and run them with Podman to start/manage your containers. It doesn’t actually use a Kubernetes cluster; it lets you pretend you’re running a big boy cluster because your command has the word “kube” in it, but in actuality you’re just running your lowly Podman containers instead. It also has the feel of being a dev toy intended for local development rather than actual production use.5 For instance, there’s no way to apply a change in-place without totally stopping and starting a container with two separate commands. What is this, 2003?
Lastly, there’s Podman Compose. It’s a third-party project (not produced by the Podman devs) that’s intended to support Docker Compose configuration files while working more “natively” with Podman. My brief experience using it (with all due respect to the devs) is that it’s total amateur hour and/or just not ready for prime time. Again, stick with Docker Compose, which works great with Podman.
Anyway, that’s all I’ve got! Use Podman if you want. Don’t use it if you don’t want. I’m not the boss of you. But you said you wanted content on Lemmy, and now you’ve got content on Lemmy. This is all your fault!
1 Where “full” is defined as: Not actually full.
2 Newer versions of Docker also have some rootless capabilities. But they’ve still got that stinky ol’ daemon.
2.5 It’s maybe not quite this simple in practice, because you’ll probably want to run multiple containers under the same Unix account unless you’re really OCD about security and/or have a hatred of the convenience of container networking.
3 You can run Podman as root and have many of the same properties as root Docker, but then what’s the point? One less daemon, I guess?
4 Where “solution” is defined as: Something that solves the problem while creating five new ones.
5 Spoiler: Red Hat’s whole positioning with Podman is like they see it is as a way for buttoned-up corporate devs to run containers locally for development while their “production” is running K8s or whatever. Personally, I don’t care how they position it as long as Podman works well to run my self-hosting shit…
One of the really nice side-effects of it running rootless is that you get all the benefits of it running as an actual Unix user.
For instance, you can set up wireguard with IP route to send all traffic from a given UID through the VPN.
Using that, I set up one user as the single user for running all the stuff I want to have VPN’d for outgoing connections, like *arr services, with absolutely no extra work. I don’t need to configure a specific container, I don’t need to change a docker-compose etc.
In rootful docker, I had to use a specific IP subnet to achieve the same, which was way more clunky.
Could you explain or show how to do that?
Yeah sure.
I’m going to assume you’re starting from the point of having a second linux user also set up to use rootless podman. That’s just following the same steps for setting up rootless podman as any other user, so there shouldn’t be too many problems there.
If you have wireguard set up and running already - i.e. with Mullvad VPN or your own VPN to a VPS - you should be able to run
ip link
to see a wireguard network interface. Mine is calledwg
. I don’t use wg-quick, which means I don’t have all my traffic routing through it by default. Instead, I use a systemd unit to bring up the WG interface and set up routing.I’ll also assume the UID you want to forward is
1001
, because that’s what I’m using. I’ll also useenp3s0
as the default network link, because that’s what mine is, but if yours iseth0
, you should use that. Finally, I’ll assume that192.168.0.0
is your standard network subnet - it’s useful to avoid routing local traffic through wireguard.YOUR_STATIC_EXTERNAL_IP#
should be whatever you get by callingcurl ifconfig.me
if you have a static IP - again, useful to avoid routing local traffic through wireguard. If you don’t have a static IP you can drop this line.[Unit] Description=Create wireguard interface After=network-online.target [Service] RemainAfterExit=yes ExecStart=/usr/bin/bash -c " \ /usr/sbin/ip link add dev wg type wireguard || true; \ /usr/bin/wg setconf wg /etc/wireguard/wg.conf || true; \ /usr/bin/resolvectl dns wg #PREFERRED_DNS#; \ /usr/sbin/ip -4 address add #WG_IPV4_ADDRESS#/32 dev wg || true; \ /usr/sbin/ip -6 address add #WG_IPV6_ADDRESS#/128 dev wg || true; \ /usr/sbin/ip link set mtu 1420 up dev wg || true; \ /usr/sbin/ip rule add uidrange 1001-1001 table 200 || true; \ /usr/sbin/ip route add #VPN_ENDPOINT# via #ROUTER_IP# dev enp3s0 table 200 || true; \ /usr/sbin/ip route add 192.168.0.0/24 via 192.168.0.1 dev enp3s0 table 200 || true; \ /usr/sbin/ip route add #YOUR_STATIC_EXTERNAL_IP#/32 via #ROUTER_IP# dev enp3s0 table 200 || true; \ /usr/sbin/ip route add default via #WG_IPV4_ADDRESS# dev wg table 200 || true; \ " ExecStop=/usr/bin/bash -c " \ /usr/sbin/ip rule del uidrange 1001-1001 table 200 || true; \ /usr/sbin/ip route flush table 200 || true; \ /usr/bin/wg set wg peer '#PEER_PUBLIC_KEY#' remove || true; \ /usr/sbin/ip link del dev wg || true; \ " [Install] WantedBy=multi-user.target
There’s a bit to go through here, so I’ll take you through why it works. Most of it is just setting up WG to receive/send traffic. The bits that are relevant are:
/usr/sbin/ip rule add uidrange 1001-1001 table 200 || true; \ /usr/sbin/ip route add #VPN_ENDPOINT# via #ROUTER_IP# dev enp3s0 table 200 || true; \ /usr/sbin/ip route add 192.168.0.0/24 via 192.168.0.1 dev enp3s0 table 200 || true; \ /usr/sbin/ip route add #YOUR_STATIC_EXTERNAL_IP#/32 via #ROUTER_IP# dev enp3s0 table 200 || true; \ /usr/sbin/ip route add default via #WG_IPV4_ADDRESS# dev wg table 200 || true; \
ip rule add uidrange 1001-1001 table 200
adds a new rule where requests from UID 1001 go through table 200. A table is a subset of ip routing rules that are only relevant to certain traffic.ip route add
ensures that traffic already going through the VPN - i.e. wireguard traffic - does. This is relevant for handshakes.ip route add 192.168.0.0/24 via 192.168.0.1 ...
is just excluding local traffic, as isip route add
Finally, we add
ip route add default via
which routes all traffic that didn’t match any of the above rules (local traffic, wireguard) to go to the wireguard interface. From there, WG handles all the rest, and passes returning traffic back.There’s going to be some individual tweaking here, but the long and short of it is, UID
1001
will have all their external traffic routed through WG. Any internal traffic between docker containers in a docker-compose should already be handled by podman pods and never reach the routing rules. Any traffic aimed at other services in the network - i.e.sonarr
callingsabnzbd
ortransmission
- will happen with a relevant local IP of the machine it’s hosted on, and so will also be skipped. Localhost is already handled by existingip route
rules, so you shouldn’t have to worry about that either.Hopefully that helps - sorry if it’s a bit confusing. I learned to set up my own IP routing to avoid
wg-quick
so that I could have greater control over the traffic flow, so this is quite a lot of my learning that I’m attempting to distill into one place.Thank you so much, i was just looking for a way to do this the other day!
Thank you very much!
Saving this for later, thank you very much for the detailed writeup. I might look into this for my main machine to partition the vpn tasks from the non-vpn tasks
You’re welcome. It has some really nice side-effects - i.e. if I want to quickly grab a file without it being from my normal IP, I can just SSH to the right user on my server and it just works - no configuration, no needing to interrupt other traffic.
Really good example of one way that Podman’s Unix Philosophy is actually helpful!
I guess this: “you run a “root” container in your completely unprivileged Unix user and everything just works” sounds like chroot. Also managing your container starts with systemd sounds pretty good to me because this is what systemd is designed for, dependencies between services, etc.
deleted by creator
Not to mention, there aren’t massive online repositories of “chroot images” you can pull down and update for every popular application you might want to run…
Fyi, you can create a pod, create containers in that pod and generate the systemd file for that. Now you have one service to start/stop everything.
That’s a great point, and honestly not something I’d thought of. Now I just need to give up my lovely Docker Compose YAML and dip my toes into
podman run
, Quadlet config, orpodman kube play
…There’s also podman-compose, which I’ve been using. It’s not quite feature complete, but it’s pretty close.
Yup, it’s briefly touched on in the original post above. Personally I had issues with it on my machine, but maybe I didn’t give it a fair shake.
I work somewhere that doesn’t have licensing with Docker Inc. And I work on a Mac. With Docker desktop out of the picture, I got some experience with the alternatives. I know this post is about the native implementation and not the VM one, but I just wanted to add my 2 cents:
Alternatives run by me: Podman, Rancher Desktop, Finch
Results:
- Podman uses a lot more energy on idle than Finch and Rancher. On AVG 4 more Wats on an M1. (Normal idle is about 5W, so 9 almost doubles it cutting greatly in my battery life)
- Podman and Finch are not compatible with some tools that expect a full docker sock. In my case the AWS CDK and SAM CLI have issues. (Which is fun as Finch is also made by AWS)
- Finch does not offer a sock at all
- Finch requires you to recreate the full VM when updated.
- If you really want to have a drop-in replacement for Docker Desktop, use Rancher Desktop. Rancher lacks in UI and the extension feature. But I never had issues with the sock, as I can run it with containerd.
- Finch has no UI
- Podman’s VM has clock drift if you put your machine in sleep. Only solution I found is to reboot the podman VM.
- Podman allows you to log in the VM with a command. I haven’t found a way on the others.
The clock drift issue has been resolved recently: https://github.com/containers/podman/issues/11541
That is awesome. I prefer podman, despite what my list might suggest.
Wow, 4 watts? That’s a lot. Any insight on what is taking up all this extra energy? I thought that podman would be thinner than docker honestly.
I did some
shallowdigging, and my guess is the virtual machine that is started for each.I see that the podman vm is a whole ass fedora image, at least back in 2021 when this article was written.
Rancher seems to use alpine if I understand the configuration correctly
Finch also uses fedora… I think. Their config is seemingly simple to the point it looks deceptive.
Oh, are you using Podman on windows? Yeah, it needs a virtual machine because it has to load the linux kernel. I would definitely believe that the windows version (or mac, I guess) of podman is way heavier than the alternatives on those platforms, but on linux it just ends up using the host kernel.
If you are doing this on linux, and still need to load a vm to use podman, that would be interesting. I haven’t run across that, but I haven’t been able to use podman too much.
You forgot the third option: A Mac ;)
How are mac’s with VMs? Windows has a bunch of Hypervisor stuff that makes them work pretty well, I don’t use macs, so I wouldn’t know how well they run VMs.
I won’t say it is the same. Mac doesn’t have subsystems, for example. But it allows for hypervisors to run. For a while apple had their own on intel CPUs, bootcamp. But they stopped updating it for the Mx line.
Nevertheless, Mac might not be Linux, it is still UNIX. So it can run qemu and all the tools built for qemu. Wine is also a tool to run windows specifically.
Didn’t know about Podman yet, I’ll probably give it a spin on some personal projects to get a bit more used to tools other than docker. Thanks for the great intro!
If you do try it out, let us know how totally obnoxious it is for you!
Sure thing, ranting is therapeutic and I wouldn’t miss the chance~
Honestly, I had to use podman at work due to… issues.
Its close enough to docker, that most docker commands will work just fine. You can even alias docker as podman, and things will for the most part, just work.
HOWEVER, there are some big changes and difference. First- podmon creates a systemctl for your containers, for starting them. This- is different, and if you forget to tell it to create the service- then your containers won’t start.
My personal opinion after using it for a few years? I strongly prefer docker. It gives me very few issues. I have spent too much time troubleshooting odd things podman does.
I’m just glad we now have multiple container engine choices. For a while there, Docker was the only game in town.
Don’t forget the multiple flavors of kubernetes too.
Say more…?
Kubernetes just runs docker containers, (and lots more).
Imagine a solution to manage containers on multiple hosts, with tons of redundancy. With network and storage automation built in.
That’s Kubernetes in a nutshell.
But, has a steep learning curve
Oh, thanks, but I’m familiar with Kubernetes. :) I was just asking what the multiple flavors of K8s have to do with Podman vs. Docker.
I see it as a feature that Podman containers are run via systemd. This makes their management consistent with the other systemd-managed services. Also, Docker does it own things with logs, while with systemd, the logs are managed in a consistent way as well.
Maybe you missed podman generate systemd? Podman will generate the systemd unit files for you.
For me, the two big benefits of podman are being able to run containers via systemd and improved security by being able to run them rootless.
I actually find this a huge problem. Not all distros are built around LSB, XDG, or FreeDesktop.org nor should they be since not everyone is running Linux as a workstation/PC replacement. While yes for the most part podman can be ran on the likes of Gentoo, Alpine, Arch and etc. It becomes a pain in the arse to decouple the tooling for podman away from freedesktop.org standards. Even more a pain in the arse for clustering options (e.g. podman-remote expects freedesktop.org norms, kubernetes expects docker containerd or freedesktop.org with podman, and nomad stack is just bulky vaporware).
The really sad part of this is that podman isn’t adding much of anything new that LXC or linux namespaces outside of not needing a daemon, allowing rootless execution (again because it doesn’t need a daemon) and giving ACLs around which OCI repos could be pulled from unlike docker’s wildcard by default. It shouldn’t be hard to do linux containerization without being tied to anything other than the linux kernel.
Thanks for the suggestion, but I actually tried
podman generate systemd
and it did not work at all with my containers; the resultant services did not start successfully and I wasn’t able to get them working with any amount of tweaking. (I’d tell you what the error messages were, but I’ve only recorded them in now-deleted comments on The Site That Shall Not Be Named.)Ok. I was already very familiar with systemd when I started using podman and didn’t have any trouble at that step. In my case, I created the systems unit files by hand.
That sounds like it’s probably the way to go…
The easiest is actually using Quadlet (integrated in Podman 4.x or later): https://www.redhat.com/sysadmin/quadlet-podman and write simple .container files.
Works very well and does auto-start, service dependencies and even container auto-updates.
Also running containers in Pods is a really nice way to handle virtual networking and allows you to manage all the containers in one go similar to docker-compose.
The container auto-updates feature does sound pretty slick. I kind of have that hacked together now with a systemd timer and Docker Compose, but it’s not pretty.
I’ve switched over my own server last week, using ansible to generate the systemd files, and it worked great. It’s just a dozen containers or so.
The only problems I had were with container interdependencies (network-mode=container:x). That didn’t work so well with systemd, restarting and updating, but when I used a pod instead these problems all went away.
So I can’t say I regret my experience so far. Now I’ll be starting to use it at work too, where the user-namespace problem rears its head, but only because we have this very specific, very dumb big lamp dev container that houses apache, sql, redis, and more under one supervisord. That’s why we have more than one user in it and frankly that’s our own damn fault! When you make proper containers they shouldn’t have more than one user in it and then userns=keep-id should work just fine.
So far, I fully recommend podman.
Using Ansible to spew out systemd service boilerplate seems like a good idea. I’ll have to try that if I can ever give up my Docker Compose security blanket. And I wish you luck with your mega-container Podman conversion. That one sounds like it’ll be… a learning experience.
I understand very well wanting to stay with the declarative nature of docker-compose. Someone should really build a better podman-compose. (or sooner or later I’ll do it myself >_<)
Do it! I think there’s a market there. Although the “Podman Compose” name is taken and you’ll have to think of something else…
The other big annoying thing about Podman is that because there’s no Big Bad Daemon managing everything, there are certain things you give up. Like containers actually starting on boot. […] until you realize that means Podman wants you to manage your containers entirely with systemd. So… running each container with a systemd service, using those services to stop/start/manage your containers, etc.
Surprisingly, they have a solution for that that doesn’t involve using systemd for everything. They put an --all option to podman start, and a systemd service to run it at boot with the correct --filter (yeah. because unix philosophy). Debian seems to enable it by default AFAICT.
No idea how well it works rootless though.
Edit: Oh and for the networking, Podman 4.4.0 seems to ship pasta which seems to be the solution to slirp4netns’s existence. Unfortunately I have no idea if it works at all because I run Debian stable which is still on 4.3.1
Surprisingly, they have a solution for that that doesn’t involve using systemd for everything. They put an --all option to podman start, and a systemd service to run it at boot with the correct --filter (yeah. because unix philosophy). Debian seems to enable it by default AFAICT.
TIL! I wonder though why that isn’t working / auto-enabled on my system… Maybe just because I am doing rootless. Thanks for the tip! I’ll have to look into this.
Lurker. Never self hosted. Nothing useful to add here. Fedora Silverblue has a lot of integration with podman. It is the basis of toolbx, which has its issues, namely that the distro it spins up can’t be upgraded in each toolbx container. The tools are there to mess with containers stuff.
Makes sense… Fedora, Red Hat, Podman… All one big ecosystem. Fortunately for other users, Podman runs fine on other distros too.
IMO it is the integration with toolbx that is interesting to me, but I struggle with these things.
Well feel free to make a post about any problems you run into with it… Someone here may be able to help!
had a knowledge sharing meeting at work recently on container security. the guy was using podman like a docker cli, but it said “this is only an emulation of docker” - is there any downsides to running podman like this? im very familiar with docker on the command line
I think calling it an emulation downplays podman. Docker and podman are both container runtimes. Docker came first and is known synonymously with containers, whereas podman is newer and attempts to fix docker’s problems.
One outcome of this is podman chose to match docker’s cli very closely so nobody needs to learn a new cli. You can even put podman on the docker socket so “docker [command]” runs with podman.
So as far as I’m understanding your description, he’s actually using Podman under the hood, just via its Docker compatibility CLI. The main downside of that IMO is you lose out on Podman-specific flags and features. Which, honestly are probably not a huge deal to you if you just want a thing that walks and talks like Docker while hopefully being more secure.
The big caveat though is root vs. non-root. If he’s running his commands as root/sudo, it’ll work a whole lot like Docker just without a daemon (and without containers starting on boot). But if he’s running as a non-root user, well… See my original post for some downsides.
I’ve tried to switch in the past, but tripped over the differences in Podman vs Docker networking. IIRC Docker is better for creating an isolated network.
I have noticed that Docker doesn’t do the best job at graceful shutdowns (say for automatic installation of updates). I suspect Podman with systemd integration could do much much beter.
At least podman does not circumvent my firewall (ufw) like docker did. Had to use a workaround to get it to work with docker.
Podman respects UFW?? That’s awesome to hear.
Yep, turns out it doesn’t insert its own iptables rules like docker does, so the special rules from ufw-docker weren’t necessary anymore.
…and of course I learn this after switching to firewalld.
At least firewalld feels relatively painless compared to rest of the redhat-container-verse.
Yeah, one of my motivations for kicking Docker to the curb (beyond security) is all the weird little bugs: Preventing shutdown, unresponsive commands, random hangs, broken upgrades, etc. Podman may not end up being any better there, but at least I can pretend for a while.
Its not too bad, you can define the UID that your podmab process runs on since it can run as its own process and not be reliant on a daemon. It makes you do a little more admin work, but honestly you have to start doing that anyway in a lot of environments.
I have had good experimental experiences with podman, but my biggest barrier to adopting it is how old the packaged version with debian is. I’m hoping to see if I can do some damage with the release of bookworm.
Oh yeah, I hear you about that ancient Debian stable version of Podman. I actually ended up upgrading some of my servers to Bookworm for that very reason.
deleted by creator
Docker and k8s aren’t the same thing though? Docker is a container runtime while k8s is an orchestration platform.
deleted by creator
Docker and k8s aren’t the same thing though? Docker is a container runtime while k8s is an orchestration platform.
Awesome summary of how podman works.
I still haven’t figured out some issues with rootless podman where I pass the PUID and PGID of “myuser” 1000 as environment variable following the linuxserver.io examples… but then get files and folders owned by 100999:100999, if I chown files to “myuser” the service gets permission denied, I give up and chown everything to 100999 as workaround it works but is a bit annoying… Maybe someone here knows what’s going on?
as far as i remember, processes run inside container as root (0:0) end up under your own UID.
everything else gets mapped to those weird UIDs.
This is a consequence of user namespaces, which tripped me up until I read this article from Red Hat about running rootless containers as a non-root user. At that point I got that (the default options) map UID 0 in the container to my UID (i.e. 1000), but the other mappings were confusing.
The short version of the useful part (for me) of that article was
podman unshare
(man podman-unshare
), which launches a shell in a user namespace, like when you start a container. You can run the following command to see how the UIDs are mapped inside of the namespace:$ podman unshare cat /proc/self/uid_map 0 1000 1 1 100000 65536
This is read (for this purpose, see
man user_namespaces
for a more detailed explanation of this) as “inside this namespace, the UIDs in column 1 map to the UID in column 2 on the caller process, for (column 3) IDs”. There is alsogid_map
which works the same way, but for groups.The snippet above is from my machine, so in a podman container, UID 0 maps to UID 1000 on the “host”, which is me, and this is “good” for only 1 user. Then, starting with UID 1, the container maps to UID 100000 in the container, and is good for 65536 UIDs. This is why when you set the
PUID
andGUID
environment variables, on your filesystem you see the files are owned by100999:100999
- you can use the mapping to figure the math out:100000+1000-1=100999
.Since
podman unshare
puts you in a shell that has the same (? terminology might not be totally right here) user namesapce as your containers, you can use it for lots of stuff – like in your comment you mentioned usingchown
to change the permissions to100999:100999
. Instead, you could have usedpodman unshare chown 1000:1000
which have correctly set the permissions for your volume mount, and on your filesystem outside the container, the permissions would be100999:100999
.I still don’t know what “unshare” means.
It doesn’t mean anything useful… Its name is an artifact of its Unix roots that should be ignored IMO. For our purposes, it basically means “run this command in the same context as my containers run so that I can see what my containers actually see.”
Not a podman user, so please take this with a whole bag of salt. That seems to me a namespace issue. Does podman by default uses user namespaces? Because if that’s the case, it’s normal that UIDs are remapped inside the container namespace, and
1000
inside it corresponds to something else (maybe100999
?) outside.One way to check could be
cat /proc/PROC_INSIDE_CONTAINER/status | grep uid
orcat /proc/PROC_INSIDE_CONTAINER/uid_map
.This https://stackoverflow.com/questions/70770437/mapping-of-user-ids seems also to be somewhat relevant to your scenario?