Setting up Python to the point to be able install packages from PyPI can be annoying and time-intensive. Even worse are OS-provided installations that start throwing cryptic error messages. Especially desktops are prone to that but it’s possible to break the whole toolchain of a server by installing some shiny package you heard about on reddit.
Your desktop system is unlikely to be a throwaway virtual machine or container. That makes it a highly mutable system with difficult rollbacks and a lot of pain if stuff breaks. So, until we all run NixOS on our desktops:
Does that sound extreme to you? Only if you haven’t found the right tools to make it effortless.
virtualenv – or as of Python 3.3 venv – has been around for a while and was a somewhat accepted standard for installing Python software. Sadly, there are many missionaries boldly proclaiming the end of virtualenv. Mostly because of containers in general and usually because of Docker in particular.
I find that unfortunate and shortsighted, because they fail to see the whole picture: virtualenv’s job isn’t just to separate your projects from each other. Its job is also to separate you from the operating system’s Python installation and the installed packages you probably have no idea about.
Let’s use the widely celebrated virtualenv-killer Docker to see why that’s a good idea. For that we look at the pre-installed packages you get after installing only
python-pip into a Ubuntu Trusty container1:
argparse (1.2.1) chardet (2.0.1) colorama (0.2.5) html5lib (0.999) pip (1.5.4) requests (2.2.1) setuptools (3.3) six (1.5.2) urllib3 (1.7.1) wsgiref (0.1.2)
Such things can happen any time and make your system fragile. Fully featured Ubuntu servers carry even more baggage of course. Whenever you install a system tool written in Python you can expect some kind of breakage. Whenever debian packagers decide that they don’t like something about how pip works and patch around it you’re involuntarily part of the “will it explode?” lottery.
macOS is no different and comes with several dozens of Python packages.
And on the desktop – no matter what platform! – the situation is even worse. I dare to say that the average
site-packages is a mess and most users have no idea why a certain package is installed. The step to break the whole installation is a very short one as many tutorial mentors will confirm.
site-packagesbelongs to the operating system.
I’ve been saying for a while now that I would prefer if OS vendors would create a virtualenv for their stuff somewhere else and let the users have the system
site-packages. But that’s not happening and no one can guarantee that some system tool you don’t even know about won’t ever install a version of a library that’s incompatible with your project’s requirements. Is it worth it to take that chance?
Only put software into
site-packages that is explicitly written for that version of the OS. In other words: system tools that only use Python packages provided by the OS. Keep everything else in virtualenv.
Stop discussing virtualenv versus system isolation as if they were mutually exclusive. You should use both at once, neither replaces the other:
- Do isolate your application server’s OS from its host using Docker / LXC / jails / zones / KVM / Xen /VMware / … to one container/virtual machine per application.
- But inside of them also do isolate your Python environment using virtualenv from unexpected surprises in the system
Less Typing, More Happiness
On your desktop you’ll want a bit more convenience than pure virtualenvs. So I urge you take five minutes to install and understand direnv, virtualenvwrapper, virtualenvwrapper-win, or virtualfish – depending on the shell and operating system you use. They take all the hassle from managing per-project virtualenvs.
Servers are a bit different because you’re unlikely to have more than one application (i.e. virtualenv) per user or even per server. You can have a look at my current approach to packaging and installing virtualenvs of server applications if you’re interested.
What About Python CLI Tools?
There’s one question that arises from this: what about all the amazing Python-based tools we love? How do we install tox, httpie, Pygments, and so forth? Arguably, they made the biggest mess in my system installations in the past.
Should you create virtualenvs for them all and link the executable scripts into some directory within you
After installation you can install Python CLI tools by calling e.g.
$ pipx install Pygments
pipx then will create a new virtualenv and install the package into it. Finally it helps you adding the installed executables into your search path so you can run
pygmentize from your shell, no matter what path you are in.
Preventing Accidental Installations
To enforce the virtualenv-only policy on your desktop, set the the environment variable
PIP_REQUIRE_VIRTUALENV=1 and pip will refuse to install packages if you’re not in an active virtualenv.
[install] require-virtualenv = true [uninstall] require-virtualenv = true
into your pip configuration file.
It saved me more often than I care to admit.
Make sure to disable it when installing Python versions using e.g. asdf.
Otherwise you’ll get a cryptic
ERROR: Could not find an activated virtualenv (required).
pip install --user?
Installing into your user-specific
site-packages is better than installing it into the global one, because you can easily nuke them all, without impacting the system-wide health of your Python installations.
Python’s inability to install more than one package version at a time still leads to version conflicts, therefore I recommend to avoid it too.
Built from an official trusty image as of 2014-09-14. ↩︎
The reason for the presence of so many high profile packages is that debian decided to recursively unwrap all dependencies that are bundled within pip (including requests). They’re begging for version conflicts. ↩︎
I used to recommend
pipsifor this purpose but that has been unmaintained for a while and
pipxis superior in every way at this point. Windows support being a notable example. ↩︎