PEP 395 – Qualified Names for Modules
- Author:
- Nick Coghlan <ncoghlan at gmail.com>
- Status:
- Withdrawn
- Type:
- Standards Track
- Created:
- 04-Mar-2011
- Python-Version:
- 3.4
- Post-History:
- 05-Mar-2011, 19-Nov-2011
Table of Contents
- PEP Withdrawal
- Abstract
- What’s in a
__name__
? - Traps for the Unwary
- Qualified Names for Modules
- Eliminating the Traps
- Explicit relative imports
- Reference Implementation
- References
- Copyright
PEP Withdrawal
This PEP was withdrawn by the author in December 2013, as other significant changes in the time since it was written have rendered several aspects obsolete. Most notably PEP 420 namespace packages rendered some of the proposals related to package detection unworkable and PEP 451 module specifications resolved the multiprocessing issues and provide a possible means to tackle the pickle compatibility issues.
A future PEP to resolve the remaining issues would still be appropriate, but it’s worth starting any such effort as a fresh PEP restating the remaining problems in an updated context rather than trying to build on this one directly.
Abstract
This PEP proposes new mechanisms that eliminate some longstanding traps for the unwary when dealing with Python’s import system, as well as serialisation and introspection of functions and classes.
It builds on the “Qualified Name” concept defined in PEP 3155.
Relationship with Other PEPs
Most significantly, this PEP is currently deferred as it requires significant changes in order to be made compatible with the removal of mandatory __init__.py files in PEP 420 (which has been implemented and released in Python 3.3).
This PEP builds on the “qualified name” concept introduced by PEP 3155, and also shares in that PEP’s aim of fixing some ugly corner cases when dealing with serialisation of arbitrary functions and classes.
It also builds on PEP 366, which took initial tentative steps towards making explicit relative imports from the main module work correctly in at least some circumstances.
Finally, PEP 328 eliminated implicit relative imports from imported modules.
This PEP proposes that the de facto implicit relative imports from main
modules that are provided by the current initialisation behaviour for
sys.path[0]
also be eliminated.
What’s in a __name__
?
Over time, a module’s __name__
attribute has come to be used to handle a
number of different tasks.
The key use cases identified for this module attribute are:
- Flagging the main module in a program, using the
if __name__ == "__main__":
convention. - As the starting point for relative imports
- To identify the location of function and class definitions within the running application
- To identify the location of classes for serialisation into pickle objects which may be shared with other interpreter instances
Traps for the Unwary
The overloading of the semantics of __name__
, along with some historically
associated behaviour in the initialisation of sys.path[0]
, has resulted in
several traps for the unwary. These traps can be quite annoying in practice,
as they are highly unobvious (especially to beginners) and can cause quite
confusing behaviour.
Why are my imports broken?
There’s a general principle that applies when modifying sys.path
: never
put a package directory directly on sys.path
. The reason this is
problematic is that every module in that directory is now potentially
accessible under two different names: as a top level module (since the
package directory is on sys.path
) and as a submodule of the package (if
the higher level directory containing the package itself is also on
sys.path
).
As an example, Django (up to and including version 1.3) is guilty of setting
up exactly this situation for site-specific applications - the application
ends up being accessible as both app
and site.app
in the module
namespace, and these are actually two different copies of the module. This
is a recipe for confusion if there is any meaningful mutable module level
state, so this behaviour is being eliminated from the default site set up in
version 1.4 (site-specific apps will always be fully qualified with the site
name).
However, it’s hard to blame Django for this, when the same part of Python
responsible for setting __name__ = "__main__"
in the main module commits
the exact same error when determining the value for sys.path[0]
.
The impact of this can be seen relatively frequently if you follow the “python” and “import” tags on Stack Overflow. When I had the time to follow it myself, I regularly encountered people struggling to understand the behaviour of straightforward package layouts like the following (I actually use package layouts along these lines in my own projects):
project/
setup.py
example/
__init__.py
foo.py
tests/
__init__.py
test_foo.py
While I would often see it without the __init__.py
files first, that’s a
trivial fix to explain. What’s hard to explain is that all of the following
ways to invoke test_foo.py
probably won’t work due to broken imports
(either failing to find example
for absolute imports, complaining
about relative imports in a non-package or beyond the toplevel package for
explicit relative imports, or issuing even more obscure errors if some other
submodule happens to shadow the name of a top-level module, such as an
example.json
module that handled serialisation or an
example.tests.unittest
test runner):
# These commands will most likely *FAIL*, even if the code is correct
# working directory: project/example/tests
./test_foo.py
python test_foo.py
python -m package.tests.test_foo
python -c "from package.tests.test_foo import main; main()"
# working directory: project/package
tests/test_foo.py
python tests/test_foo.py
python -m package.tests.test_foo
python -c "from package.tests.test_foo import main; main()"
# working directory: project
example/tests/test_foo.py
python example/tests/test_foo.py
# working directory: project/..
project/example/tests/test_foo.py
python project/example/tests/test_foo.py
# The -m and -c approaches don't work from here either, but the failure
# to find 'package' correctly is easier to explain in this case
That’s right, that long list is of all the methods of invocation that will almost certainly break if you try them, and the error messages won’t make any sense if you’re not already intimately familiar not only with the way Python’s import system works, but also with how it gets initialised.
For a long time, the only way to get sys.path
right with that kind of
setup was to either set it manually in test_foo.py
itself (hardly
something a novice, or even many veteran, Python programmers are going to
know how to do) or else to make sure to import the module instead of
executing it directly:
# working directory: project
python -c "from package.tests.test_foo import main; main()"
Since the implementation of PEP 366 (which defined a mechanism that allows
relative imports to work correctly when a module inside a package is executed
via the -m
switch), the following also works properly:
# working directory: project
python -m package.tests.test_foo
The fact that most methods of invoking Python code from the command line break when that code is inside a package, and the two that do work are highly sensitive to the current working directory is all thoroughly confusing for a beginner. I personally believe it is one of the key factors leading to the perception that Python packages are complicated and hard to get right.
This problem isn’t even limited to the command line - if test_foo.py
is
open in Idle and you attempt to run it by pressing F5, or if you try to run
it by clicking on it in a graphical filebrowser, then it will fail in just
the same way it would if run directly from the command line.
There’s a reason the general “no package directories on sys.path
”
guideline exists, and the fact that the interpreter itself doesn’t follow
it when determining sys.path[0]
is the root cause of all sorts of grief.
In the past, this couldn’t be fixed due to backwards compatibility concerns.
However, scripts potentially affected by this problem will already require
fixes when porting to the Python 3.x (due to the elimination of implicit
relative imports when importing modules normally). This provides a convenient
opportunity to implement a corresponding change in the initialisation
semantics for sys.path[0]
.
Importing the main module twice
Another venerable trap is the issue of importing __main__
twice. This
occurs when the main module is also imported under its real name, effectively
creating two instances of the same module under different names.
If the state stored in __main__
is significant to the correct operation
of the program, or if there is top-level code in the main module that has
non-idempotent side effects, then this duplication can cause obscure and
surprising errors.
In a bit of a pickle
Something many users may not realise is that the pickle
module sometimes
relies on the __module__
attribute when serialising instances of arbitrary
classes. So instances of classes defined in __main__
are pickled that way,
and won’t be unpickled correctly by another python instance that only imported
that module instead of running it directly. This behaviour is the underlying
reason for the advice from many Python veterans to do as little as possible
in the __main__
module in any application that involves any form of
object serialisation and persistence.
Similarly, when creating a pseudo-module (see next paragraph), pickles rely on the name of the module where a class is actually defined, rather than the officially documented location for that class in the module hierarchy.
For the purposes of this PEP, a “pseudo-module” is a package designed like
the Python 3.2 unittest
and concurrent.futures
packages. These
packages are documented as if they were single modules, but are in fact
internally implemented as a package. This is supposed to be an
implementation detail that users and other implementations don’t need to
worry about, but, thanks to pickle
(and serialisation in general),
the details are often exposed and can effectively become part of the public
API.
While this PEP focuses specifically on pickle
as the principal
serialisation scheme in the standard library, this issue may also affect
other mechanisms that support serialisation of arbitrary class instances
and rely on __module__
attributes to determine how to handle
deserialisation.
Where’s the source?
Some sophisticated users of the pseudo-module technique described
above recognise the problem with implementation details leaking out via the
pickle
module, and choose to address it by altering __name__
to refer
to the public location for the module before defining any functions or classes
(or else by modifying the __module__
attributes of those objects after
they have been defined).
This approach is effective at eliminating the leakage of information via
pickling, but comes at the cost of breaking introspection for functions and
classes (as their __module__
attribute now points to the wrong place).
Forkless Windows
To get around the lack of os.fork
on Windows, the multiprocessing
module attempts to re-execute Python with the same main module, but skipping
over any code guarded by if __name__ == "__main__":
checks. It does the
best it can with the information it has, but is forced to make assumptions
that simply aren’t valid whenever the main module isn’t an ordinary directly
executed script or top-level module. Packages and non-top-level modules
executed via the -m
switch, as well as directly executed zipfiles or
directories, are likely to make multiprocessing on Windows do the wrong thing
(either quietly or noisily, depending on application details) when spawning a
new process.
While this issue currently only affects Windows directly, it also impacts any proposals to provide Windows-style “clean process” invocation via the multiprocessing module on other platforms.
Qualified Names for Modules
To make it feasible to fix these problems once and for all, it is proposed
to add a new module level attribute: __qualname__
. This abbreviation of
“qualified name” is taken from PEP 3155, where it is used to store the naming
path to a nested class or function definition relative to the top level
module.
For modules, __qualname__
will normally be the same as __name__
, just
as it is for top-level functions and classes in PEP 3155. However, it will
differ in some situations so that the above problems can be addressed.
Specifically, whenever __name__
is modified for some other purpose (such
as to denote the main module), then __qualname__
will remain unchanged,
allowing code that needs it to access the original unmodified value.
If a module loader does not initialise __qualname__
itself, then the
import system will add it automatically (setting it to the same value as
__name__
).
Alternative Names
Two alternative names were also considered for the new attribute: “full name”
(__fullname__
) and “implementation name” (__implname__
).
Either of those would actually be valid for the use case in this PEP.
However, as a meta-issue, PEP 3155 is also adding a new attribute (for
functions and classes) that is “like __name__
, but different in some cases
where __name__
is missing necessary information” and those terms aren’t
accurate for the PEP 3155 function and class use case.
PEP 3155 deliberately omits the module information, so the term “full name”
is simply untrue, and “implementation name” implies that it may specify an
object other than that specified by __name__
, and that is never the
case for PEP 3155 (in that PEP, __name__
and __qualname__
always
refer to the same function or class, it’s just that __name__
is
insufficient to accurately identify nested functions and classes).
Since it seems needlessly inconsistent to add two new terms for attributes
that only exist because backwards compatibility concerns keep us from
changing the behaviour of __name__
itself, this PEP instead chose to
adopt the PEP 3155 terminology.
If the relative inscrutability of “qualified name” and __qualname__
encourages interested developers to look them up at least once rather than
assuming they know what they mean just from the name and guessing wrong,
that’s not necessarily a bad outcome.
Besides, 99% of Python developers should never need to even care these extra attributes exist - they’re really an implementation detail to let us fix a few problematic behaviours exhibited by imports, pickling and introspection, not something people are going to be dealing with on a regular basis.
Eliminating the Traps
The following changes are interrelated and make the most sense when considered together. They collectively either completely eliminate the traps for the unwary noted above, or else provide straightforward mechanisms for dealing with them.
A rough draft of some of the concepts presented here was first posted on the python-ideas list ([1]), but they have evolved considerably since first being discussed in that thread. Further discussion has subsequently taken place on the import-sig mailing list ([2]. [3]).
Fixing main module imports inside packages
To eliminate this trap, it is proposed that an additional filesystem check be
performed when determining a suitable value for sys.path[0]
. This check
will look for Python’s explicit package directory markers and use them to find
the appropriate directory to add to sys.path
.
The current algorithm for setting sys.path[0]
in relevant cases is roughly
as follows:
# Interactive prompt, -m switch, -c switch
sys.path.insert(0, '')
# Valid sys.path entry execution (i.e. directory and zip execution)
sys.path.insert(0, sys.argv[0])
# Direct script execution
sys.path.insert(0, os.path.dirname(sys.argv[0]))
It is proposed that this initialisation process be modified to take package details stored on the filesystem into account:
# Interactive prompt, -m switch, -c switch
in_package, path_entry, _ignored = split_path_module(os.getcwd(), '')
if in_package:
sys.path.insert(0, path_entry)
else:
sys.path.insert(0, '')
# Start interactive prompt or run -c command as usual
# __main__.__qualname__ is set to "__main__"
# The -m switches uses the same sys.path[0] calculation, but:
# modname is the argument to the -m switch
# modname is passed to ``runpy._run_module_as_main()`` as usual
# __main__.__qualname__ is set to modname
# Valid sys.path entry execution (i.e. directory and zip execution)
modname = "__main__"
path_entry, modname = split_path_module(sys.argv[0], modname)
sys.path.insert(0, path_entry)
# modname (possibly adjusted) is passed to ``runpy._run_module_as_main()``
# __main__.__qualname__ is set to modname
# Direct script execution
in_package, path_entry, modname = split_path_module(sys.argv[0])
sys.path.insert(0, path_entry)
if in_package:
# Pass modname to ``runpy._run_module_as_main()``
else:
# Run script directly
# __main__.__qualname__ is set to modname
The split_path_module()
supporting function used in the above pseudo-code
would have the following semantics:
def _splitmodname(fspath):
path_entry, fname = os.path.split(fspath)
modname = os.path.splitext(fname)[0]
return path_entry, modname
def _is_package_dir(fspath):
return any(os.exists("__init__" + info[0]) for info
in imp.get_suffixes())
def split_path_module(fspath, modname=None):
"""Given a filesystem path and a relative module name, determine an
appropriate sys.path entry and a fully qualified module name.
Returns a 3-tuple of (package_depth, fspath, modname). A reported
package depth of 0 indicates that this would be a top level import.
If no relative module name is given, it is derived from the final
component in the supplied path with the extension stripped.
"""
if modname is None:
fspath, modname = _splitmodname(fspath)
package_depth = 0
while _is_package_dir(fspath):
fspath, pkg = _splitmodname(fspath)
modname = pkg + '.' + modname
return package_depth, fspath, modname
This PEP also proposes that the split_path_module()
functionality be
exposed directly to Python users via the runpy
module.
With this fix in place, and the same simple package layout described earlier, all of the following commands would invoke the test suite correctly:
# working directory: project/example/tests
./test_foo.py
python test_foo.py
python -m package.tests.test_foo
python -c "from .test_foo import main; main()"
python -c "from ..tests.test_foo import main; main()"
python -c "from package.tests.test_foo import main; main()"
# working directory: project/package
tests/test_foo.py
python tests/test_foo.py
python -m package.tests.test_foo
python -c "from .tests.test_foo import main; main()"
python -c "from package.tests.test_foo import main; main()"
# working directory: project
example/tests/test_foo.py
python example/tests/test_foo.py
python -m package.tests.test_foo
python -c "from package.tests.test_foo import main; main()"
# working directory: project/..
project/example/tests/test_foo.py
python project/example/tests/test_foo.py
# The -m and -c approaches still don't work from here, but the failure
# to find 'package' correctly is pretty easy to explain in this case
With these changes, clicking Python modules in a graphical file browser
should always execute them correctly, even if they live inside a package.
Depending on the details of how it invokes the script, Idle would likely also
be able to run test_foo.py
correctly with F5, without needing any Idle
specific fixes.
Optional addition: command line relative imports
With the above changes in place, it would be a fairly minor addition to allow
explicit relative imports as arguments to the -m
switch:
# working directory: project/example/tests
python -m .test_foo
python -m ..tests.test_foo
# working directory: project/example/
python -m .tests.test_foo
With this addition, system initialisation for the -m
switch would change
as follows:
# -m switch (permitting explicit relative imports)
in_package, path_entry, pkg_name = split_path_module(os.getcwd(), '')
qualname= <<arguments to -m switch>>
if qualname.startswith('.'):
modname = qualname
while modname.startswith('.'):
modname = modname[1:]
pkg_name, sep, _ignored = pkg_name.rpartition('.')
if not sep:
raise ImportError("Attempted relative import beyond top level package")
qualname = pkg_name + '.' modname
if in_package:
sys.path.insert(0, path_entry)
else:
sys.path.insert(0, '')
# qualname is passed to ``runpy._run_module_as_main()``
# _main__.__qualname__ is set to qualname
Compatibility with PEP 382
Making this proposal compatible with the PEP 382 namespace packaging PEP is
trivial. The semantics of _is_package_dir()
are merely changed to be:
def _is_package_dir(fspath):
return (fspath.endswith(".pyp") or
any(os.exists("__init__" + info[0]) for info
in imp.get_suffixes()))
Incompatibility with PEP 402
PEP 402 proposes the elimination of explicit markers in the file system for
Python packages. This fundamentally breaks the proposed concept of being able
to take a filesystem path and a Python module name and work out an unambiguous
mapping to the Python module namespace. Instead, the appropriate mapping
would depend on the current values in sys.path
, rendering it impossible
to ever fix the problems described above with the calculation of
sys.path[0]
when the interpreter is initialised.
While some aspects of this PEP could probably be salvaged if PEP 402 were adopted, the core concept of making import semantics from main and other modules more consistent would no longer be feasible.
This incompatibility is discussed in more detail in the relevant import-sig threads ([2], [3]).
Potential incompatibilities with scripts stored in packages
The proposed change to sys.path[0]
initialisation may break some
existing code. Specifically, it will break scripts stored in package
directories that rely on the implicit relative imports from __main__
in
order to run correctly under Python 3.
While such scripts could be imported in Python 2 (due to implicit relative imports) it is already the case that they cannot be imported in Python 3, as implicit relative imports are no longer permitted when a module is imported.
By disallowing implicit relatives imports from the main module as well, such modules won’t even work as scripts with this PEP. Switching them over to explicit relative imports will then get them working again as both executable scripts and as importable modules.
To support earlier versions of Python, a script could be written to use different forms of import based on the Python version:
if __name__ == "__main__" and sys.version_info < (3, 3):
import peer # Implicit relative import
else:
from . import peer # explicit relative import
Fixing dual imports of the main module
Given the above proposal to get __qualname__
consistently set correctly
in the main module, one simple change is proposed to eliminate the problem
of dual imports of the main module: the addition of a sys.metapath
hook
that detects attempts to import __main__
under its real name and returns
the original main module instead:
class AliasImporter:
def __init__(self, module, alias):
self.module = module
self.alias = alias
def __repr__(self):
fmt = "{0.__class__.__name__}({0.module.__name__}, {0.alias})"
return fmt.format(self)
def find_module(self, fullname, path=None):
if path is None and fullname == self.alias:
return self
return None
def load_module(self, fullname):
if fullname != self.alias:
raise ImportError("{!r} cannot load {!r}".format(self, fullname))
return self.main_module
This metapath hook would be added automatically during import system initialisation based on the following logic:
main = sys.modules["__main__"]
if main.__name__ != main.__qualname__:
sys.metapath.append(AliasImporter(main, main.__qualname__))
This is probably the least important proposal in the PEP - it just
closes off the last mechanism that is likely to lead to module duplication
after the configuration of sys.path[0]
at interpreter startup is
addressed.
Fixing pickling without breaking introspection
To fix this problem, it is proposed to make use of the new module level
__qualname__
attributes to determine the real module location when
__name__
has been modified for any reason.
In the main module, __qualname__
will automatically be set to the main
module’s “real” name (as described above) by the interpreter.
Pseudo-modules that adjust __name__
to point to the public namespace will
leave __qualname__
untouched, so the implementation location remains readily
accessible for introspection.
If __name__
is adjusted at the top of a module, then this will
automatically adjust the __module__
attribute for all functions and
classes subsequently defined in that module.
Since multiple submodules may be set to use the same “public” namespace,
functions and classes will be given a new __qualmodule__
attribute
that refers to the __qualname__
of their module.
This isn’t strictly necessary for functions (you could find out their module’s qualified name by looking in their globals dictionary), but it is needed for classes, since they don’t hold a reference to the globals of their defining module. Once a new attribute is added to classes, it is more convenient to keep the API consistent and add a new attribute to functions as well.
These changes mean that adjusting __name__
(and, either directly or
indirectly, the corresponding function and class __module__
attributes)
becomes the officially sanctioned way to implement a namespace as a package,
while exposing the API as if it were still a single module.
All serialisation code that currently uses __name__
and __module__
attributes will then avoid exposing implementation details by default.
To correctly handle serialisation of items from the main module, the class
and function definition logic will be updated to also use __qualname__
for the __module__
attribute in the case where __name__ == "__main__"
.
With __name__
and __module__
being officially blessed as being used
for the public names of things, the introspection tools in the standard
library will be updated to use __qualname__
and __qualmodule__
where appropriate. For example:
pydoc
will report both public and qualified names for modulesinspect.getsource()
(and similar tools) will use the qualified names that point to the implementation of the code- additional
pydoc
and/orinspect
APIs may be provided that report all modules with a given public__name__
.
Fixing multiprocessing on Windows
With __qualname__
now available to tell multiprocessing
the real
name of the main module, it will be able to simply include it in the
serialised information passed to the child process, eliminating the
need for the current dubious introspection of the __file__
attribute.
For older Python versions, multiprocessing
could be improved by applying
the split_path_module()
algorithm described above when attempting to
work out how to execute the main module based on its __file__
attribute.
Explicit relative imports
This PEP proposes that __package__
be unconditionally defined in the
main module as __qualname__.rpartition('.')[0]
. Aside from that, it
proposes that the behaviour of explicit relative imports be left alone.
In particular, if __package__
is not set in a module when an explicit
relative import occurs, the automatically cached value will continue to be
derived from __name__
rather than __qualname__
. This minimises any
backwards incompatibilities with existing code that deliberately manipulates
relative imports by adjusting __name__
rather than setting __package__
directly.
This PEP does not propose that __package__
be deprecated. While it is
technically redundant following the introduction of __qualname__
, it just
isn’t worth the hassle of deprecating it within the lifetime of Python 3.x.
Reference Implementation
None as yet.
References
Copyright
This document has been placed in the public domain.
Source: https://github.com/python/peps/blob/main/pep-0395.txt
Last modified: 2022-09-16 05:15:18 GMT