Python's importlib and inspect for static class members - python

In a long-running app I need to dynamically modify static class members based on path to the class' module and the class name.
Ex. I have a class pack1.mod1.Person and by definition I know it has a age property. So utilizing the importlib and inspect I try to load the class using the module path and class name and update the age property. It all seems fine until I read the the age property from my naturally imported Person class and find it's not updated.
Here are some more details:
.
├── app.py
└── pack1
├── __init__.py
└── mod1.py
mod1.py
class Person:
age = 42
app.py
import inspect
import os
from importlib import util
from pack1.mod1 import Person
if __name__ == '__main__':
Person.age = 3
print(Person.age) # => 3
spec = util.spec_from_file_location('pack1.mod1', os.path.join('pack1', 'mod1.py'))
module = util.module_from_spec(spec)
spec.loader.exec_module(module)
members = inspect.getmembers(module)
for x, member in inspect.getmembers(module, lambda i: inspect.isclass(i) and i.__name__ == Person.__name__):
print('Person:', Person.age) # => Person: 3
print('Person from inspect:', member.age) # => Person from inspect: 42
Person.age = 11
member.age = 66
print('Person:', Person.age) # => Person: 11
print('Person from inspect:', member.age) # => Person from inspect: 66
In the app.py I would expect member and Person to be the same thing but as the example shows they aren't.
What am I missing and how to achieve such an update on the static members of a class?

Python has no way of knowing that the regularly imported module and the manually module are "the same": Using util.spec_from_file_location up to spec.loader.exec_module side-steps Python's module registry and explicitly creates a new instance of the module.
Instead, use the native operations of the interpreter (import, ...) or their programmatic equivalents (importlib.load_module, ...)
If the module/class are well-known, one can import it regularly and directly inspect it.
import pack1.mod1
pack1.mod1.Person.age = 66
If module and class are only known by name, one can look them up from the existing modules.
import importlib
module_name, qualname, attribute, value = 'pack1.mod1', 'Person', 'age', 66
obj = importlib.import_module(module_name) # same as `import {module_name}`
for part in qualname.split('.'):
obj = getattr(obj, part) # same as `{obj}.{part}
setattr(obj, attribute, value) # same as `{obj}.{attribute} = {value}`

Related

How to access constant with class pointer in python - AUTH_USER_MODEL = 'django_restframework_2fa.User'

I have defined AUTH_USER_MODEL = 'django_restframework_2fa.User' in the settings.py file of Django application.
django_restframework_2fa is the name of the package and it has module named models where class User is defined.
Now, I want to access that User() class using AUTH_USER_MODEL constant. How can this be done?
I tried to access it like this settings.AUTH_USER_MODEL(**data) which should be equivalent to User(**data) but it doesn't work.
In python module can be imported using import_module() method of import importlib. Then to get the attributes of that module use method getattr() which takes two arguments module and attribute name.
This is how I have accomplished it -
def get_class(module_class_string):
"""
:param module_class_string: full name of the class to create an object of
:return: class
"""
module_name, class_name = module_class_string.rsplit(".", 1)
module_name = module_name + '.models'
module = importlib.import_module(module_name)
return getattr(module, class_name)

Does python support to import package

I am using Python 2.7, and have following code strucure
model
__init__.py
order.py
cart.py
That is, I define a package named model, and in this package, I define a module order, and I define a class in order.py
class MyOrder(object):
def __init__(self, name):
self.name = name
def getname(self):
return self.name
In the cart.py, the code is:
import model
x = model.order.MyOrder("Book")
print x.getname()
When I run it, it complains that AttributeError: 'module' object has no attribute 'order',
But the following is correct:
import model.order
x = model.order.MyOrder("Book")
print x.getname()
It looks that I can't import package (like import model) ?
If you want to have model automatically import order so it's available, you should do that in __init__.py. Simply put the following inside model/__init__.py:
from . import order
After that, you should be able to access model.order with just import model.

Passing a parameter between python classes

I'm attempting to write my first python module. The module is a wrapper for an api.
I'd like to allow users to set an api key in one of two ways: using a static configuration file or dynamically when they initiate the class.
How do I pass the key from __init__.py to functions.py if the key is in fact set dynamically?
File structure:
package/
__init__.py
config.py
functions.py
File contents:
__init__.py
import config
class MyClass(object):
def __init__(self, key):
self.key = key if key else config.key
...
functions.py
import config
class MyFunctions(object):
def __init__(self):
self.key = ?
self.base_url = config.base
def function1(self, my_id):
endpoint = urlencode({'id':my_id, 'key':self.key})
...
config.py
key = 'xxxxxxxxxxx'
base= 'http://xxxxx.com/api'
Assuming you want to set the same key for all the classes in the functions module, do so in the initialization of the package, overwriting the variables set by default in the config module.
Then import the config module and use the variables
config.py
key = 1
base_url = 'www.xyz.com/'
__init__.py
import config
import functions
class MyClass(object):
def __init__(self, key=config.key):
config.key = key
functions.py
import config
class MyFunctionsA(object):
def function1(self, my_id):
print(config.base_url)
print(config.key)
class MyFunctionsB(object):
def function1(self, my_id):
print(config.base_url)
print(config.key)
class MyFunctionsC(object):
def function1(self, my_id):
print(config.base_url)
print(config.key)
So you can use them as
>>> import package as pk
>>> pk.MyClass(9)
<package.MyClass object at 0x7f2133e96d10>
>>> f1 = pk.functions.MyFunctionsA()
>>> f2 = pk.functions.MyFunctionsB()
>>> f3 = pk.functions.MyFunctionsC()
>>> f1.function1(100)
www.xyz.com/
9
>>> f2.function1(100)
www.xyz.com/
9
>>> f3.function1(100)
www.xyz.com/
9
However, in case the key is an instance attribute of MyFunctions, let the users pass the key to its __init__, as usual.
Note: I ignore why you need to instantiate MyClass to set things up. It may be simpler to use a plain function instead.

Python metaClass and import *

Main Goal: Automatically register classes (by a string) in a factory to be created dynamically at run time using that string, classes can be in their own file and not grouped in one file.
I have couple of classes which all inherit from the same base class and they define a string as their type.
A user wants to get an instance of one of these classes but only knows the type at run time.
Therefore I have a factory to create an instance given a type.
I didn't want to hard code an "if then statements" so I have a meta class to register all the sub classes of the base class:
class MetaRegister(type):
# we use __init__ rather than __new__ here because we want
# to modify attributes of the class *after* they have been
# created
def __init__(cls, name, bases, dct):
if not hasattr(cls, 'registry'):
# this is the base class. Create an empty registry
cls.registry = {}
else:
# this is a derived class. Add cls to the registry
interface_id = cls().get_model_type()
cls.registry[interface_id] = cls
super(MetaRegister, cls).__init__(name, bases, dct)
The problem is that for this to work the factory has to import all the subclass (So the meta class runs).
To fix this you can use from X import *
But for this to work you need to define an __all__ var in the __init__.py file of the package to include all the sub classes.
I don't want to hard code the sub classes because it beats the purpose of using the meta class.
I can go over the file in the package using:
import glob
from os.path import dirname, basename, isfile
modules = glob.glob(dirname(__file__) + "/*.py")
__all__ = [basename(f)[:-3] for f in modules if isfile(f)]
Which works great, but the project needs to compile to a single .so file, which nullifies the use of the file system.
So how could I achieve my main goal of creating instances at run time without hard codding the type?
Is there a way to populate an __all__ var at run time without touching the filesystem?
In Java I'd probably decorate the class with an annotation and then get all the classes with that annotation at run time, is there something similar on python?
I know there are decorators in python but I'm not sure I can use them in this way.
Edit 1:
Each subclass must be in a file:
- Models
-- __init__.py
-- ModelFactory.py
-- Regression
--- __init__.py
--- Base.py
--- Subclass1.py
--- Subclass2ExtendsSubclass1.py
Edit 2: Some code to Illustrate the problem:
+ main.py
|__ Models
|__ __init__.py
|__ ModelFactory.py
|__ Regression
|__ init__.py
|__ Base.py
|__ SubClass.py
|__ ModelRegister.py
main.py
from models.ModelFactory import ModelFactory
if __name__ == '__main__':
ModelFactory()
ModelFactory.py
from models.regression.Base import registry
import models.regression
class ModelFactory(object):
def get(self, some_type):
return registry[some_type]
ModelRegister.py
class ModelRegister(type):
# we use __init__ rather than __new__ here because we want
# to modify attributes of the class *after* they have been
# created
def __init__(cls, name, bases, dct):
print cls.__name__
if not hasattr(cls, 'registry'):
# this is the base class. Create an empty registry
cls.registry = {}
else:
# this is a derived class. Add cls to the registry
interface_id = cls().get_model_type()
cls.registry[interface_id] = cls
super(ModelRegister, cls).__init__(name, bases, dct)
Base.py
from models.regression.ModelRegister import ModelRegister
class Base(object):
__metaclass__ = ModelRegister
def get_type(self):
return "BASE"
SubClass.py
from models.regression.Base import Base
class SubClass(Base):
def get_type(self):
return "SUB_CLASS"
Running it you can see only "Base" it printed.
Using a decorator gives the same results.
A simple way to register classes as runtime is to use decorators:
registry = {}
def register(cls):
registry[cls.__name__] = cls
return cls
#register
class Foo(object):
pass
#register
class Bar(object):
pass
This will work if all of your classes are defined in the same module, and if that module is imported at runtime. Your situation, however, complicates things. First, you want to define your classes in different modules. This means that we must be able to dynamically determine which modules exist within our package at runtime. This would be straightforward using Python's pkgutil module, however, you also state that you are using Nuitka to compile your package into an extension module. pkgutil doesn't work with such extension modules.
I cannot find any documented way of determining the modules contained within an Nuitka extension module from within Python. If one does exist, the decorator approach above would work after dynamically importing each submodule.
As it is, I believe the most straightforward solution is to write a script to generate an __init__.py before compiling. Suppose we have the following package structure:
.
├── __init__.py
├── plugins
│   ├── alpha.py
│   └── beta.py
└── register.py
The "plugins" are contained within the plugins directory. The contents of the files are:
# register.py
# -----------
registry = {}
def register(cls):
registry[cls.__name__] = cls
return cls
# __init__.py
# -----------
from . import plugins
from . import register
# ./plugins/alpha.py
# ------------------
from ..register import register
#register
class Alpha(object):
pass
# ./plugins/beta.py
# ------------------
from ..register import register
#register
class Beta(object):
pass
As it stands, importing the package above will not result in any of the classes being registered. This is because the class definitions are never run, since the modules containing them are never imported. The remedy is to automatically generate an __init__.py for the plugins folder. Below is a script which does exactly this -- this script can be made part of your compilation process.
import pathlib
root = pathlib.Path('./mypkg/plugins')
exclude = {'__init__.py'}
def gen_modules(root):
for entry in root.iterdir():
if entry.suffix == '.py' and entry.name not in exclude:
yield entry.stem
with (root / '__init__.py').open('w') as fh:
for module in gen_modules(root):
fh.write('from . import %s\n' % module)
Placing this script one directory above your package root (assuming your package is called mypkg) and running it yields:
from . import alpha
from . import beta
Now for the test: we compile the package:
nuitka --module mypkg --recurse-to=mypkg
and try importing it, checking to see if all of the classes were properly registered:
>>> import mypkg
>>> mypkg.register.registry
{'Beta': <class 'mypkg.plugins.beta.Beta'>,
'Alpha': <class 'mypkg.plugins.alpha.Alpha'>}
Note that the same approach will work with using metaclasses to register the plugin classes, I simply preferred to use decorators here.
If the reflected classes are using your metaclass, you don't need to use from X import * to get them registered. Only import X should be enough. As soon as the module containing the classes is imported, the classes will be created and available in your metaclass registry.
I would do this with dynamic imports.
models/regression/base.py:
class Base(object):
def get_type(self):
return "BASE"
models/regression/subclass.py:
from models.regression.base import Base
class SubClass(Base):
def get_type(self):
return "SUB_CLASS"
__myclass__ = SubClass
loader.py:
from importlib import import_module
class_name = "subclass"
module = import_module("models.regression.%s" % class_name)
model = module.__myclass__()
print(model.get_type())
And empty __init__.py files in models/ and models/regression/
With:
nuitka --recurse-none --recurse-directory models --module loader.py
The resulting loader.so contains all the modules under the models/ subdirectory.

How to document class instatiated within class using Sphinx

I'm trying to use sphinx and autodoc for a large set of python modules. How can it document a class from a module, that has been imported and insatiated in another module:
# module1.py
class Class1():
def method1():
pass
# module2.py
import module1
class Class2():
class1 = module1.Class1()
I want the class1 instance in Class2() to show up in the docs, and refer back to the module1 document.

Categories

Resources