An easy to use decorator for persistent memoization: like functools.lrucache, but results persist between runs and can be stored in any format to any storage.
from perscache import Cache cache = Cache() @cache def get_data(): ...- Cache the results of a function that uses a lot of resources: runs for a long time, consumes a lot of traffic, uses up paid API calls etc.
- Speed up retreival of data that doesn't change often.
- Inspect the results of a decorated function while debugging.
- Build a simple data store.
- Easy to swap out the cache configuration when switching environments.
- Async functions supported.
- Instance-specific caching for both sync and async class methods.
- Time-to-live (TTL) support - automatically invalidate cache entries after a certain time.
- Automatic cache invalidation when the decorated function arguments or code have been changed.
- You can ignore changes in certain arguments of the decorated function.
- Each instance of a class maintains its own separate cache.
- Smart detection of instance methods vs standalone functions.
- Various serialization formats: JSON, YAML, pickle, Parquet, CSV etc.
- Various storage backends: local disk, GCS (Google Cloud Storage) and others to be implemented soon (S3, Azure Blob Storage).
- Serialization and storage are separated into two different classes, so that you can mix various serialization formats and storage back-ends as you like - JSON to local storage, Pickle to AWS, Parquet to Google Cloud Storage etc.
- You can set default serialization format and storage backend and then change them on a per-function basis.
- You can easily add new serialization formats and storage back-ends.
- Local storage is file-based, so you can use human-readable serialization (JSON, YAML, CSV) and inspect cached results.
- Automatic cleanup: least recently used results can be removed from storage when the total storage size exceeds a given threshold.
pip install perscacheThere are also some optional dependencies you need to install to use some of the features:
yaml- for YAML serializationpyarrow- for Parquet serializationpandas- for CSV serializationgcsfs- for GCS storage
from perscache import Cache cache = Cache() counter = 0 @cache def get_data(): print("Fetching data...") global counter counter += 1 return "abc" print(get_data()) # the function is called # Fetching data... # abc print(get_data()) # the cache is used # abc print(counter) # the function was called only once # 1The cache decorator works seamlessly with both regular functions and instance methods:
class Calculator: def __init__(self): self.compute_count = 0 @cache # Default per_instance=True def add(self, a: int, b: int) -> int: self.compute_count += 1 return a + b @cache(per_instance=False) # Share cache between instances def multiply(self, a: int, b: int) -> int: self.compute_count += 1 return a * b calc1 = Calculator() calc2 = Calculator() # First call computes the result result1 = calc1.add(5, 3) # compute_count = 1 # Second call uses cache result2 = calc1.add(5, 3) # compute_count still 1 # Different instance gets its own cache result3 = calc2.add(5, 3) # calc2.compute_count = 1 # Shared cache between instances result4 = calc1.multiply(4, 2) # compute_count = 2 # Second instance uses the same cache result5 = calc2.multiply(4, 2) # compute_count still 2By default, each instance maintains its own cache (per_instance=True). However, you can set per_instance=False to share the cache between all instances of a class. This works for both synchronous and asynchronous methods.
The cache works with async functions and methods:
@cache async def fetch_data(url: str): print("Fetching data...") response = await some_async_request(url) return response # First call fetches data1 = await fetch_data("example.com") # Second call uses cache data2 = await fetch_data("example.com")@cache def get_data(key): print("The function has been called...") return key print(get_data("abc")) # the function has been called # The function has been called... # abc print(get_data("fgh")) # the function has been called again # The function has been called... # fgh print(get_data("abc")) # using the cache # abc @cache def get_data(key): print("This function has been changed...") return key print(get_data("abc")) # the function has been called again # This function has been changed... # abcNOTE:
perscachehashes the function arguments, its code and the name of the class of the serializer, so that the cache is invalidated when any of these change. You can use theignore_argsparameter to ignore changes in certain arguments of the decorated function.However, if you change the code of the serializer, the cache is not invalidated. (This is because we cannot hash the code of the serializer when in was made with the factory function
make_serializer. See Make your own serialization and storage back-ends for more details.)
import datetime as dt @cache(ttl=dt.timedelta(days=1)) def get_data(): """This function will be cached for 1 day and called again after this period expires.""" ...By specifying the arguments that should be ignored, you can still use the cache even in the values of these arguments have changed.
@cache(ignore="ignore_this") def get_data(key, ignore_this): print("The function has been called...") return key print(get_data("abc", "ignore_1")) # the function has been called # The function has been called... # abc # using the cache although the the second argument is different print(get_data("abc", "ignore_2")) # abc# set up serialization format and storage backend cache = Cache( serializer=JSONSerializer(), storage=GoogleCloudStorage("/bucket/folder") ) ... # change the default serialization format @cache(serialization=PickleSerializer()) def get_data(key): ...import os from perscache import Cache, NoCache from perscache.storage import LocalFileStorage if os.environ.get["DEBUG"]: cache = NoCache() # turn off caching in debug mode else: cache = ( GoogleCloudStorage("/bucket/folder") if os.environ.get["GOOGLE_PROJECT_NAME"] # if running in the cloud else LocalFileStorage() ) @cache def function(): ...Use human-readable serialization (JSONSerializer, YAMLSerializer, CSVSerializer) and a file storage (LocalFileStorage, GoogleCloudStorage) to inspect cached results.
When using LocalFileStorage(max_size=...) or GoogleCloudStorage(max_size=...) , the least recently used cache entries are automatically removed to keep the total cache size with the max_size limit.
Although you can use the standard CloudPickleSerializer() for almost any type of data, sometimes you want to inspect the results of a decorated function by lookin into the cache files. This requires the data to be serialized in a human-readable format. But the included human-readable serializers (JSONSerializer(), YAMLSerializer(), CSVSerializer()) sometimes cannot process complex objects.
To see which serializers are compatible with which data types, see the compatibility.py file.
That's when making your own serializer comes in handy.
To do this, you should:
- Derive your own serialization classe from the abstract
Serializerclass and override the abstract methods. You should also provide theextensionclass variable that specifies the file extension. - Use your class with the
Cacheclass.
from perscache.serializers import Serializer class MySerializer(Serializer): extension = "data" def dumps(self, data: Any) -> bytes: ... def loads(self, data: bytes) -> Any: ... cache = Cache(serializer=MySerializer())You can also use the perscache.serializers.make_serializer() function to create a serializer for a given data type.
import pyrogram from perscache.serializers import make_serializer PyrogramSerializer = make_serializer( "PyrogramSerializer", "pyro", dumps_fn = lambda data: str(data).encode("utf-8"), loads_fn = lambda data: eval(data.decode("utf-8")), ) cache = Cache(serializer=PyrogramSerializer()) @cache async def some_pyrogram_func() -> pyrogram.Message: ...Making a custom storage backed is similar:
class MyStorage(Storage): def read(self, path, deadline: datetime.datetime) -> bytes: """Read the file at the given path and return its contents as bytes. If the file does not exist, raise FileNotFoundError. If the file is older than the given deadline, raise CacheExpired. """ ... def write(self, path, data: bytes) -> None: """Write the file at the given path.""" ... cache = Cache(storage=MyStorage())You can also derive your storage class from perscache.storage.FileStorage if you are building a filesystem-based storage back-end. Refer to the storage.py file for more information.