import time import pendulum import redis.asyncio as redis import uvicorn from fastapi import FastAPI from redis.asyncio.connection import ConnectionPool from starlette.requests import Request from starlette.responses import Response from fastapi_cache import FastAPICache from fastapi_cache.backends.redis import RedisBackend from fastapi_cache.decorator import cache app = FastAPI() ret = 0 @cache(namespace="test", expire=1) async def get_ret(): global ret ret = ret + 1 return ret @app.get("/") @cache(namespace="test", expire=10) async def index(): return dict(ret=await get_ret()) @app.get("/clear") async def clear(): return await FastAPICache.clear(namespace="test") @app.get("/date") @cache(namespace="test", expire=10) async def get_data(request: Request, response: Response): return pendulum.today() # Note: This function MUST be sync to demonstrate fastapi-cache's correct handling, # i.e. running cached sync functions in threadpool just like FastAPI itself! @app.get("/blocking") @cache(namespace="test", expire=10) def blocking(): time.sleep(5) return dict(ret=get_ret()) @app.get("/datetime") @cache(namespace="test", expire=2) async def get_datetime(request: Request, response: Response): print(request, response) return pendulum.now() @app.on_event("startup") async def startup(): pool = ConnectionPool.from_url(url="redis://redis") r = redis.Redis(connection_pool=pool) FastAPICache.init(RedisBackend(r), prefix="fastapi-cache") if __name__ == "__main__": uvicorn.run("main:app", debug=True, reload=True)