Implements Better Abstraction on Top of StorageService and File Handlers

Adds Expiration Time to Config

Reformats Project
This commit is contained in:
2023-08-10 01:54:17 -03:00
parent 0b320a3222
commit 2bd7ae10b9
12 changed files with 125 additions and 66 deletions

View File

@@ -1,3 +1,5 @@
from resize_image_service.utils.enums.storage_type import StorageType
from dotenv import load_dotenv from dotenv import load_dotenv
import os import os
@@ -5,9 +7,11 @@ import os
def get_config_s3(): def get_config_s3():
load_dotenv() load_dotenv()
return { return {
"aws_access_key_id": os.environ.get("AWS_ACCESS_KEY_ID", None), "aws_access_key_id": os.environ.get("AWS_ACCESS_KEY_ID", None),
"aws_secret_access_key": os.environ.get("AWS_SECRET_ACCESS_KEY", None), "aws_secret_access_key": os.environ.get("AWS_SECRET_ACCESS_KEY", None),
"region_name": os.environ.get("AWS_REGION_NAME", None), "region_name": os.environ.get("AWS_REGION_NAME", None),
"bucket_name": os.environ.get("AWS_BUCKET_NAME", None), "bucket_name": os.environ.get("AWS_BUCKET_NAME", None),
"expires_in": os.environ.get("EXPIRES_IN", 3600),
} }

View File

@@ -1,11 +1,11 @@
from resize_image_service.depends.depend_queue import dependency_queue from resize_image_service.depends.depend_queue import dependency_queue
from resize_image_service.depends.depend_s3_service import ( from resize_image_service.depends.depend_s3_service import (
dependency_s3_service, dependency_storage_service,
) )
from resize_image_service.service.s3_service import S3Service from resize_image_service.service.storage_service import StorageService
from resize_image_service.utils.enums.file_type import FileType from resize_image_service.utils.enums.file_type import FileType
from resize_image_service.utils.file_name_hash import file_name_hash from resize_image_service.utils.file_name_hash import file_name_hash
from resize_image_service.worker.s3_image_worker import s3_image_worker from resize_image_service.worker.storage_file_worker import storage_file_worker
from fastapi import Body, Depends, Form from fastapi import Body, Depends, Form
from fastapi_utils.cbv import cbv from fastapi_utils.cbv import cbv
@@ -20,27 +20,29 @@ s3_router = InferringRouter()
@cbv(s3_router) @cbv(s3_router)
class S3Controller: class S3Controller:
queue: Queue = Depends(dependency_queue, use_cache=True) queue: Queue = Depends(dependency_queue, use_cache=True)
s3_service: S3Service = Depends(dependency_s3_service, use_cache=True) storage_service: StorageService = Depends(dependency_storage_service, use_cache=True)
@s3_router.get("/new_file_url/", status_code=200) @s3_router.get("/new_file_url/", status_code=200)
def new_file_url( def new_file_url(
self, self,
username: Annotated[str, Form()], username: Annotated[str, Body(embed=True)],
file_postfix: Annotated[str, Form()], file_postfix: Annotated[str, Body(embed=True)],
file_type: Annotated[FileType, Form()], file_type: Annotated[FileType, Body(embed=True)],
) -> dict[str, str]: ) -> dict[str, str]:
return self.s3_service.get_temp_upload_link( return self.storage_service.get_temp_upload_link(
file_name_hash(username, file_postfix), file_type file_name_hash(username, file_postfix), file_type
) )
@s3_router.get("/file_url/", status_code=200) @s3_router.get("/file_url/", status_code=200)
def file_url( def file_url(
self, username: Annotated[str, Form()], file_postfix: Annotated[str, Form()] self,
username: Annotated[str, Body(embed=True)],
file_postfix: Annotated[str, Body(embed=True)],
) -> dict[str, str]: ) -> dict[str, str]:
return self.s3_service.get_temp_read_link( return self.storage_service.get_temp_read_link(
file_name_hash(username, file_postfix) file_name_hash(username, file_postfix)
) )
@s3_router.post("/process_image/", status_code=200) @s3_router.post("/process_file/", status_code=200)
def process_image(self, string_url: Annotated[str, Body(embed=True)]): def process_file(self, string_url: Annotated[str, Body(embed=True)]):
self.queue.enqueue(s3_image_worker, string_url) self.queue.enqueue(storage_file_worker, string_url)

View File

@@ -1,9 +1,19 @@
from resize_image_service.config.config_s3 import get_config_s3 from resize_image_service.config.config_s3 import get_config_s3
from resize_image_service.service.s3_service import S3Service from resize_image_service.service.amazon_s3_service import AmazonS3Service
from resize_image_service.service.storage_service import StorageService
from resize_image_service.utils.enums.storage_type import StorageType
from dotenv import load_dotenv
import os
from functools import cache from functools import cache
@cache @cache
def dependency_s3_service() -> S3Service: def dependency_storage_service() -> StorageService:
return S3Service(**get_config_s3()) load_dotenv()
if StorageType(os.environ["STORAGE_TYPE"]) == StorageType.S3_STORAGE:
return AmazonS3Service(**get_config_s3())
raise RuntimeError("Invalid Storage Type")

View File

@@ -1,18 +1,27 @@
from resize_image_service.utils.enums.file_type import CONTENT_TYPE, FileType from __future__ import annotations
from resize_image_service.service.storage_service import StorageService
from resize_image_service.utils.enums.file_type import FileType
from resize_image_service.utils.file_handler import FILE_HANDLER
import boto3 import boto3
from PIL import Image from PIL import Image
import io import io
from typing import Any, Dict from typing import Any
class S3Service: class AmazonS3Service(StorageService):
def __init__(self, **kwargs): def __init__(self, **kwargs):
super().__init__(**kwargs)
self.__validate_config(**kwargs) self.__validate_config(**kwargs)
self.bucket_name = kwargs.get("bucket_name") self.bucket_name = kwargs.get("bucket_name")
self.region_name = kwargs.get("region_name") self.region_name = kwargs.get("region_name")
self.expires_in = kwargs.get("expires_in")
self.s3 = boto3.client( self.s3 = boto3.client(
"s3", "s3",
aws_access_key_id=kwargs.get("aws_access_key_id"), aws_access_key_id=kwargs.get("aws_access_key_id"),
@@ -24,69 +33,47 @@ class S3Service:
self, file_name, file_type: FileType self, file_name, file_type: FileType
) -> dict[str, str | Any]: ) -> dict[str, str | Any]:
return { return {
"presigned_url": self._get_presigned_right_url(file_name, file_type), "presigned_url": self._get_presigned_write_url(file_name, file_type),
"file_key": self._get_object_url(file_name), "file_key": self._get_object_url(file_name),
} }
def get_temp_read_link(self, file_name) -> dict[str, str | Any]: def get_temp_read_link(self, file_name) -> dict[str, str | Any]:
return {"presigned_url": self._get_presigned_read_url(file_name)} return {"presigned_url": self._get_presigned_read_url(file_name)}
def process_image(self, file_name) -> None: def process_file(self, file_name: str, file_type: FileType) -> None:
img = self._get_image_obj(file_name) file_bytes = self._get_file_obj(file_name)
handler = FILE_HANDLER[file_type]["handler"]
img = self._resize_img(img) self._upload_file(file_name, handler(file_bytes))
img = self._remove_img_metadata(img)
self._upload_image(file_name, img) def _get_object_url(self, file_name: str) -> str:
def _get_object_url(self, file_name: str):
return f"https://{self.bucket_name}.s3.{self.region_name}.amazonaws.com/{file_name}" return f"https://{self.bucket_name}.s3.{self.region_name}.amazonaws.com/{file_name}"
def _get_presigned_right_url(self, file_name, file_type: FileType): def _get_presigned_write_url(self, file_name, file_type: FileType) -> str:
return self.s3.generate_presigned_url( return self.s3.generate_presigned_url(
"put_object", "put_object",
Params={ Params={
"Bucket": self.bucket_name, "Bucket": self.bucket_name,
"Key": file_name, "Key": file_name,
"ContentType": CONTENT_TYPE[file_type], "ContentType": FILE_HANDLER[file_type]["content_type"],
}, },
ExpiresIn=3600, ExpiresIn=self.expires_in,
) )
def _get_presigned_read_url(self, file_name): def _get_presigned_read_url(self, file_name) -> str:
return self.s3.generate_presigned_url( return self.s3.generate_presigned_url(
"get_object", "get_object",
Params={"Bucket": self.bucket_name, "Key": file_name}, Params={"Bucket": self.bucket_name, "Key": file_name},
ExpiresIn=3600, ExpiresIn=self.expires_in,
) )
def _get_image_obj(self, file_name: str): def _get_file_obj(self, file_name: str) -> io.BytesIO:
object_byte = io.BytesIO( return io.BytesIO(
self.s3.get_object(Bucket=self.bucket_name, Key=file_name)["Body"].read() self.s3.get_object(Bucket=self.bucket_name, Key=file_name)["Body"].read()
) )
return Image.open(object_byte) def _upload_file(self, file_name: str, file_bytes: io.BytesIO) -> None:
self.s3.upload_fileobj(file_bytes, Bucket=self.bucket_name, Key=file_name)
def _upload_image(self, file_name: str, img: Image):
new_byte_img = io.BytesIO()
img.save(new_byte_img, format="PNG")
new_byte_img.seek(0)
self.s3.upload_fileobj(new_byte_img, Bucket=self.bucket_name, Key=file_name)
@staticmethod
def _resize_img(img):
img.thumbnail((320, 320))
return img
@staticmethod
def _remove_img_metadata(img):
data = list(img.getdata())
image_without_exif = Image.new(img.mode, img.size)
image_without_exif.putdata(data)
return image_without_exif
@staticmethod @staticmethod
def __validate_config(**kwargs): def __validate_config(**kwargs):

View File

@@ -0,0 +1,25 @@
from __future__ import annotations
from resize_image_service.utils.enums.file_type import FileType
from abc import ABC, abstractmethod
from typing import Any
class StorageService(ABC):
def __init__(self, **kwargs):
pass
@abstractmethod
def get_temp_upload_link(
self, file_name, file_type: FileType
) -> dict[str, str | Any]:
pass
@abstractmethod
def get_temp_read_link(self, file_name) -> dict[str, str | Any]:
pass
@abstractmethod
def process_file(self, file_name) -> None:
pass

View File

@@ -4,6 +4,3 @@ from enum import Enum
class FileType(Enum): class FileType(Enum):
PNG = "png" PNG = "png"
JPEG = "jpeg" JPEG = "jpeg"
CONTENT_TYPE = {FileType.PNG: "image/png", FileType.JPEG: "image/jpeg"}

View File

@@ -0,0 +1,5 @@
from enum import Enum
class StorageType(Enum):
S3_STORAGE = "s3"

View File

@@ -0,0 +1,9 @@
from resize_image_service.utils.enums.file_type import FileType
from resize_image_service.utils.file_handler.handlers.image_handler import (
image_handler,
)
FILE_HANDLER = {
FileType.PNG: {"content_type": "image/png", "handler": image_handler},
FileType.JPEG: {"content_type": "image/jpeg", "handler": image_handler},
}

View File

@@ -0,0 +1,20 @@
from PIL import Image
import io
def image_handler(file_bytes: io.BytesIO) -> io.BytesIO:
img = Image.open(file_bytes)
img.thumbnail((320, 320))
data = list(img.getdata())
image_without_exif = Image.new(img.mode, img.size)
image_without_exif.putdata(data)
new_byte_img = io.BytesIO()
img.save(new_byte_img, format="PNG")
new_byte_img.seek(0)
return new_byte_img

View File

@@ -1,7 +0,0 @@
from resize_image_service.depends.depend_s3_service import (
dependency_s3_service,
)
def s3_image_worker(string_url: str) -> None:
dependency_s3_service().process_image(string_url)

View File

@@ -0,0 +1,7 @@
from resize_image_service.depends.depend_s3_service import (
dependency_storage_service,
)
def storage_file_worker(string_url: str) -> None:
dependency_storage_service().process_image(string_url)