mirror of
https://github.com/mealie-recipes/mealie.git
synced 2025-08-02 20:15:24 +02:00
* add dependency injection for get_repositories * convert events api to controller * update generic typing * add abstract controllers * update test naming * migrate admin services to controllers * add additional admin route tests * remove print * add public shared dependencies * add types * fix typo * add static variables for recipe json keys * add coverage gutters config * update controller routers * add generic success response * add category/tag/tool tests * add token refresh test * add coverage utilities * covert comments to controller * add todo * add helper properties * delete old service * update test notes * add unit test for pretty_stats * remove dead code from post_webhooks * update group routes to use controllers * add additional group test coverage * abstract common permission checks * convert ingredient parser to controller * update recipe crud to use controller * remove dead-code * add class lifespan tracker for debugging * convert bulk export to controller * migrate tools router to controller * update recipe share to controller * move customer router to _base * ignore prints in flake8 * convert units and foods to new controllers * migrate user routes to controllers * centralize error handling * fix invalid ref * reorder fields * update routers to share common handling * update tests * remove prints * fix cookbooks delete * fix cookbook get * add controller for mealplanner * cover report routes to controller * remove __future__ imports * remove dead code * remove all base_http children and remove dead code
310 lines
10 KiB
Python
310 lines
10 KiB
Python
import json
|
|
import shutil
|
|
import zipfile
|
|
from pathlib import Path
|
|
from typing import Callable
|
|
|
|
from pydantic.main import BaseModel
|
|
from sqlalchemy.orm.session import Session
|
|
|
|
from mealie.core.config import get_app_dirs
|
|
from mealie.repos.all_repositories import get_repositories
|
|
from mealie.schema.admin import CommentImport, GroupImport, RecipeImport, UserImport
|
|
from mealie.schema.recipe import Recipe, RecipeCommentOut
|
|
from mealie.schema.user import PrivateUser, UpdateGroup
|
|
from mealie.services.image import minify
|
|
|
|
app_dirs = get_app_dirs()
|
|
|
|
|
|
class ImportDatabase:
|
|
def __init__(
|
|
self,
|
|
user: PrivateUser,
|
|
session: Session,
|
|
zip_archive: str,
|
|
force_import: bool = False,
|
|
) -> None:
|
|
"""Import a database.zip file exported from mealie.
|
|
|
|
Args:
|
|
session (Session): SqlAlchemy Session
|
|
zip_archive (str): The filename contained in the backups directory
|
|
force_import (bool, optional): Force import will update all existing recipes. If False existing recipes are skipped. Defaults to False.
|
|
|
|
Raises:
|
|
Exception: If the zip file does not exists an exception raise.
|
|
"""
|
|
self.user = user
|
|
self.session = session
|
|
self.db = get_repositories(session)
|
|
self.archive = app_dirs.BACKUP_DIR.joinpath(zip_archive)
|
|
self.force_imports = force_import
|
|
|
|
if self.archive.is_file():
|
|
self.import_dir = app_dirs.TEMP_DIR.joinpath("active_import")
|
|
self.import_dir.mkdir(parents=True, exist_ok=True)
|
|
|
|
with zipfile.ZipFile(self.archive, "r") as zip_ref:
|
|
zip_ref.extractall(self.import_dir)
|
|
else:
|
|
raise Exception("Import file does not exist")
|
|
|
|
def import_recipes(self):
|
|
recipe_dir: Path = self.import_dir.joinpath("recipes")
|
|
imports = []
|
|
successful_imports = {}
|
|
|
|
recipes = ImportDatabase.read_models_file(
|
|
file_path=recipe_dir,
|
|
model=Recipe,
|
|
single_file=False,
|
|
migrate=ImportDatabase._recipe_migration,
|
|
)
|
|
|
|
for recipe in recipes:
|
|
recipe: Recipe
|
|
|
|
recipe.group_id = self.user.group_id
|
|
recipe.user_id = self.user.id
|
|
|
|
import_status = self.import_model(
|
|
db_table=self.db.recipes,
|
|
model=recipe,
|
|
return_model=RecipeImport,
|
|
name_attr="name",
|
|
search_key="slug",
|
|
slug=recipe.slug,
|
|
)
|
|
|
|
if import_status.status:
|
|
successful_imports[recipe.slug] = recipe
|
|
|
|
imports.append(import_status)
|
|
|
|
self._import_images(successful_imports)
|
|
|
|
return imports
|
|
|
|
def import_comments(self):
|
|
comment_dir: Path = self.import_dir.joinpath("comments", "comments.json")
|
|
|
|
if not comment_dir.exists():
|
|
return
|
|
|
|
comments = ImportDatabase.read_models_file(file_path=comment_dir, model=RecipeCommentOut)
|
|
|
|
for comment in comments:
|
|
comment: RecipeCommentOut
|
|
|
|
self.import_model(
|
|
db_table=self.db.comments,
|
|
model=comment,
|
|
return_model=CommentImport,
|
|
name_attr="uuid",
|
|
search_key="uuid",
|
|
)
|
|
|
|
@staticmethod
|
|
def _recipe_migration(recipe_dict: dict) -> dict:
|
|
if recipe_dict.get("categories", False):
|
|
recipe_dict["recipeCategory"] = recipe_dict.get("categories")
|
|
del recipe_dict["categories"]
|
|
try:
|
|
del recipe_dict["_id"]
|
|
del recipe_dict["date_added"]
|
|
except Exception:
|
|
pass
|
|
# Migration from list to Object Type Data
|
|
try:
|
|
if "" in recipe_dict["tags"]:
|
|
recipe_dict["tags"] = [tag for tag in recipe_dict["tags"] if tag != ""]
|
|
except Exception:
|
|
pass
|
|
|
|
try:
|
|
if "" in recipe_dict["categories"]:
|
|
recipe_dict["categories"] = [cat for cat in recipe_dict["categories"] if cat != ""]
|
|
|
|
except Exception:
|
|
pass
|
|
|
|
if type(recipe_dict["extras"]) == list:
|
|
recipe_dict["extras"] = {}
|
|
|
|
recipe_dict["comments"] = []
|
|
|
|
return recipe_dict
|
|
|
|
def _import_images(self, successful_imports: list[Recipe]):
|
|
image_dir = self.import_dir.joinpath("images")
|
|
|
|
if image_dir.exists(): # Migrate from before v0.5.0
|
|
for image in image_dir.iterdir():
|
|
item: Recipe = successful_imports.get(image.stem)
|
|
|
|
if item:
|
|
dest_dir = item.image_dir
|
|
|
|
if image.is_dir():
|
|
shutil.copytree(image, dest_dir, dirs_exist_ok=True)
|
|
|
|
if image.is_file():
|
|
shutil.copy(image, dest_dir)
|
|
|
|
else:
|
|
recipe_dir = self.import_dir.joinpath("recipes")
|
|
shutil.copytree(recipe_dir, app_dirs.RECIPE_DATA_DIR, dirs_exist_ok=True)
|
|
|
|
minify.migrate_images()
|
|
|
|
def import_settings(self):
|
|
return []
|
|
|
|
def import_groups(self):
|
|
groups_file = self.import_dir.joinpath("groups", "groups.json")
|
|
groups = ImportDatabase.read_models_file(groups_file, UpdateGroup)
|
|
group_imports = []
|
|
|
|
for group in groups:
|
|
import_status = self.import_model(self.db.groups, group, GroupImport, search_key="name")
|
|
group_imports.append(import_status)
|
|
|
|
return group_imports
|
|
|
|
def import_users(self):
|
|
users_file = self.import_dir.joinpath("users", "users.json")
|
|
users = ImportDatabase.read_models_file(users_file, PrivateUser)
|
|
user_imports = []
|
|
for user in users:
|
|
if user.id == 1: # Update Default User
|
|
self.db.users.update(1, user.dict())
|
|
import_status = UserImport(name=user.full_name, status=True)
|
|
user_imports.append(import_status)
|
|
continue
|
|
|
|
import_status = self.import_model(
|
|
db_table=self.db.users,
|
|
model=user,
|
|
return_model=UserImport,
|
|
name_attr="full_name",
|
|
search_key="email",
|
|
)
|
|
|
|
user_imports.append(import_status)
|
|
|
|
return user_imports
|
|
|
|
@staticmethod
|
|
def read_models_file(file_path: Path, model: BaseModel, single_file=True, migrate: Callable = None):
|
|
"""A general purpose function that is used to process a backup `.json` file created by mealie
|
|
note that if the file doesn't not exists the function will return any empty list
|
|
|
|
Args:
|
|
file_path (Path): The path to the .json file or directory
|
|
model (BaseModel): The pydantic model that will be created from the .json file entries
|
|
single_file (bool, optional): If true, the json data will be treated as list, if false it will use glob style matches and treat each file as its own entry. Defaults to True.
|
|
migrate (Callable, optional): A migrate function that will be called on the data prior to creating a model. Defaults to None.
|
|
|
|
Returns:
|
|
[type]: [description]
|
|
"""
|
|
if not file_path.exists():
|
|
return []
|
|
|
|
if single_file:
|
|
with open(file_path, "r") as f:
|
|
file_data = json.loads(f.read())
|
|
|
|
if migrate:
|
|
file_data = [migrate(x) for x in file_data]
|
|
|
|
return [model(**g) for g in file_data]
|
|
|
|
all_models = []
|
|
for file in file_path.glob("**/*.json"):
|
|
with open(file, "r") as f:
|
|
file_data = json.loads(f.read())
|
|
|
|
if migrate:
|
|
file_data = migrate(file_data)
|
|
|
|
all_models.append(model(**file_data))
|
|
|
|
return all_models
|
|
|
|
def import_model(self, db_table, model, return_model, name_attr="name", search_key="id", **kwargs):
|
|
"""A general purpose function used to insert a list of pydantic modelsi into the database.
|
|
The assumption at this point is that the models that are inserted. If self.force_imports is true
|
|
any existing entries will be removed prior to creation
|
|
|
|
Args:
|
|
db_table ([type]): A database table like `db.users`
|
|
model ([type]): The Pydantic model that matches the database
|
|
return_model ([type]): The return model that will be used for the 'report'
|
|
name_attr (str, optional): The name property on the return model. Defaults to "name".
|
|
search_key (str, optional): The key used to identify if an the entry already exists. Defaults to "id"
|
|
**kwargs (): Any kwargs passed will be used to set attributes on the `return_model`
|
|
|
|
Returns:
|
|
[type]: Returns the `return_model` specified.
|
|
"""
|
|
model_name = getattr(model, name_attr)
|
|
search_value = getattr(model, search_key)
|
|
|
|
item = db_table.get(search_value, search_key)
|
|
if item:
|
|
if not self.force_imports:
|
|
return return_model(
|
|
name=model_name,
|
|
status=False,
|
|
exception=f"Table entry with matching '{search_key}': '{search_value}' exists",
|
|
)
|
|
|
|
primary_key = getattr(item, db_table.primary_key)
|
|
db_table.delete(primary_key)
|
|
try:
|
|
db_table.create(model.dict())
|
|
import_status = return_model(name=model_name, status=True)
|
|
|
|
except Exception as inst:
|
|
self.session.rollback()
|
|
import_status = return_model(name=model_name, status=False, exception=str(inst))
|
|
|
|
for key, value in kwargs.items():
|
|
setattr(return_model, key, value)
|
|
|
|
return import_status
|
|
|
|
def clean_up(self):
|
|
shutil.rmtree(app_dirs.TEMP_DIR)
|
|
|
|
|
|
def import_database(
|
|
session: Session,
|
|
user: PrivateUser,
|
|
archive,
|
|
import_recipes=True,
|
|
import_settings=True,
|
|
import_users=True,
|
|
import_groups=True,
|
|
force_import: bool = False,
|
|
**_,
|
|
):
|
|
import_session = ImportDatabase(user, session, archive, force_import)
|
|
|
|
recipe_report = import_session.import_recipes() if import_recipes else []
|
|
settings_report = import_session.import_settings() if import_settings else []
|
|
group_report = import_session.import_groups() if import_groups else []
|
|
user_report = import_session.import_users() if import_users else []
|
|
notification_report = []
|
|
|
|
import_session.clean_up()
|
|
|
|
return {
|
|
"recipeImports": recipe_report,
|
|
"settingsImports": settings_report,
|
|
"groupImports": group_report,
|
|
"userImports": user_report,
|
|
"notificationImports": notification_report,
|
|
}
|