This commit is contained in:
2026-01-14 15:56:04 -08:00
parent 0f726b11dc
commit d474916874
2 changed files with 159 additions and 126 deletions

View File

@@ -864,6 +864,7 @@ class API_folder_store:
def get_metadata(self, file_hash: str) -> Optional[Dict[str, Any]]: def get_metadata(self, file_hash: str) -> Optional[Dict[str, Any]]:
"""Get metadata for a file by hash.""" """Get metadata for a file by hash."""
try: try:
with self._db_lock:
cursor = self.connection.cursor() cursor = self.connection.cursor()
cursor.execute( cursor.execute(
@@ -1236,6 +1237,7 @@ class API_folder_store:
def get_tags(self, file_hash: str) -> List[str]: def get_tags(self, file_hash: str) -> List[str]:
"""Get all tags for a file by hash.""" """Get all tags for a file by hash."""
try: try:
with self._db_lock:
cursor = self.connection.cursor() cursor = self.connection.cursor()
cursor.execute( cursor.execute(
@@ -1833,6 +1835,7 @@ class API_folder_store:
def search_hash(self, file_hash: str) -> Optional[Path]: def search_hash(self, file_hash: str) -> Optional[Path]:
"""Search for a file by hash.""" """Search for a file by hash."""
try: try:
with self._db_lock:
cursor = self.connection.cursor() cursor = self.connection.cursor()
cursor.execute( cursor.execute(
@@ -3525,13 +3528,15 @@ class LocalLibrarySearchOptimizer:
"""Get tags from database cache.""" """Get tags from database cache."""
if not self.db: if not self.db:
return [] return []
return self.db.get_tags(file_path) file_hash = self.db.get_file_hash(file_path)
return self.db.get_tags(file_hash) if file_hash else []
def get_cached_metadata(self, file_path: Path) -> Optional[Dict[str, Any]]: def get_cached_metadata(self, file_path: Path) -> Optional[Dict[str, Any]]:
"""Get metadata from database cache.""" """Get metadata from database cache."""
if not self.db: if not self.db:
return None return None
return self.db.get_metadata(file_path) file_hash = self.db.get_file_hash(file_path)
return self.db.get_metadata(file_hash) if file_hash else None
def prefetch_metadata(self, file_paths: List[Path]) -> None: def prefetch_metadata(self, file_paths: List[Path]) -> None:
"""Pre-cache metadata for multiple files.""" """Pre-cache metadata for multiple files."""
@@ -3554,11 +3559,15 @@ class LocalLibrarySearchOptimizer:
return return
try: try:
tags = self.db.get_tags(file_path) file_hash = self.db.get_file_hash(file_path)
if not file_hash:
return
tags = self.db.get_tags(file_hash)
if tags: if tags:
search_result.tag_summary = ", ".join(tags) search_result.tag_summary = ", ".join(tags)
metadata = self.db.get_metadata(file_path) metadata = self.db.get_metadata(file_hash)
if metadata: if metadata:
if "hash" in metadata: if "hash" in metadata:
search_result.hash_hex = metadata["hash"] search_result.hash_hex = metadata["hash"]
@@ -3575,6 +3584,7 @@ class LocalLibrarySearchOptimizer:
return [] return []
try: try:
with self.db._db_lock:
cursor = self.db.connection.cursor() cursor = self.db.connection.cursor()
cursor.execute( cursor.execute(
""" """
@@ -3606,6 +3616,7 @@ class LocalLibrarySearchOptimizer:
return [] return []
try: try:
with self.db._db_lock:
cursor = self.db.connection.cursor() cursor = self.db.connection.cursor()
cursor.execute( cursor.execute(
""" """

View File

@@ -69,6 +69,17 @@ logger = logging.getLogger(__name__)
STORAGE_PATH: Optional[Path] = None STORAGE_PATH: Optional[Path] = None
API_KEY: Optional[str] = None # API key for authentication (None = no auth required) API_KEY: Optional[str] = None # API key for authentication (None = no auth required)
# Cache for database connection to prevent "database is locked" on high frequency requests
_DB_CACHE: Dict[str, Any] = {}
def get_db(path: Path):
from API.folder import LocalLibrarySearchOptimizer
p_str = str(path)
if p_str not in _DB_CACHE:
_DB_CACHE[p_str] = LocalLibrarySearchOptimizer(path)
_DB_CACHE[p_str].__enter__()
return _DB_CACHE[p_str]
# Try importing Flask - will be used in main() only # Try importing Flask - will be used in main() only
try: try:
from flask import Flask, request, jsonify from flask import Flask, request, jsonify
@@ -199,24 +210,33 @@ def create_app():
# ======================================================================== # ========================================================================
@app.route("/health", methods=["GET"]) @app.route("/health", methods=["GET"])
@require_auth()
def health(): def health():
"""Check server health and storage availability.""" """Check server health and storage availability."""
# Check auth manually to allow discovery even if locked
authed = True
if API_KEY:
provided_key = request.headers.get("X-API-Key") or request.args.get("api_key")
if not provided_key or provided_key != API_KEY:
authed = False
status = { status = {
"status": "ok", "status": "ok",
"service": "remote_storage", "service": "remote_storage",
"name": os.environ.get("MM_SERVER_NAME", "Remote Storage"), "name": os.environ.get("MM_SERVER_NAME", "Remote Storage"),
"storage_configured": STORAGE_PATH is not None, "storage_configured": STORAGE_PATH is not None,
"timestamp": datetime.now().isoformat(), "timestamp": datetime.now().isoformat(),
"locked": not authed and API_KEY is not None
} }
# If not authed but API_KEY is required, return minimal info for discovery
if not authed and API_KEY:
return jsonify(status), 200
if STORAGE_PATH: if STORAGE_PATH:
status["storage_path"] = str(STORAGE_PATH) status["storage_path"] = str(STORAGE_PATH)
status["storage_exists"] = STORAGE_PATH.exists() status["storage_exists"] = STORAGE_PATH.exists()
try: try:
from API.folder import API_folder_store search_db = get_db(STORAGE_PATH)
with API_folder_store(STORAGE_PATH) as db:
status["database_accessible"] = True status["database_accessible"] = True
except Exception as e: except Exception as e:
status["database_accessible"] = False status["database_accessible"] = False
@@ -233,8 +253,6 @@ def create_app():
@require_storage() @require_storage()
def search_files(): def search_files():
"""Search for files by name or tag.""" """Search for files by name or tag."""
from API.folder import LocalLibrarySearchOptimizer, API_folder_store
query = request.args.get("q", "") query = request.args.get("q", "")
limit = request.args.get("limit", 100, type=int) limit = request.args.get("limit", 100, type=int)
@@ -242,7 +260,7 @@ def create_app():
db_query = query if query and query != "*" else "" db_query = query if query and query != "*" else ""
try: try:
with LocalLibrarySearchOptimizer(STORAGE_PATH) as search_db: search_db = get_db(STORAGE_PATH)
results = search_db.search_by_name(db_query, limit) results = search_db.search_by_name(db_query, limit)
tag_results = search_db.search_by_tag(db_query, limit) tag_results = search_db.search_by_tag(db_query, limit)
all_results_dict = { all_results_dict = {
@@ -251,10 +269,12 @@ def create_app():
} }
# Fetch tags for each result to support title extraction on client # Fetch tags for each result to support title extraction on client
with API_folder_store(STORAGE_PATH) as db: if search_db.db:
for res in all_results_dict.values(): for res in all_results_dict.values():
if res.get("file_path"): file_hash = res.get("hash")
res["tag"] = db.get_tags(Path(res["file_path"])) if file_hash:
tags = search_db.db.get_tags(file_hash)
res["tag"] = tags
return ( return (
jsonify( jsonify(
@@ -275,17 +295,19 @@ def create_app():
@require_storage() @require_storage()
def get_file_metadata(file_hash: str): def get_file_metadata(file_hash: str):
"""Get metadata for a specific file by hash.""" """Get metadata for a specific file by hash."""
from API.folder import API_folder_store
try: try:
with API_folder_store(STORAGE_PATH) as db: search_db = get_db(STORAGE_PATH)
db = search_db.db
if not db:
return jsonify({"error": "Database unavailable"}), 500
file_path = db.search_hash(file_hash) file_path = db.search_hash(file_hash)
if not file_path or not file_path.exists(): if not file_path or not file_path.exists():
return jsonify({"error": "File not found"}), 404 return jsonify({"error": "File not found"}), 404
metadata = db.get_metadata(file_path) metadata = db.get_metadata(file_hash)
tags = db.get_tags(file_path) tags = db.get_tags(file_hash) # Use hash string
return ( return (
jsonify( jsonify(