media-rip/backend/app/routers/admin.py
xpltd 1e9014f569 Error log: failed download diagnostics for admin
Backend:
- New error_log table: url, domain, error, format_id, media_type,
  session_id, created_at
- log_download_error() called when yt-dlp throws during download
- GET /admin/errors returns recent entries (limit 200)
- DELETE /admin/errors clears all entries
- Manual purge also clears error log
- Domain extracted from URL via urlparse for grouping

Frontend:
- New 'Errors' tab in admin panel (Sessions, Storage, Errors, Settings)
- Each error entry shows: domain, timestamp, full URL, error message,
  format/media type metadata
- Red left border + error-colored message for visual scanning
- Clear Log button to wipe entries
- Empty state: 'No errors logged.'

Error entries contain enough context (full URL, error message, domain,
format, media type) to paste into an LLM for domain-specific debugging.
2026-03-19 06:34:08 -05:00

343 lines
11 KiB
Python

"""Admin API endpoints — protected by require_admin dependency."""
from __future__ import annotations
import logging
from fastapi import APIRouter, Depends, Request
from app.dependencies import require_admin
logger = logging.getLogger("mediarip.admin")
router = APIRouter(prefix="/admin", tags=["admin"])
@router.get("/sessions")
async def list_sessions(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""List all sessions with basic stats."""
db = request.app.state.db
cursor = await db.execute(
"""
SELECT s.id, s.created_at, s.last_seen,
COUNT(j.id) as job_count
FROM sessions s
LEFT JOIN jobs j ON j.session_id = s.id
GROUP BY s.id
ORDER BY s.last_seen DESC
"""
)
rows = await cursor.fetchall()
sessions = [
{
"id": row["id"],
"created_at": row["created_at"],
"last_seen": row["last_seen"],
"job_count": row["job_count"],
}
for row in rows
]
return {"sessions": sessions, "total": len(sessions)}
@router.get("/sessions/{session_id}/jobs")
async def session_jobs(
session_id: str,
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""List jobs for a specific session with file details."""
db = request.app.state.db
cursor = await db.execute(
"""
SELECT id, url, status, filename, filesize,
created_at, started_at, completed_at
FROM jobs
WHERE session_id = ?
ORDER BY created_at DESC
""",
(session_id,),
)
rows = await cursor.fetchall()
jobs = [
{
"id": row["id"],
"url": row["url"],
"status": row["status"],
"filename": row["filename"],
"filesize": row["filesize"],
"created_at": row["created_at"],
"started_at": row["started_at"],
"completed_at": row["completed_at"],
}
for row in rows
]
return {"jobs": jobs}
@router.get("/storage")
async def storage_info(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Return storage usage information."""
import shutil
from pathlib import Path
config = request.app.state.config
db = request.app.state.db
output_dir = Path(config.downloads.output_dir)
# Disk usage
try:
usage = shutil.disk_usage(output_dir)
disk = {
"total": usage.total,
"used": usage.used,
"free": usage.free,
}
except OSError:
disk = {"total": 0, "used": 0, "free": 0}
# Job counts by status
cursor = await db.execute(
"SELECT status, COUNT(*) as count FROM jobs GROUP BY status"
)
rows = await cursor.fetchall()
by_status = {row["status"]: row["count"] for row in rows}
return {"disk": disk, "jobs_by_status": by_status}
@router.get("/unsupported-urls")
async def list_unsupported_urls(
request: Request,
_admin: str = Depends(require_admin),
limit: int = 100,
offset: int = 0,
) -> dict:
"""List logged unsupported URL extraction failures."""
db = request.app.state.db
cursor = await db.execute(
"SELECT * FROM unsupported_urls ORDER BY created_at DESC LIMIT ? OFFSET ?",
(limit, offset),
)
rows = await cursor.fetchall()
items = [
{
"id": row["id"],
"url": row["url"],
"session_id": row["session_id"],
"error": row["error"],
"created_at": row["created_at"],
}
for row in rows
]
# Total count
count_cursor = await db.execute("SELECT COUNT(*) FROM unsupported_urls")
count_row = await count_cursor.fetchone()
total = count_row[0] if count_row else 0
return {"items": items, "total": total, "limit": limit, "offset": offset}
@router.get("/errors")
async def get_errors(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Return recent download error log entries."""
from app.core.database import get_error_log
db = request.app.state.db
entries = await get_error_log(db, limit=200)
return {"errors": entries}
@router.delete("/errors")
async def clear_errors(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Clear all error log entries."""
from app.core.database import clear_error_log
db = request.app.state.db
count = await clear_error_log(db)
return {"cleared": count}
@router.post("/purge")
async def manual_purge(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Manually trigger a purge of expired downloads."""
from app.services.purge import run_purge
config = request.app.state.config
db = request.app.state.db
# Attach runtime overrides so purge service can read them
overrides = getattr(request.app.state, "settings_overrides", {})
config._runtime_overrides = overrides
result = await run_purge(db, config, purge_all=True)
# Broadcast job_removed events to all SSE clients
broker = request.app.state.broker
for job_id in result.get("deleted_job_ids", []):
broker.publish_all({"event": "job_removed", "data": {"job_id": job_id}})
# Don't send internal field to client
result.pop("deleted_job_ids", None)
return result
@router.put("/settings")
async def update_settings(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Update runtime settings (in-memory only — resets on restart).
Accepts a JSON body with optional fields:
- welcome_message: str
- default_video_format: str (auto, mp4, webm)
- default_audio_format: str (auto, mp3, m4a, flac, wav, opus)
"""
body = await request.json()
if not hasattr(request.app.state, "settings_overrides"):
request.app.state.settings_overrides = {}
updated = []
if "welcome_message" in body:
msg = body["welcome_message"]
if not isinstance(msg, str):
from fastapi.responses import JSONResponse
return JSONResponse(
status_code=422,
content={"detail": "welcome_message must be a string"},
)
request.app.state.settings_overrides["welcome_message"] = msg
updated.append("welcome_message")
logger.info("Admin updated welcome_message to: %s", msg[:80])
valid_video_formats = {"auto", "mp4", "webm"}
valid_audio_formats = {"auto", "mp3", "m4a", "flac", "wav", "opus"}
if "default_video_format" in body:
fmt = body["default_video_format"]
if fmt in valid_video_formats:
request.app.state.settings_overrides["default_video_format"] = fmt
updated.append("default_video_format")
logger.info("Admin updated default_video_format to: %s", fmt)
if "default_audio_format" in body:
fmt = body["default_audio_format"]
if fmt in valid_audio_formats:
request.app.state.settings_overrides["default_audio_format"] = fmt
updated.append("default_audio_format")
logger.info("Admin updated default_audio_format to: %s", fmt)
if "privacy_mode" in body:
val = body["privacy_mode"]
if isinstance(val, bool):
request.app.state.settings_overrides["privacy_mode"] = val
# When enabling privacy mode, also enable the purge scheduler
config = request.app.state.config
if val and not config.purge.enabled:
config.purge.enabled = True
# Start the scheduler if APScheduler is available
try:
from apscheduler.schedulers.asyncio import AsyncIOScheduler
from apscheduler.triggers.cron import CronTrigger
from app.services.purge import run_purge
if not hasattr(request.app.state, "scheduler"):
scheduler = AsyncIOScheduler()
scheduler.add_job(
run_purge,
CronTrigger(minute="*/30"), # every 30 min for privacy
args=[request.app.state.db, config],
id="purge_job",
name="Privacy purge",
replace_existing=True,
)
scheduler.start()
request.app.state.scheduler = scheduler
logger.info("Privacy mode: started purge scheduler (every 30 min)")
except Exception as e:
logger.warning("Could not start purge scheduler: %s", e)
updated.append("privacy_mode")
logger.info("Admin updated privacy_mode to: %s", val)
if "privacy_retention_hours" in body:
val = body["privacy_retention_hours"]
if isinstance(val, (int, float)) and 1 <= val <= 8760: # 1 hour to 1 year
request.app.state.settings_overrides["privacy_retention_hours"] = int(val)
updated.append("privacy_retention_hours")
logger.info("Admin updated privacy_retention_hours to: %d", int(val))
logger.info("Admin updated default_audio_format to: %s", fmt)
return {"updated": updated, "status": "ok"}
@router.put("/password")
async def change_password(
request: Request,
_admin: str = Depends(require_admin),
) -> dict:
"""Change admin password (in-memory only — resets on restart).
Accepts JSON body:
- current_password: str (required, must match current password)
- new_password: str (required, min 4 chars)
"""
import bcrypt
body = await request.json()
current = body.get("current_password", "")
new_pw = body.get("new_password", "")
if not current or not new_pw:
from fastapi.responses import JSONResponse
return JSONResponse(
status_code=422,
content={"detail": "current_password and new_password are required"},
)
if len(new_pw) < 4:
from fastapi.responses import JSONResponse
return JSONResponse(
status_code=422,
content={"detail": "New password must be at least 4 characters"},
)
# Verify current password
config = request.app.state.config
try:
valid = bcrypt.checkpw(
current.encode("utf-8"),
config.admin.password_hash.encode("utf-8"),
)
except (ValueError, TypeError):
valid = False
if not valid:
from fastapi.responses import JSONResponse
return JSONResponse(
status_code=403,
content={"detail": "Current password is incorrect"},
)
# Hash and store new password
new_hash = bcrypt.hashpw(new_pw.encode("utf-8"), bcrypt.gensalt()).decode("utf-8")
config.admin.password_hash = new_hash
logger.info("Admin password changed by user '%s'", _admin)
return {"status": "ok", "message": "Password changed successfully"}