V0.2.0
All checks were successful
Build Docker Image / build (push) Successful in 12m39s

This commit is contained in:
sudacode 2025-02-13 17:46:15 -08:00
parent 9544b0415c
commit b7e89d9c22
Signed by: sudacode
SSH Key Fingerprint: SHA256:lT5C2bB398DcX6daCF/gYFNSTK3y+Du3oTGUnYzfTEw
17 changed files with 455 additions and 24 deletions

2
.gitignore vendored
View File

@ -3,3 +3,5 @@ env/*
.git
db/*
app/__pycache__/*
migrations/versions/__pycache__/
migrations/__pycache__/

View File

@ -17,10 +17,6 @@ WORKDIR /app
# Copy the current directory contents into the container at /app
COPY . /app
# Install any needed packages specified in requirements.txt
# If there are no external dependencies, you can skip this step
# RUN pip install --no-cache-dir -r requirements.txt
# Make port 8080 available to the world outside this container
EXPOSE "${PORT_NUMBER}"

View File

@ -1 +1 @@
0.1.2
0.2.0

119
alembic.ini Normal file
View File

@ -0,0 +1,119 @@
# A generic, single database configuration.
[alembic]
# path to migration scripts
# Use forward slashes (/) also on windows to provide an os agnostic path
script_location = migrations
# template used to generate migration file names; The default value is %%(rev)s_%%(slug)s
# Uncomment the line below if you want the files to be prepended with date and time
# see https://alembic.sqlalchemy.org/en/latest/tutorial.html#editing-the-ini-file
# for all available tokens
# file_template = %%(year)d_%%(month).2d_%%(day).2d_%%(hour).2d%%(minute).2d-%%(rev)s_%%(slug)s
# sys.path path, will be prepended to sys.path if present.
# defaults to the current working directory.
prepend_sys_path = .
# timezone to use when rendering the date within the migration file
# as well as the filename.
# If specified, requires the python>=3.9 or backports.zoneinfo library and tzdata library.
# Any required deps can installed by adding `alembic[tz]` to the pip requirements
# string value is passed to ZoneInfo()
# leave blank for localtime
# timezone =
# max length of characters to apply to the "slug" field
# truncate_slug_length = 40
# set to 'true' to run the environment during
# the 'revision' command, regardless of autogenerate
# revision_environment = false
# set to 'true' to allow .pyc and .pyo files without
# a source .py file to be detected as revisions in the
# versions/ directory
# sourceless = false
# version location specification; This defaults
# to migrations/versions. When using multiple version
# directories, initial revisions must be specified with --version-path.
# The path separator used here should be the separator specified by "version_path_separator" below.
# version_locations = %(here)s/bar:%(here)s/bat:migrations/versions
# version path separator; As mentioned above, this is the character used to split
# version_locations. The default within new alembic.ini files is "os", which uses os.pathsep.
# If this key is omitted entirely, it falls back to the legacy behavior of splitting on spaces and/or commas.
# Valid values for version_path_separator are:
#
# version_path_separator = :
# version_path_separator = ;
# version_path_separator = space
# version_path_separator = newline
#
# Use os.pathsep. Default configuration used for new projects.
version_path_separator = os
# set to 'true' to search source files recursively
# in each "version_locations" directory
# new in Alembic version 1.10
# recursive_version_locations = false
# the output encoding used when revision files
# are written from script.py.mako
# output_encoding = utf-8
# sqlalchemy.url = driver://user:pass@localhost/dbname
[post_write_hooks]
# post_write_hooks defines scripts or Python functions that are run
# on newly generated revision scripts. See the documentation for further
# detail and examples
# format using "black" - use the console_scripts runner, against the "black" entrypoint
# hooks = black
# black.type = console_scripts
# black.entrypoint = black
# black.options = -l 79 REVISION_SCRIPT_FILENAME
# lint with attempts to fix using "ruff" - use the exec runner, execute a binary
# hooks = ruff
# ruff.type = exec
# ruff.executable = %(here)s/.venv/bin/ruff
# ruff.options = --fix REVISION_SCRIPT_FILENAME
# Logging configuration
[loggers]
keys = root,sqlalchemy,alembic
[handlers]
keys = console
[formatters]
keys = generic
[logger_root]
level = WARNING
handlers = console
qualname =
[logger_sqlalchemy]
level = WARNING
handlers =
qualname = sqlalchemy.engine
[logger_alembic]
level = INFO
handlers =
qualname = alembic
[handler_console]
class = StreamHandler
args = (sys.stderr,)
level = NOTSET
formatter = generic
[formatter_generic]
format = %(levelname)-5.5s [%(name)s] %(message)s
datefmt = %H:%M:%S

View File

@ -4,16 +4,21 @@ from app.config import Config
from app.database import engine
from app.models import Base
from app.views import bp
from app.migrations import run_migrations
def create_app():
app = Flask(__name__)
def create_app(test_config=None):
app = Flask(__name__, instance_relative_config=True)
app.config.from_object(Config)
# Ensure database tables exist
# Create database tables if they don't exist
Base.metadata.create_all(engine)
# Register blueprints
app.register_blueprint(bp)
# Run migrations after tables are created
with app.app_context():
run_migrations(app)
return app

View File

@ -4,6 +4,6 @@ import os
class Config:
LISTEN_ADDRESS = os.getenv("LISTEN_ADDRESS", "0.0.0.0")
LISTEN_PORT = int(os.getenv("LISTEN_PORT", "8080"))
DATABASE_URL = os.getenv("DATABASE_URL", "sqlite:///./mpv.db")
SQLALCHEMY_DATABASE_URI = os.getenv("SQLALCHEMY_DATABASE_URI", "sqlite:///./mpv.db")
MPV_SOCKET = os.getenv("MPV_SOCKET", "/tmp/mpvsocket")
LOGLEVEL = os.getenv("LOGLEVEL", "INFO").strip().upper()

View File

@ -2,7 +2,7 @@ from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from app.config import Config
engine = create_engine(Config.DATABASE_URL, pool_pre_ping=True)
engine = create_engine(Config.SQLALCHEMY_DATABASE_URI, pool_pre_ping=True)
Session = scoped_session(sessionmaker(bind=engine))

32
app/migrations.py Normal file
View File

@ -0,0 +1,32 @@
import logging
import os
from alembic.config import Config
from alembic import command
from flask import Blueprint
from app.config import Config as AppConfig
migrations_bp = Blueprint("migrations", __name__)
def run_migrations(app):
"""Run database migrations"""
try:
# Create Alembic configuration object
alembic_cfg = Config("alembic.ini")
# Get the database URL from your app's config
database_url = AppConfig.SQLALCHEMY_DATABASE_URI
print(database_url)
# Set the SQLAlchemy URL
alembic_cfg.set_main_option("sqlalchemy.url", database_url)
logging.debug(f"Database URL set to: {database_url}")
# Run the migration
command.upgrade(alembic_cfg, "head")
logging.info("Database migration completed successfully")
return True
except Exception as e:
logging.error(f"Migration failed: {str(e)}")
return False

View File

@ -12,6 +12,11 @@ class WatchHistory(Base):
video_name = Column(String(255), nullable=False)
channel_url = Column(String(255), nullable=False)
channel_name = Column(String(255), nullable=False)
category = Column(String(100), nullable=True)
view_count = Column(Integer, nullable=True)
subscriber_count = Column(Integer, nullable=True)
thumbnail_url = Column(String(255), nullable=True)
upload_date = Column(DateTime, nullable=True)
watch_date = Column(DateTime, nullable=False, server_default=func.now())
created_by = Column(
String(100), nullable=False, server_default="mpv-youtube-queue-server"

View File

@ -2,6 +2,7 @@ import logging
from flask import Blueprint, g, jsonify, request
from sqlalchemy.exc import SQLAlchemyError
from sqlalchemy import inspect
from app.database import get_db_session
from app.models import SavedQueue, WatchHistory
@ -61,6 +62,10 @@ def add_video():
if not all(
k in data for k in ["video_url", "video_name", "channel_url", "channel_name"]
):
logging.error("Missing required fields")
logging.error(
"Required fields: video_url, video_name, channel_url, channel_name"
)
return jsonify(message="Missing required fields"), 400
new_entry = WatchHistory(
@ -68,12 +73,18 @@ def add_video():
video_name=data["video_name"],
channel_url=data["channel_url"],
channel_name=data["channel_name"],
category=data.get("category") if data.get("category") else None,
view_count=data.get("view_count") if data.get("view_count") else None,
subscriber_count=data.get("subscribers") if data.get("subscribers") else None,
thumbnail_url=data.get("thumbnail_url") if data.get("thumbnail_url") else None,
upload_date=data.get("upload_date") if data.get("upload_date") else None,
)
db_session = g.db_session
db_session.add(new_entry)
try:
logging.debug("Adding video to watch history")
db_session.add(new_entry)
db_session.commit()
logging.debug("Video added to watch history")
logging.debug(f"URL: {data['video_url']}")
@ -99,3 +110,78 @@ def handle_request():
if send_to_mpv(command):
return "URL added to mpv queue", 200
return "Failed to add URL to mpv queue", 500
@bp.route("/migrate_watch_history", methods=["POST"])
def migrate_watch_history():
db_session = g.db_session
engine = db_session.get_bind()
try:
# First check and add missing columns
inspector = inspect(engine)
existing_columns = [
col["name"] for col in inspector.get_columns("watch_history")
]
# Define new columns and their SQL
new_columns = {
"category": "ALTER TABLE watch_history ADD COLUMN category VARCHAR(100)",
"view_count": "ALTER TABLE watch_history ADD COLUMN view_count INTEGER",
"subscriber_count": "ALTER TABLE watch_history ADD COLUMN subscriber_count INTEGER",
"thumbnail_url": "ALTER TABLE watch_history ADD COLUMN thumbnail_url VARCHAR(255)",
"upload_date": "ALTER TABLE watch_history ADD COLUMN upload_date TIMESTAMP",
}
# Add missing columns
columns_added = []
for col_name, sql in new_columns.items():
if col_name not in existing_columns:
engine.execute(sql)
columns_added.append(col_name)
logging.info(f"Added column: {col_name}")
# Now backfill with default values
entries = db_session.query(WatchHistory).all()
updated_count = 0
for entry in entries:
updated = False
# Check and set defaults for new columns if they're None
if entry.category is None:
entry.category = "Unknown"
updated = True
if entry.view_count is None:
entry.view_count = 0
updated = True
if entry.subscriber_count is None:
entry.subscriber_count = 0
updated = True
if entry.thumbnail_url is None:
entry.thumbnail_url = ""
updated = True
if entry.upload_date is None:
# Set to watch_date as a fallback
entry.upload_date = entry.watch_date
updated = True
if updated:
updated_count += 1
db_session.commit()
return (
jsonify(
{
"message": "Migration completed successfully",
"columns_added": columns_added,
"records_updated": updated_count,
"total_records": len(entries),
}
),
200,
)
except SQLAlchemyError as e:
db_session.rollback()
logging.error(f"Migration failed: {e}")
return jsonify(message=f"Failed to migrate watch history: {str(e)}"), 500

View File

@ -9,9 +9,4 @@ services:
ports:
- 42069:8080
env_file: .env
networks:
- mpv-youtube-queue-server
restart: unless-stopped
networks:
mpv-youtube-queue-server:
external: true

View File

@ -8,4 +8,4 @@ LOGLEVEL=info
# postgresql+psycopg2://user:password@host:port/dbname[?key=value&key=value...]
# sqlite:///path
# oracle+oracledb://user:pass@hostname:port[/dbname][?service_name=<service>[&key=value&key=value...]]
DATABASE_URL=mysql+pymysql://user:password@localhost:3306/mpv
SQLALCHEMY_DATABASE_URI=mysql+pymysql://user:password@localhost:3306/mpv

83
migrations/env.py Normal file
View File

@ -0,0 +1,83 @@
from logging.config import fileConfig
from sqlalchemy import engine_from_config
from sqlalchemy import pool, create_engine
from alembic import context
from app.models import Base
from app.config import Config
# this is the Alembic Config object, which provides
# access to the values within the .ini file in use.
config = context.config
# Set the database URL from our app config
config.set_main_option('sqlalchemy.url', Config.SQLALCHEMY_DATABASE_URI)
# Interpret the config file for Python logging.
# This line sets up loggers basically.
if config.config_file_name is not None:
fileConfig(config.config_file_name)
# add your model's MetaData object here
# for 'autogenerate' support
# from myapp import mymodel
# target_metadata = mymodel.Base.metadata
target_metadata = Base.metadata
# other values from the config, defined by the needs of env.py,
# can be acquired:
# my_important_option = config.get_main_option("my_important_option")
# ... etc.
def run_migrations_offline() -> None:
"""Run migrations in 'offline' mode.
This configures the context with just a URL
and not an Engine, though an Engine is acceptable
here as well. By skipping the Engine creation
we don't even need a DBAPI to be available.
Calls to context.execute() here emit the given string to the
script output.
"""
url = config.get_main_option("sqlalchemy.url")
context.configure(
url=url,
target_metadata=target_metadata,
literal_binds=True,
dialect_opts={"paramstyle": "named"},
)
with context.begin_transaction():
context.run_migrations()
def run_migrations_online() -> None:
"""Run migrations in 'online' mode.
In this scenario we need to create an Engine
and associate a connection with the context.
"""
# Use create_engine() directly with our URL
connectable = create_engine(Config.SQLALCHEMY_DATABASE_URI)
with connectable.connect() as connection:
context.configure(
connection=connection,
target_metadata=target_metadata
)
with context.begin_transaction():
context.run_migrations()
if context.is_offline_mode():
run_migrations_offline()
else:
run_migrations_online()

26
migrations/script.py.mako Normal file
View File

@ -0,0 +1,26 @@
"""${message}
Revision ID: ${up_revision}
Revises: ${down_revision | comma,n}
Create Date: ${create_date}
"""
from typing import Sequence, Union
from alembic import op
import sqlalchemy as sa
${imports if imports else ""}
# revision identifiers, used by Alembic.
revision: str = ${repr(up_revision)}
down_revision: Union[str, None] = ${repr(down_revision)}
branch_labels: Union[str, Sequence[str], None] = ${repr(branch_labels)}
depends_on: Union[str, Sequence[str], None] = ${repr(depends_on)}
def upgrade() -> None:
${upgrades if upgrades else "pass"}
def downgrade() -> None:
${downgrades if downgrades else "pass"}

View File

@ -0,0 +1,80 @@
"""add watch history columns
Revision ID: 8911624d0776
Revises:
Create Date: 2024-xx-xx xx:xx:xx.xxx
"""
from alembic import op
import sqlalchemy as sa
from datetime import datetime
from sqlalchemy.engine.reflection import Inspector
from sqlalchemy import text
# revision identifiers, used by Alembic.
revision = '8911624d0776'
down_revision = None
branch_labels = None
depends_on = None
def column_exists(table, column):
# Get inspector
conn = op.get_bind()
inspector = Inspector.from_engine(conn)
columns = [c['name'] for c in inspector.get_columns(table)]
return column in columns
def upgrade():
# Add new columns if they don't exist
with op.batch_alter_table('watch_history') as batch_op:
# Add category column
if not column_exists('watch_history', 'category'):
batch_op.add_column(sa.Column('category', sa.String(100), nullable=True))
# Add view_count column
if not column_exists('watch_history', 'view_count'):
batch_op.add_column(sa.Column('view_count', sa.Integer(), nullable=True))
# Add subscriber_count column
if not column_exists('watch_history', 'subscriber_count'):
batch_op.add_column(sa.Column('subscriber_count', sa.Integer(), nullable=True))
# Add thumbnail_url column
if not column_exists('watch_history', 'thumbnail_url'):
batch_op.add_column(sa.Column('thumbnail_url', sa.String(255), nullable=True))
# Add upload_date column
if not column_exists('watch_history', 'upload_date'):
batch_op.add_column(sa.Column('upload_date', sa.DateTime(), nullable=True))
# Backfill data
conn = op.get_bind()
conn.execute(text("""
UPDATE watch_history
SET category = COALESCE(category, 'Unknown'),
view_count = COALESCE(view_count, 0),
subscriber_count = COALESCE(subscriber_count, 0),
thumbnail_url = COALESCE(thumbnail_url, ''),
upload_date = COALESCE(upload_date, watch_date)
WHERE category IS NULL
OR view_count IS NULL
OR subscriber_count IS NULL
OR thumbnail_url IS NULL
OR upload_date IS NULL
"""))
def downgrade():
with op.batch_alter_table('watch_history') as batch_op:
if column_exists('watch_history', 'upload_date'):
batch_op.drop_column('upload_date')
if column_exists('watch_history', 'thumbnail_url'):
batch_op.drop_column('thumbnail_url')
if column_exists('watch_history', 'subscriber_count'):
batch_op.drop_column('subscriber_count')
if column_exists('watch_history', 'view_count'):
batch_op.drop_column('view_count')
if column_exists('watch_history', 'category'):
batch_op.drop_column('category')

View File

@ -10,7 +10,7 @@ Restart=on-failure
Environment="MPV_SOCKET=/tmp/mpvsocket"
Environment="LISTEN_ADDRESS=0.0.0.0"
Environment="LISTEN_PORT=42069"
Environment="DATABASE_URL=mysql+mysqldb://user:password@localhost:3306/mpv"
Environment="SQLALCHEMY_DATABASE_URI=mysql+mysqldb://user:password@localhost:3306/mpv"
[Install]
WantedBy=multi-user.target

View File

@ -1,18 +1,20 @@
blinker==1.8.2
alembic==1.14.1
blinker==1.9.0
cffi==1.17.1
click==8.1.7
cryptography==43.0.1
Flask==3.0.3
Flask==3.1.0
greenlet==3.1.1
gunicorn==23.0.0
itsdangerous==2.2.0
Jinja2==3.1.4
Jinja2==3.1.5
Mako==1.3.9
MarkupSafe==2.1.5
oracledb==2.4.1
oracledb==2.5.1
packaging==24.1
psycopg2-binary==2.9.10
pycparser==2.22
PyMySQL==1.1.1
SQLAlchemy==2.0.34
SQLAlchemy==2.0.38
typing_extensions==4.12.2
Werkzeug==3.0.4
Werkzeug==3.1.3