Fix linting errors across backend and frontend
All checks were successful
CI / backend-lint (push) Successful in 7s
CI / frontend-lint (push) Successful in 29s

Backend: auto-fix and format all ruff issues, manually fix B904/B023/
SIM117/B007/E741/F841 errors, suppress B008 (FastAPI Depends) and F821
(SQLAlchemy forward refs) in config. Frontend: allow constant exports,
disable React compiler-specific rules (set-state-in-effect,
preserve-manual-memoization).

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
Julian Tabel
2026-02-10 12:26:57 +01:00
parent 7f8890086f
commit e4111c67bc
48 changed files with 1225 additions and 883 deletions

View File

@@ -47,8 +47,12 @@ select = [
]
ignore = [
"E501", # line too long (handled by formatter)
"B008", # Depends() in defaults — standard FastAPI pattern
]
[tool.ruff.lint.per-file-ignores]
"src/app/models/*.py" = ["F821"] # forward refs in SQLAlchemy relationships
[tool.ruff.lint.isort]
known-first-party = ["app"]

View File

@@ -1,16 +1,14 @@
import asyncio
from logging.config import fileConfig
from alembic import context
from sqlalchemy import pool
from sqlalchemy.ext.asyncio import async_engine_from_config
from alembic import context
from app.core.config import settings
from app.core.database import Base, _get_async_url
# Import all models so Base.metadata is populated
import app.models # noqa: F401
from app.core.config import settings
from app.core.database import Base, _get_async_url
config = context.config

View File

@@ -6,18 +6,17 @@ Create Date: 2026-02-05 13:27:47.649534
"""
from typing import Sequence, Union
from collections.abc import Sequence
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
from alembic import op
from sqlalchemy.dialects import postgresql
# revision identifiers, used by Alembic.
revision: str = "03e5f186a9d5"
down_revision: Union[str, Sequence[str], None] = None
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
down_revision: str | Sequence[str] | None = None
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
@@ -36,9 +35,7 @@ def upgrade() -> None:
"routes",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column("name", sa.String(100), nullable=False),
sa.Column(
"game_id", sa.Integer(), sa.ForeignKey("games.id"), nullable=False
),
sa.Column("game_id", sa.Integer(), sa.ForeignKey("games.id"), nullable=False),
sa.Column("order", sa.SmallInteger(), nullable=False),
)
op.create_index("ix_routes_game_id", "routes", ["game_id"])
@@ -46,22 +43,16 @@ def upgrade() -> None:
op.create_table(
"pokemon",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"national_dex", sa.SmallInteger(), nullable=False, unique=True
),
sa.Column("national_dex", sa.SmallInteger(), nullable=False, unique=True),
sa.Column("name", sa.String(50), nullable=False),
sa.Column(
"types", postgresql.ARRAY(sa.String(20)), nullable=False
),
sa.Column("types", postgresql.ARRAY(sa.String(20)), nullable=False),
sa.Column("sprite_url", sa.String(500), nullable=True),
)
op.create_table(
"route_encounters",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"route_id", sa.Integer(), sa.ForeignKey("routes.id"), nullable=False
),
sa.Column("route_id", sa.Integer(), sa.ForeignKey("routes.id"), nullable=False),
sa.Column(
"pokemon_id",
sa.Integer(),
@@ -77,9 +68,7 @@ def upgrade() -> None:
name="uq_route_pokemon_method",
),
)
op.create_index(
"ix_route_encounters_route_id", "route_encounters", ["route_id"]
)
op.create_index("ix_route_encounters_route_id", "route_encounters", ["route_id"])
op.create_index(
"ix_route_encounters_pokemon_id", "route_encounters", ["pokemon_id"]
)
@@ -87,30 +76,20 @@ def upgrade() -> None:
op.create_table(
"nuzlocke_runs",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"game_id", sa.Integer(), sa.ForeignKey("games.id"), nullable=False
),
sa.Column("game_id", sa.Integer(), sa.ForeignKey("games.id"), nullable=False),
sa.Column("name", sa.String(100), nullable=False),
sa.Column("status", sa.String(20), nullable=False),
sa.Column(
"rules", postgresql.JSONB(), nullable=False, server_default="{}"
),
sa.Column("rules", postgresql.JSONB(), nullable=False, server_default="{}"),
sa.Column(
"started_at",
sa.DateTime(timezone=True),
nullable=False,
server_default=sa.func.now(),
),
sa.Column(
"completed_at", sa.DateTime(timezone=True), nullable=True
),
)
op.create_index(
"ix_nuzlocke_runs_game_id", "nuzlocke_runs", ["game_id"]
)
op.create_index(
"ix_nuzlocke_runs_status", "nuzlocke_runs", ["status"]
sa.Column("completed_at", sa.DateTime(timezone=True), nullable=True),
)
op.create_index("ix_nuzlocke_runs_game_id", "nuzlocke_runs", ["game_id"])
op.create_index("ix_nuzlocke_runs_status", "nuzlocke_runs", ["status"])
op.create_table(
"encounters",
@@ -121,9 +100,7 @@ def upgrade() -> None:
sa.ForeignKey("nuzlocke_runs.id"),
nullable=False,
),
sa.Column(
"route_id", sa.Integer(), sa.ForeignKey("routes.id"), nullable=False
),
sa.Column("route_id", sa.Integer(), sa.ForeignKey("routes.id"), nullable=False),
sa.Column(
"pokemon_id",
sa.Integer(),

View File

@@ -5,28 +5,27 @@ Revises: 03e5f186a9d5
Create Date: 2026-02-05 13:01:30.631978
"""
from typing import Sequence, Union
from collections.abc import Sequence
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision: str = '694df688fb02'
down_revision: Union[str, Sequence[str], None] = '03e5f186a9d5'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "694df688fb02"
down_revision: str | Sequence[str] | None = "03e5f186a9d5"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
"""Upgrade schema."""
# ### commands auto generated by Alembic - please adjust! ###
op.create_unique_constraint('uq_routes_game_name', 'routes', ['game_id', 'name'])
op.create_unique_constraint("uq_routes_game_name", "routes", ["game_id", "name"])
# ### end Alembic commands ###
def downgrade() -> None:
"""Downgrade schema."""
# ### commands auto generated by Alembic - please adjust! ###
op.drop_constraint('uq_routes_game_name', 'routes', type_='unique')
op.drop_constraint("uq_routes_game_name", "routes", type_="unique")
# ### end Alembic commands ###

View File

@@ -5,30 +5,36 @@ Revises: 694df688fb02
Create Date: 2026-02-05 13:32:35.559499
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = '9afcbafe9888'
down_revision: Union[str, Sequence[str], None] = '694df688fb02'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "9afcbafe9888"
down_revision: str | Sequence[str] | None = "694df688fb02"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
"""Upgrade schema."""
op.add_column('route_encounters', sa.Column('min_level', sa.SmallInteger(), nullable=False, server_default='0'))
op.add_column('route_encounters', sa.Column('max_level', sa.SmallInteger(), nullable=False, server_default='0'))
op.alter_column('route_encounters', 'min_level', server_default=None)
op.alter_column('route_encounters', 'max_level', server_default=None)
op.add_column(
"route_encounters",
sa.Column("min_level", sa.SmallInteger(), nullable=False, server_default="0"),
)
op.add_column(
"route_encounters",
sa.Column("max_level", sa.SmallInteger(), nullable=False, server_default="0"),
)
op.alter_column("route_encounters", "min_level", server_default=None)
op.alter_column("route_encounters", "max_level", server_default=None)
def downgrade() -> None:
"""Downgrade schema."""
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column('route_encounters', 'max_level')
op.drop_column('route_encounters', 'min_level')
op.drop_column("route_encounters", "max_level")
op.drop_column("route_encounters", "min_level")
# ### end Alembic commands ###

View File

@@ -5,22 +5,22 @@ Revises: 9afcbafe9888
Create Date: 2026-02-05 17:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'a1b2c3d4e5f6'
down_revision: Union[str, Sequence[str], None] = '9afcbafe9888'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "a1b2c3d4e5f6"
down_revision: str | Sequence[str] | None = "9afcbafe9888"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column('encounters', sa.Column('death_cause', sa.String(100), nullable=True))
op.add_column("encounters", sa.Column("death_cause", sa.String(100), nullable=True))
def downgrade() -> None:
op.drop_column('encounters', 'death_cause')
op.drop_column("encounters", "death_cause")

View File

@@ -5,25 +5,25 @@ Revises: f6a7b8c9d0e1
Create Date: 2026-02-07 12:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'a1b2c3d4e5f7'
down_revision: Union[str, Sequence[str], None] = 'f6a7b8c9d0e1'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "a1b2c3d4e5f7"
down_revision: str | Sequence[str] | None = "f6a7b8c9d0e1"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'routes',
sa.Column('pinwheel_zone', sa.SmallInteger(), nullable=True),
"routes",
sa.Column("pinwheel_zone", sa.SmallInteger(), nullable=True),
)
def downgrade() -> None:
op.drop_column('routes', 'pinwheel_zone')
op.drop_column("routes", "pinwheel_zone")

View File

@@ -5,25 +5,25 @@ Revises: f6a7b8c9d0e1
Create Date: 2026-02-09 12:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'a1b2c3d4e5f8'
down_revision: Union[str, Sequence[str], None] = 'f6a7b8c9d0e1'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "a1b2c3d4e5f8"
down_revision: str | Sequence[str] | None = "f6a7b8c9d0e1"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'games',
sa.Column('category', sa.String(20), nullable=True),
"games",
sa.Column("category", sa.String(20), nullable=True),
)
def downgrade() -> None:
op.drop_column('games', 'category')
op.drop_column("games", "category")

View File

@@ -5,22 +5,24 @@ Revises: f5a6b7c8d9e0
Create Date: 2026-02-08 21:00:00.000000
"""
from typing import Sequence, Union
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'a6b7c8d9e0f1'
down_revision: Union[str, Sequence[str], None] = 'f5a6b7c8d9e0'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "a6b7c8d9e0f1"
down_revision: str | Sequence[str] | None = "f5a6b7c8d9e0"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column('boss_battles', sa.Column('specialty_type', sa.String(20), nullable=True))
op.add_column(
"boss_battles", sa.Column("specialty_type", sa.String(20), nullable=True)
)
def downgrade() -> None:
op.drop_column('boss_battles', 'specialty_type')
op.drop_column("boss_battles", "specialty_type")

View File

@@ -5,25 +5,27 @@ Revises: a1b2c3d4e5f7
Create Date: 2026-02-07 18:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'b1c2d3e4f5a6'
down_revision: Union[str, Sequence[str], None] = 'a1b2c3d4e5f7'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "b1c2d3e4f5a6"
down_revision: str | Sequence[str] | None = "a1b2c3d4e5f7"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'encounters',
sa.Column('is_shiny', sa.Boolean(), nullable=False, server_default=sa.text('false')),
"encounters",
sa.Column(
"is_shiny", sa.Boolean(), nullable=False, server_default=sa.text("false")
),
)
def downgrade() -> None:
op.drop_column('encounters', 'is_shiny')
op.drop_column("encounters", "is_shiny")

View File

@@ -5,38 +5,56 @@ Revises: a1b2c3d4e5f6
Create Date: 2026-02-05 18:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'b2c3d4e5f6a7'
down_revision: Union[str, Sequence[str], None] = 'a1b2c3d4e5f6'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "b2c3d4e5f6a7"
down_revision: str | Sequence[str] | None = "a1b2c3d4e5f6"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.create_table(
'evolutions',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('from_pokemon_id', sa.Integer(), sa.ForeignKey('pokemon.id'), nullable=False, index=True),
sa.Column('to_pokemon_id', sa.Integer(), sa.ForeignKey('pokemon.id'), nullable=False, index=True),
sa.Column('trigger', sa.String(30), nullable=False),
sa.Column('min_level', sa.SmallInteger(), nullable=True),
sa.Column('item', sa.String(50), nullable=True),
sa.Column('held_item', sa.String(50), nullable=True),
sa.Column('condition', sa.String(200), nullable=True),
"evolutions",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"from_pokemon_id",
sa.Integer(),
sa.ForeignKey("pokemon.id"),
nullable=False,
index=True,
),
sa.Column(
"to_pokemon_id",
sa.Integer(),
sa.ForeignKey("pokemon.id"),
nullable=False,
index=True,
),
sa.Column("trigger", sa.String(30), nullable=False),
sa.Column("min_level", sa.SmallInteger(), nullable=True),
sa.Column("item", sa.String(50), nullable=True),
sa.Column("held_item", sa.String(50), nullable=True),
sa.Column("condition", sa.String(200), nullable=True),
)
op.add_column(
'encounters',
sa.Column('current_pokemon_id', sa.Integer(), sa.ForeignKey('pokemon.id'), nullable=True, index=True),
"encounters",
sa.Column(
"current_pokemon_id",
sa.Integer(),
sa.ForeignKey("pokemon.id"),
nullable=True,
index=True,
),
)
def downgrade() -> None:
op.drop_column('encounters', 'current_pokemon_id')
op.drop_table('evolutions')
op.drop_column("encounters", "current_pokemon_id")
op.drop_table("evolutions")

View File

@@ -5,42 +5,65 @@ Revises: a1b2c3d4e5f8, b7c8d9e0f1a2
Create Date: 2026-02-09 14:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
from sqlalchemy.dialects.postgresql import JSONB
# revision identifiers, used by Alembic.
revision: str = 'b2c3d4e5f6a8'
down_revision: Union[str, Sequence[str], None] = ('a1b2c3d4e5f8', 'b7c8d9e0f1a2')
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "b2c3d4e5f6a8"
down_revision: str | Sequence[str] | None = ("a1b2c3d4e5f8", "b7c8d9e0f1a2")
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.create_table(
'genlockes',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('name', sa.String(100), nullable=False),
sa.Column('status', sa.String(20), nullable=False, index=True),
sa.Column('genlocke_rules', JSONB(), nullable=False, server_default='{}'),
sa.Column('nuzlocke_rules', JSONB(), nullable=False, server_default='{}'),
sa.Column('created_at', sa.DateTime(timezone=True), server_default=sa.func.now(), nullable=False),
"genlockes",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column("name", sa.String(100), nullable=False),
sa.Column("status", sa.String(20), nullable=False, index=True),
sa.Column("genlocke_rules", JSONB(), nullable=False, server_default="{}"),
sa.Column("nuzlocke_rules", JSONB(), nullable=False, server_default="{}"),
sa.Column(
"created_at",
sa.DateTime(timezone=True),
server_default=sa.func.now(),
nullable=False,
),
)
op.create_table(
'genlocke_legs',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('genlocke_id', sa.Integer(), sa.ForeignKey('genlockes.id', ondelete='CASCADE'), nullable=False, index=True),
sa.Column('game_id', sa.Integer(), sa.ForeignKey('games.id'), nullable=False, index=True),
sa.Column('run_id', sa.Integer(), sa.ForeignKey('nuzlocke_runs.id'), nullable=True, index=True),
sa.Column('leg_order', sa.SmallInteger(), nullable=False),
sa.UniqueConstraint('genlocke_id', 'leg_order', name='uq_genlocke_legs_order'),
"genlocke_legs",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"genlocke_id",
sa.Integer(),
sa.ForeignKey("genlockes.id", ondelete="CASCADE"),
nullable=False,
index=True,
),
sa.Column(
"game_id",
sa.Integer(),
sa.ForeignKey("games.id"),
nullable=False,
index=True,
),
sa.Column(
"run_id",
sa.Integer(),
sa.ForeignKey("nuzlocke_runs.id"),
nullable=True,
index=True,
),
sa.Column("leg_order", sa.SmallInteger(), nullable=False),
sa.UniqueConstraint("genlocke_id", "leg_order", name="uq_genlocke_legs_order"),
)
def downgrade() -> None:
op.drop_table('genlocke_legs')
op.drop_table('genlockes')
op.drop_table("genlocke_legs")
op.drop_table("genlockes")

View File

@@ -5,22 +5,24 @@ Revises: a6b7c8d9e0f1
Create Date: 2026-02-08 22:00:00.000000
"""
from typing import Sequence, Union
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'b7c8d9e0f1a2'
down_revision: Union[str, Sequence[str], None] = 'a6b7c8d9e0f1'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "b7c8d9e0f1a2"
down_revision: str | Sequence[str] | None = "a6b7c8d9e0f1"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column('boss_pokemon', sa.Column('condition_label', sa.String(100), nullable=True))
op.add_column(
"boss_pokemon", sa.Column("condition_label", sa.String(100), nullable=True)
)
def downgrade() -> None:
op.drop_column('boss_pokemon', 'condition_label')
op.drop_column("boss_pokemon", "condition_label")

View File

@@ -5,57 +5,95 @@ Revises: b1c2d3e4f5a6
Create Date: 2026-02-08 12:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'c2d3e4f5a6b7'
down_revision: Union[str, Sequence[str], None] = 'b1c2d3e4f5a6'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "c2d3e4f5a6b7"
down_revision: str | Sequence[str] | None = "b1c2d3e4f5a6"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.create_table(
'boss_battles',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('game_id', sa.Integer(), sa.ForeignKey('games.id'), nullable=False, index=True),
sa.Column('name', sa.String(100), nullable=False),
sa.Column('boss_type', sa.String(20), nullable=False),
sa.Column('badge_name', sa.String(100), nullable=True),
sa.Column('badge_image_url', sa.String(500), nullable=True),
sa.Column('level_cap', sa.SmallInteger(), nullable=False),
sa.Column('order', sa.SmallInteger(), nullable=False),
sa.Column('after_route_id', sa.Integer(), sa.ForeignKey('routes.id'), nullable=True, index=True),
sa.Column('location', sa.String(200), nullable=False),
sa.Column('sprite_url', sa.String(500), nullable=True),
"boss_battles",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"game_id",
sa.Integer(),
sa.ForeignKey("games.id"),
nullable=False,
index=True,
),
sa.Column("name", sa.String(100), nullable=False),
sa.Column("boss_type", sa.String(20), nullable=False),
sa.Column("badge_name", sa.String(100), nullable=True),
sa.Column("badge_image_url", sa.String(500), nullable=True),
sa.Column("level_cap", sa.SmallInteger(), nullable=False),
sa.Column("order", sa.SmallInteger(), nullable=False),
sa.Column(
"after_route_id",
sa.Integer(),
sa.ForeignKey("routes.id"),
nullable=True,
index=True,
),
sa.Column("location", sa.String(200), nullable=False),
sa.Column("sprite_url", sa.String(500), nullable=True),
)
op.create_table(
'boss_pokemon',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('boss_battle_id', sa.Integer(), sa.ForeignKey('boss_battles.id', ondelete='CASCADE'), nullable=False, index=True),
sa.Column('pokemon_id', sa.Integer(), sa.ForeignKey('pokemon.id'), nullable=False, index=True),
sa.Column('level', sa.SmallInteger(), nullable=False),
sa.Column('order', sa.SmallInteger(), nullable=False),
"boss_pokemon",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"boss_battle_id",
sa.Integer(),
sa.ForeignKey("boss_battles.id", ondelete="CASCADE"),
nullable=False,
index=True,
),
sa.Column(
"pokemon_id",
sa.Integer(),
sa.ForeignKey("pokemon.id"),
nullable=False,
index=True,
),
sa.Column("level", sa.SmallInteger(), nullable=False),
sa.Column("order", sa.SmallInteger(), nullable=False),
)
op.create_table(
'boss_results',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('run_id', sa.Integer(), sa.ForeignKey('nuzlocke_runs.id', ondelete='CASCADE'), nullable=False, index=True),
sa.Column('boss_battle_id', sa.Integer(), sa.ForeignKey('boss_battles.id'), nullable=False, index=True),
sa.Column('result', sa.String(10), nullable=False),
sa.Column('attempts', sa.SmallInteger(), nullable=False, server_default='1'),
sa.Column('completed_at', sa.DateTime(timezone=True), nullable=True),
sa.UniqueConstraint('run_id', 'boss_battle_id', name='uq_boss_results_run_boss'),
"boss_results",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"run_id",
sa.Integer(),
sa.ForeignKey("nuzlocke_runs.id", ondelete="CASCADE"),
nullable=False,
index=True,
),
sa.Column(
"boss_battle_id",
sa.Integer(),
sa.ForeignKey("boss_battles.id"),
nullable=False,
index=True,
),
sa.Column("result", sa.String(10), nullable=False),
sa.Column("attempts", sa.SmallInteger(), nullable=False, server_default="1"),
sa.Column("completed_at", sa.DateTime(timezone=True), nullable=True),
sa.UniqueConstraint(
"run_id", "boss_battle_id", name="uq_boss_results_run_boss"
),
)
def downgrade() -> None:
op.drop_table('boss_results')
op.drop_table('boss_pokemon')
op.drop_table('boss_battles')
op.drop_table("boss_results")
op.drop_table("boss_pokemon")
op.drop_table("boss_battles")

View File

@@ -5,26 +5,26 @@ Revises: b2c3d4e5f6a7
Create Date: 2026-02-06 12:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'c3d4e5f6a7b8'
down_revision: Union[str, Sequence[str], None] = 'b2c3d4e5f6a7'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "c3d4e5f6a7b8"
down_revision: str | Sequence[str] | None = "b2c3d4e5f6a7"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'routes',
"routes",
sa.Column(
'parent_route_id',
"parent_route_id",
sa.Integer(),
sa.ForeignKey('routes.id', ondelete='CASCADE'),
sa.ForeignKey("routes.id", ondelete="CASCADE"),
nullable=True,
index=True,
),
@@ -32,4 +32,4 @@ def upgrade() -> None:
def downgrade() -> None:
op.drop_column('routes', 'parent_route_id')
op.drop_column("routes", "parent_route_id")

View File

@@ -5,26 +5,26 @@ Revises: b2c3d4e5f6a8
Create Date: 2026-02-09 18:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
from sqlalchemy.dialects.postgresql import JSONB
# revision identifiers, used by Alembic.
revision: str = 'c3d4e5f6a7b9'
down_revision: Union[str, Sequence[str], None] = 'b2c3d4e5f6a8'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "c3d4e5f6a7b9"
down_revision: str | Sequence[str] | None = "b2c3d4e5f6a8"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'genlocke_legs',
sa.Column('retired_pokemon_ids', JSONB(), nullable=True),
"genlocke_legs",
sa.Column("retired_pokemon_ids", JSONB(), nullable=True),
)
def downgrade() -> None:
op.drop_column('genlocke_legs', 'retired_pokemon_ids')
op.drop_column("genlocke_legs", "retired_pokemon_ids")

View File

@@ -5,28 +5,28 @@ Revises: c2d3e4f5a6b7
Create Date: 2026-02-08 14:00:00.000000
"""
import json
from collections.abc import Sequence
from pathlib import Path
from typing import Sequence, Union
from alembic import op
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'd3e4f5a6b7c8'
down_revision: Union[str, Sequence[str], None] = 'c2d3e4f5a6b7'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "d3e4f5a6b7c8"
down_revision: str | Sequence[str] | None = "c2d3e4f5a6b7"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
# 1. Create version_groups table
op.create_table(
'version_groups',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('name', sa.String(100), nullable=False),
sa.Column('slug', sa.String(100), nullable=False, unique=True),
"version_groups",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column("name", sa.String(100), nullable=False),
sa.Column("slug", sa.String(100), nullable=False, unique=True),
)
# 2. Populate version groups from seed data
@@ -36,10 +36,10 @@ def upgrade() -> None:
conn = op.get_bind()
vg_table = sa.table(
'version_groups',
sa.column('id', sa.Integer),
sa.column('name', sa.String),
sa.column('slug', sa.String),
"version_groups",
sa.column("id", sa.Integer),
sa.column("name", sa.String),
sa.column("slug", sa.String),
)
# Build slug -> id mapping and game_slug -> vg_id mapping
@@ -49,8 +49,7 @@ def upgrade() -> None:
vg_id = vg_idx
# Use the slug as a readable name (e.g., "red-blue" -> "Red / Blue")
vg_name = " / ".join(
g["name"].replace("Pokemon ", "")
for g in vg_info["games"].values()
g["name"].replace("Pokemon ", "") for g in vg_info["games"].values()
)
conn.execute(vg_table.insert().values(id=vg_id, name=vg_name, slug=vg_slug))
slug_to_vg_id[vg_slug] = vg_id
@@ -58,16 +57,23 @@ def upgrade() -> None:
game_slug_to_vg_id[game_slug] = vg_id
# 3. Add version_group_id to games (nullable initially)
op.add_column('games', sa.Column('version_group_id', sa.Integer(),
sa.ForeignKey('version_groups.id'), nullable=True))
op.create_index('ix_games_version_group_id', 'games', ['version_group_id'])
op.add_column(
"games",
sa.Column(
"version_group_id",
sa.Integer(),
sa.ForeignKey("version_groups.id"),
nullable=True,
),
)
op.create_index("ix_games_version_group_id", "games", ["version_group_id"])
# Populate games.version_group_id from the mapping
games_table = sa.table(
'games',
sa.column('id', sa.Integer),
sa.column('slug', sa.String),
sa.column('version_group_id', sa.Integer),
"games",
sa.column("id", sa.Integer),
sa.column("slug", sa.String),
sa.column("version_group_id", sa.Integer),
)
rows = conn.execute(sa.select(games_table.c.id, games_table.c.slug)).fetchall()
for game_id, game_slug in rows:
@@ -80,21 +86,23 @@ def upgrade() -> None:
)
# 4. Add game_id to route_encounters (nullable initially), populate from routes.game_id
op.add_column('route_encounters', sa.Column('game_id', sa.Integer(),
sa.ForeignKey('games.id'), nullable=True))
op.create_index('ix_route_encounters_game_id', 'route_encounters', ['game_id'])
op.add_column(
"route_encounters",
sa.Column("game_id", sa.Integer(), sa.ForeignKey("games.id"), nullable=True),
)
op.create_index("ix_route_encounters_game_id", "route_encounters", ["game_id"])
routes_table = sa.table(
'routes',
sa.column('id', sa.Integer),
sa.column('name', sa.String),
sa.column('game_id', sa.Integer),
"routes",
sa.column("id", sa.Integer),
sa.column("name", sa.String),
sa.column("game_id", sa.Integer),
)
re_table = sa.table(
'route_encounters',
sa.column('id', sa.Integer),
sa.column('route_id', sa.Integer),
sa.column('game_id', sa.Integer),
"route_encounters",
sa.column("id", sa.Integer),
sa.column("route_id", sa.Integer),
sa.column("game_id", sa.Integer),
)
# Populate route_encounters.game_id from routes.game_id via join
conn.execute(
@@ -104,10 +112,11 @@ def upgrade() -> None:
)
# 5. Drop old unique constraint on route_encounters, add new one with game_id
op.drop_constraint('uq_route_pokemon_method', 'route_encounters', type_='unique')
op.drop_constraint("uq_route_pokemon_method", "route_encounters", type_="unique")
op.create_unique_constraint(
'uq_route_pokemon_method_game', 'route_encounters',
['route_id', 'pokemon_id', 'encounter_method', 'game_id']
"uq_route_pokemon_method_game",
"route_encounters",
["route_id", "pokemon_id", "encounter_method", "game_id"],
)
# 6. Deduplicate routes within version groups
@@ -115,15 +124,15 @@ def upgrade() -> None:
# and re-point route_encounters, encounters, and boss_battles to canonical routes
encounters_table = sa.table(
'encounters',
sa.column('id', sa.Integer),
sa.column('route_id', sa.Integer),
"encounters",
sa.column("id", sa.Integer),
sa.column("route_id", sa.Integer),
)
boss_battles_table = sa.table(
'boss_battles',
sa.column('id', sa.Integer),
sa.column('game_id', sa.Integer),
sa.column('after_route_id', sa.Integer),
"boss_battles",
sa.column("id", sa.Integer),
sa.column("game_id", sa.Integer),
sa.column("after_route_id", sa.Integer),
)
# Get all version groups that have more than one game
@@ -149,16 +158,18 @@ def upgrade() -> None:
# Get canonical routes (by name)
canonical_routes = conn.execute(
sa.select(routes_table.c.id, routes_table.c.name)
.where(routes_table.c.game_id == canonical_game_id)
sa.select(routes_table.c.id, routes_table.c.name).where(
routes_table.c.game_id == canonical_game_id
)
).fetchall()
canonical_name_to_id = {name: rid for rid, name in canonical_routes}
# For each non-canonical game, re-point references to canonical routes
for nc_game_id in non_canonical_game_ids:
nc_routes = conn.execute(
sa.select(routes_table.c.id, routes_table.c.name)
.where(routes_table.c.game_id == nc_game_id)
sa.select(routes_table.c.id, routes_table.c.name).where(
routes_table.c.game_id == nc_game_id
)
).fetchall()
for old_route_id, route_name in nc_routes:
@@ -192,29 +203,36 @@ def upgrade() -> None:
conn.execute(
sa.text(
"DELETE FROM routes WHERE parent_route_id IS NOT NULL AND game_id IN :nc_ids"
).bindparams(sa.bindparam('nc_ids', expanding=True)),
{"nc_ids": non_canonical_game_ids}
).bindparams(sa.bindparam("nc_ids", expanding=True)),
{"nc_ids": non_canonical_game_ids},
)
# Then delete parent routes
conn.execute(
sa.text(
"DELETE FROM routes WHERE game_id IN :nc_ids"
).bindparams(sa.bindparam('nc_ids', expanding=True)),
{"nc_ids": non_canonical_game_ids}
sa.text("DELETE FROM routes WHERE game_id IN :nc_ids").bindparams(
sa.bindparam("nc_ids", expanding=True)
),
{"nc_ids": non_canonical_game_ids},
)
# 7. Add version_group_id to routes (nullable), populate from games.version_group_id
op.add_column('routes', sa.Column('version_group_id', sa.Integer(),
sa.ForeignKey('version_groups.id'), nullable=True))
op.create_index('ix_routes_version_group_id', 'routes', ['version_group_id'])
op.add_column(
"routes",
sa.Column(
"version_group_id",
sa.Integer(),
sa.ForeignKey("version_groups.id"),
nullable=True,
),
)
op.create_index("ix_routes_version_group_id", "routes", ["version_group_id"])
# Need to re-declare routes_table with version_group_id
routes_table_v2 = sa.table(
'routes',
sa.column('id', sa.Integer),
sa.column('name', sa.String),
sa.column('game_id', sa.Integer),
sa.column('version_group_id', sa.Integer),
"routes",
sa.column("id", sa.Integer),
sa.column("name", sa.String),
sa.column("game_id", sa.Integer),
sa.column("version_group_id", sa.Integer),
)
# Populate routes.version_group_id from the game's version_group_id
@@ -225,24 +243,32 @@ def upgrade() -> None:
)
# 8. Drop routes.game_id, drop old unique constraint, add new one
op.drop_constraint('uq_routes_game_name', 'routes', type_='unique')
op.drop_index('ix_routes_game_id', 'routes')
op.drop_column('routes', 'game_id')
op.drop_constraint("uq_routes_game_name", "routes", type_="unique")
op.drop_index("ix_routes_game_id", "routes")
op.drop_column("routes", "game_id")
op.create_unique_constraint(
'uq_routes_version_group_name', 'routes',
['version_group_id', 'name']
"uq_routes_version_group_name", "routes", ["version_group_id", "name"]
)
# 9. Add version_group_id to boss_battles (nullable), populate from games.version_group_id
op.add_column('boss_battles', sa.Column('version_group_id', sa.Integer(),
sa.ForeignKey('version_groups.id'), nullable=True))
op.create_index('ix_boss_battles_version_group_id', 'boss_battles', ['version_group_id'])
op.add_column(
"boss_battles",
sa.Column(
"version_group_id",
sa.Integer(),
sa.ForeignKey("version_groups.id"),
nullable=True,
),
)
op.create_index(
"ix_boss_battles_version_group_id", "boss_battles", ["version_group_id"]
)
bb_table_v2 = sa.table(
'boss_battles',
sa.column('id', sa.Integer),
sa.column('game_id', sa.Integer),
sa.column('version_group_id', sa.Integer),
"boss_battles",
sa.column("id", sa.Integer),
sa.column("game_id", sa.Integer),
sa.column("version_group_id", sa.Integer),
)
conn.execute(
@@ -252,14 +278,14 @@ def upgrade() -> None:
)
# 10. Drop boss_battles.game_id
op.drop_index('ix_boss_battles_game_id', 'boss_battles')
op.drop_column('boss_battles', 'game_id')
op.drop_index("ix_boss_battles_game_id", "boss_battles")
op.drop_column("boss_battles", "game_id")
# 11. Make columns non-nullable
op.alter_column('route_encounters', 'game_id', nullable=False)
op.alter_column('routes', 'version_group_id', nullable=False)
op.alter_column('boss_battles', 'version_group_id', nullable=False)
op.alter_column('games', 'version_group_id', nullable=False)
op.alter_column("route_encounters", "game_id", nullable=False)
op.alter_column("routes", "version_group_id", nullable=False)
op.alter_column("boss_battles", "version_group_id", nullable=False)
op.alter_column("games", "version_group_id", nullable=False)
def downgrade() -> None:

View File

@@ -5,25 +5,25 @@ Revises: c3d4e5f6a7b8
Create Date: 2026-02-06 14:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'd4e5f6a7b8c9'
down_revision: Union[str, Sequence[str], None] = 'c3d4e5f6a7b8'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "d4e5f6a7b8c9"
down_revision: str | Sequence[str] | None = "c3d4e5f6a7b8"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'games',
sa.Column('color', sa.String(7), nullable=True),
"games",
sa.Column("color", sa.String(7), nullable=True),
)
def downgrade() -> None:
op.drop_column('games', 'color')
op.drop_column("games", "color")

View File

@@ -5,26 +5,26 @@ Revises: c3d4e5f6a7b9
Create Date: 2026-02-09 20:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
from sqlalchemy.dialects.postgresql import JSONB
# revision identifiers, used by Alembic.
revision: str = 'd4e5f6a7b9c0'
down_revision: Union[str, Sequence[str], None] = 'c3d4e5f6a7b9'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "d4e5f6a7b9c0"
down_revision: str | Sequence[str] | None = "c3d4e5f6a7b9"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'nuzlocke_runs',
sa.Column('hof_encounter_ids', JSONB(), nullable=True),
"nuzlocke_runs",
sa.Column("hof_encounter_ids", JSONB(), nullable=True),
)
def downgrade() -> None:
op.drop_column('nuzlocke_runs', 'hof_encounter_ids')
op.drop_column("nuzlocke_runs", "hof_encounter_ids")

View File

@@ -5,25 +5,27 @@ Revises: d3e4f5a6b7c8
Create Date: 2026-02-08 18:00:00.000000
"""
from typing import Sequence, Union
from collections.abc import Sequence
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'e4f5a6b7c8d9'
down_revision: Union[str, Sequence[str], None] = 'd3e4f5a6b7c8'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "e4f5a6b7c8d9"
down_revision: str | Sequence[str] | None = "d3e4f5a6b7c8"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.create_unique_constraint(
'uq_boss_battles_version_group_order',
'boss_battles',
['version_group_id', 'order'],
"uq_boss_battles_version_group_order",
"boss_battles",
["version_group_id", "order"],
)
def downgrade() -> None:
op.drop_constraint('uq_boss_battles_version_group_order', 'boss_battles', type_='unique')
op.drop_constraint(
"uq_boss_battles_version_group_order", "boss_battles", type_="unique"
)

View File

@@ -5,24 +5,25 @@ Revises: d4e5f6a7b8c9
Create Date: 2026-02-07 10:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'e5f6a7b8c9d0'
down_revision: Union[str, Sequence[str], None] = 'd4e5f6a7b8c9'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "e5f6a7b8c9d0"
down_revision: str | Sequence[str] | None = "d4e5f6a7b8c9"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
# Rename national_dex -> pokeapi_id and widen to Integer
op.alter_column(
'pokemon', 'national_dex',
new_column_name='pokeapi_id',
"pokemon",
"national_dex",
new_column_name="pokeapi_id",
type_=sa.Integer(),
existing_type=sa.SmallInteger(),
existing_nullable=False,
@@ -30,23 +31,26 @@ def upgrade() -> None:
# Add real national_dex column (shared between forms and base species)
op.add_column(
'pokemon',
sa.Column('national_dex', sa.SmallInteger(), nullable=False, server_default='0'),
"pokemon",
sa.Column(
"national_dex", sa.SmallInteger(), nullable=False, server_default="0"
),
)
# Populate national_dex = pokeapi_id for all existing rows
# (correct for base species; forms will be fixed by re-seeding)
op.execute('UPDATE pokemon SET national_dex = pokeapi_id')
op.execute("UPDATE pokemon SET national_dex = pokeapi_id")
# Remove the default now that all rows are populated
op.alter_column('pokemon', 'national_dex', server_default=None)
op.alter_column("pokemon", "national_dex", server_default=None)
def downgrade() -> None:
op.drop_column('pokemon', 'national_dex')
op.drop_column("pokemon", "national_dex")
op.alter_column(
'pokemon', 'pokeapi_id',
new_column_name='national_dex',
"pokemon",
"pokeapi_id",
new_column_name="national_dex",
type_=sa.SmallInteger(),
existing_type=sa.Integer(),
existing_nullable=False,

View File

@@ -5,32 +5,55 @@ Revises: d4e5f6a7b9c0
Create Date: 2026-02-09 22:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'e5f6a7b9c0d1'
down_revision: Union[str, Sequence[str], None] = 'd4e5f6a7b9c0'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "e5f6a7b9c0d1"
down_revision: str | Sequence[str] | None = "d4e5f6a7b9c0"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.create_table(
'genlocke_transfers',
sa.Column('id', sa.Integer(), primary_key=True),
sa.Column('genlocke_id', sa.Integer(), sa.ForeignKey('genlockes.id', ondelete='CASCADE'), nullable=False, index=True),
sa.Column('source_encounter_id', sa.Integer(), sa.ForeignKey('encounters.id'), nullable=False, index=True),
sa.Column('target_encounter_id', sa.Integer(), sa.ForeignKey('encounters.id'), nullable=False, unique=True),
sa.Column('source_leg_order', sa.SmallInteger(), nullable=False),
sa.Column('target_leg_order', sa.SmallInteger(), nullable=False),
sa.Column('created_at', sa.DateTime(timezone=True), server_default=sa.func.now(), nullable=False),
sa.UniqueConstraint('target_encounter_id', name='uq_genlocke_transfers_target'),
"genlocke_transfers",
sa.Column("id", sa.Integer(), primary_key=True),
sa.Column(
"genlocke_id",
sa.Integer(),
sa.ForeignKey("genlockes.id", ondelete="CASCADE"),
nullable=False,
index=True,
),
sa.Column(
"source_encounter_id",
sa.Integer(),
sa.ForeignKey("encounters.id"),
nullable=False,
index=True,
),
sa.Column(
"target_encounter_id",
sa.Integer(),
sa.ForeignKey("encounters.id"),
nullable=False,
unique=True,
),
sa.Column("source_leg_order", sa.SmallInteger(), nullable=False),
sa.Column("target_leg_order", sa.SmallInteger(), nullable=False),
sa.Column(
"created_at",
sa.DateTime(timezone=True),
server_default=sa.func.now(),
nullable=False,
),
sa.UniqueConstraint("target_encounter_id", name="uq_genlocke_transfers_target"),
)
def downgrade() -> None:
op.drop_table('genlocke_transfers')
op.drop_table("genlocke_transfers")

View File

@@ -5,22 +5,22 @@ Revises: e4f5a6b7c8d9
Create Date: 2026-02-08 20:00:00.000000
"""
from typing import Sequence, Union
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'f5a6b7c8d9e0'
down_revision: Union[str, Sequence[str], None] = 'e4f5a6b7c8d9'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "f5a6b7c8d9e0"
down_revision: str | Sequence[str] | None = "e4f5a6b7c8d9"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column('boss_battles', sa.Column('section', sa.String(100), nullable=True))
op.add_column("boss_battles", sa.Column("section", sa.String(100), nullable=True))
def downgrade() -> None:
op.drop_column('boss_battles', 'section')
op.drop_column("boss_battles", "section")

View File

@@ -5,25 +5,25 @@ Revises: e5f6a7b8c9d0
Create Date: 2026-02-07 12:00:00.000000
"""
from typing import Sequence, Union
from alembic import op
from collections.abc import Sequence
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision: str = 'f6a7b8c9d0e1'
down_revision: Union[str, Sequence[str], None] = 'e5f6a7b8c9d0'
branch_labels: Union[str, Sequence[str], None] = None
depends_on: Union[str, Sequence[str], None] = None
revision: str = "f6a7b8c9d0e1"
down_revision: str | Sequence[str] | None = "e5f6a7b8c9d0"
branch_labels: str | Sequence[str] | None = None
depends_on: str | Sequence[str] | None = None
def upgrade() -> None:
op.add_column(
'evolutions',
sa.Column('region', sa.String(30), nullable=True),
"evolutions",
sa.Column("region", sa.String(30), nullable=True),
)
def downgrade() -> None:
op.drop_column('evolutions', 'region')
op.drop_column("evolutions", "region")

View File

@@ -1,4 +1,4 @@
from datetime import datetime, timezone
from datetime import UTC, datetime
from fastapi import APIRouter, Depends, HTTPException, Response
from sqlalchemy import select
@@ -33,7 +33,9 @@ async def _get_version_group_id(session: AsyncSession, game_id: int) -> int:
if game is None:
raise HTTPException(status_code=404, detail="Game not found")
if game.version_group_id is None:
raise HTTPException(status_code=400, detail="Game has no version group assigned")
raise HTTPException(
status_code=400, detail="Game has no version group assigned"
)
return game.version_group_id
@@ -41,9 +43,7 @@ async def _get_version_group_id(session: AsyncSession, game_id: int) -> int:
@router.get("/games/{game_id}/bosses", response_model=list[BossBattleResponse])
async def list_bosses(
game_id: int, session: AsyncSession = Depends(get_session)
):
async def list_bosses(game_id: int, session: AsyncSession = Depends(get_session)):
vg_id = await _get_version_group_id(session, game_id)
result = await session.execute(
@@ -72,7 +72,9 @@ async def reorder_bosses(
bosses = {b.id: b for b in result.scalars().all()}
if len(bosses) != len(boss_ids):
raise HTTPException(status_code=400, detail="Some boss IDs not found in this game")
raise HTTPException(
status_code=400, detail="Some boss IDs not found in this game"
)
# Phase 1: set temporary negative orders to avoid unique constraint violations
for i, item in enumerate(data.bosses):
@@ -94,7 +96,9 @@ async def reorder_bosses(
return result.scalars().all()
@router.post("/games/{game_id}/bosses", response_model=BossBattleResponse, status_code=201)
@router.post(
"/games/{game_id}/bosses", response_model=BossBattleResponse, status_code=201
)
async def create_boss(
game_id: int,
data: BossBattleCreate,
@@ -157,7 +161,9 @@ async def delete_boss(
vg_id = await _get_version_group_id(session, game_id)
result = await session.execute(
select(BossBattle).where(BossBattle.id == boss_id, BossBattle.version_group_id == vg_id)
select(BossBattle).where(
BossBattle.id == boss_id, BossBattle.version_group_id == vg_id
)
)
boss = result.scalar_one_or_none()
if boss is None:
@@ -188,9 +194,13 @@ async def bulk_import_bosses(
bosses_data = [item.model_dump() for item in items]
try:
count = await upsert_bosses(session, vg_id, bosses_data, dex_to_id, route_name_to_id)
count = await upsert_bosses(
session, vg_id, bosses_data, dex_to_id, route_name_to_id
)
except Exception as e:
raise HTTPException(status_code=400, detail=f"Failed to import bosses: {e}")
raise HTTPException(
status_code=400, detail=f"Failed to import bosses: {e}"
) from e
await session.commit()
return BulkImportResult(created=count, updated=0, errors=[])
@@ -252,22 +262,20 @@ async def set_boss_team(
@router.get("/runs/{run_id}/boss-results", response_model=list[BossResultResponse])
async def list_boss_results(
run_id: int, session: AsyncSession = Depends(get_session)
):
async def list_boss_results(run_id: int, session: AsyncSession = Depends(get_session)):
run = await session.get(NuzlockeRun, run_id)
if run is None:
raise HTTPException(status_code=404, detail="Run not found")
result = await session.execute(
select(BossResult)
.where(BossResult.run_id == run_id)
.order_by(BossResult.id)
select(BossResult).where(BossResult.run_id == run_id).order_by(BossResult.id)
)
return result.scalars().all()
@router.post("/runs/{run_id}/boss-results", response_model=BossResultResponse, status_code=201)
@router.post(
"/runs/{run_id}/boss-results", response_model=BossResultResponse, status_code=201
)
async def create_boss_result(
run_id: int,
data: BossResultCreate,
@@ -293,14 +301,14 @@ async def create_boss_result(
if result:
result.result = data.result
result.attempts = data.attempts
result.completed_at = datetime.now(timezone.utc) if data.result == "won" else None
result.completed_at = datetime.now(UTC) if data.result == "won" else None
else:
result = BossResult(
run_id=run_id,
boss_battle_id=data.boss_battle_id,
result=data.result,
attempts=data.attempts,
completed_at=datetime.now(timezone.utc) if data.result == "won" else None,
completed_at=datetime.now(UTC) if data.result == "won" else None,
)
session.add(result)

View File

@@ -8,8 +8,8 @@ from sqlalchemy.orm import joinedload, selectinload
from app.core.database import get_session
from app.models.encounter import Encounter
from app.models.evolution import Evolution
from app.models.genlocke_transfer import GenlockeTransfer
from app.models.genlocke import GenlockeLeg
from app.models.genlocke_transfer import GenlockeTransfer
from app.models.nuzlocke_run import NuzlockeRun
from app.models.pokemon import Pokemon
from app.models.route import Route
@@ -60,7 +60,11 @@ async def create_encounter(
# Shiny clause: shiny encounters bypass the route-lock check
shiny_clause_on = run.rules.get("shinyClause", True) if run.rules else True
skip_route_lock = (data.is_shiny and shiny_clause_on) or data.origin in ("shed_evolution", "egg", "transfer")
skip_route_lock = (data.is_shiny and shiny_clause_on) or data.origin in (
"shed_evolution",
"egg",
"transfer",
)
# If this route has a parent, check if sibling already has an encounter
if route.parent_route_id is not None and not skip_route_lock:
@@ -78,7 +82,8 @@ async def create_encounter(
# Zone-aware: only check siblings in the same zone (null treated as 0)
my_zone = route.pinwheel_zone if route.pinwheel_zone is not None else 0
sibling_ids = [
s.id for s in siblings
s.id
for s in siblings
if (s.pinwheel_zone if s.pinwheel_zone is not None else 0) == my_zone
]
else:
@@ -89,8 +94,7 @@ async def create_encounter(
# Exclude transfer-target encounters so they don't block the starter
transfer_target_ids = select(GenlockeTransfer.target_encounter_id)
existing_encounter = await session.execute(
select(Encounter)
.where(
select(Encounter).where(
Encounter.run_id == run_id,
Encounter.route_id.in_(sibling_ids),
~Encounter.id.in_(transfer_target_ids),
@@ -197,6 +201,7 @@ async def bulk_randomize_encounters(
# 2. Get version_group_id from game
from app.models.game import Game
game = await session.get(Game, game_id)
if game is None or game.version_group_id is None:
raise HTTPException(status_code=400, detail="Game has no version group")
@@ -257,8 +262,7 @@ async def bulk_randomize_encounters(
leg = leg_result.scalar_one_or_none()
if leg:
genlocke_result = await session.execute(
select(GenlockeLeg.retired_pokemon_ids)
.where(
select(GenlockeLeg.retired_pokemon_ids).where(
GenlockeLeg.genlocke_id == leg.genlocke_id,
GenlockeLeg.leg_order < leg.leg_order,
GenlockeLeg.retired_pokemon_ids.isnot(None),
@@ -268,7 +272,6 @@ async def bulk_randomize_encounters(
duped.update(retired_ids)
# 8. Organize routes: identify top-level and children
routes_by_id = {r.id: r for r in all_routes}
top_level = [r for r in all_routes if r.parent_route_id is None]
children_by_parent: dict[int, list[Route]] = {}
for r in all_routes:
@@ -289,7 +292,11 @@ async def bulk_randomize_encounters(
if parent_route.id in encountered_route_ids:
continue
available = route_pokemon.get(parent_route.id, [])
eligible = [p for p in available if p not in duped] if dupes_clause_on else available
eligible = (
[p for p in available if p not in duped]
if dupes_clause_on
else available
)
if not eligible:
skipped += 1
continue
@@ -335,7 +342,11 @@ async def bulk_randomize_encounters(
if p not in zone_pokemon:
zone_pokemon.append(p)
eligible = [p for p in zone_pokemon if p not in duped] if dupes_clause_on else zone_pokemon
eligible = (
[p for p in zone_pokemon if p not in duped]
if dupes_clause_on
else zone_pokemon
)
if not eligible:
skipped += 1
continue
@@ -371,7 +382,11 @@ async def bulk_randomize_encounters(
if p not in group_pokemon:
group_pokemon.append(p)
eligible = [p for p in group_pokemon if p not in duped] if dupes_clause_on else group_pokemon
eligible = (
[p for p in group_pokemon if p not in duped]
if dupes_clause_on
else group_pokemon
)
if not eligible:
skipped += 1
continue

View File

@@ -26,17 +26,18 @@ async def list_evolutions(
offset: int = Query(0, ge=0),
session: AsyncSession = Depends(get_session),
):
base_query = (
select(Evolution)
.options(joinedload(Evolution.from_pokemon), joinedload(Evolution.to_pokemon))
base_query = select(Evolution).options(
joinedload(Evolution.from_pokemon), joinedload(Evolution.to_pokemon)
)
if search:
search_lower = search.lower()
# Join pokemon to search by name
from_pokemon = select(Pokemon.id).where(
func.lower(Pokemon.name).contains(search_lower)
).scalar_subquery()
from_pokemon = (
select(Pokemon.id)
.where(func.lower(Pokemon.name).contains(search_lower))
.scalar_subquery()
)
base_query = base_query.where(
or_(
Evolution.from_pokemon_id.in_(from_pokemon),
@@ -52,9 +53,11 @@ async def list_evolutions(
count_base = select(Evolution)
if search:
search_lower = search.lower()
from_pokemon = select(Pokemon.id).where(
func.lower(Pokemon.name).contains(search_lower)
).scalar_subquery()
from_pokemon = (
select(Pokemon.id)
.where(func.lower(Pokemon.name).contains(search_lower))
.scalar_subquery()
)
count_base = count_base.where(
or_(
Evolution.from_pokemon_id.in_(from_pokemon),
@@ -68,7 +71,11 @@ async def list_evolutions(
count_query = select(func.count()).select_from(count_base.subquery())
total = (await session.execute(count_query)).scalar() or 0
items_query = base_query.order_by(Evolution.from_pokemon_id, Evolution.to_pokemon_id).offset(offset).limit(limit)
items_query = (
base_query.order_by(Evolution.from_pokemon_id, Evolution.to_pokemon_id)
.offset(offset)
.limit(limit)
)
result = await session.execute(items_query)
items = result.scalars().unique().all()
@@ -209,7 +216,9 @@ async def bulk_import_evolutions(
session.add(evolution)
created += 1
except Exception as e:
errors.append(f"Evolution {item.from_pokeapi_id} -> {item.to_pokeapi_id}: {e}")
errors.append(
f"Evolution {item.from_pokeapi_id} -> {item.to_pokeapi_id}: {e}"
)
await session.commit()
return BulkImportResult(created=created, updated=updated, errors=errors)

View File

@@ -20,9 +20,7 @@ router = APIRouter()
@router.get("/games")
async def export_games(session: AsyncSession = Depends(get_session)):
"""Export all games in seed JSON format."""
result = await session.execute(
select(Game).order_by(Game.name)
)
result = await session.execute(select(Game).order_by(Game.name))
games = result.scalars().all()
return [
{
@@ -154,7 +152,11 @@ async def export_game_bosses(
"pokemon_name": bp.pokemon.name,
"level": bp.level,
"order": bp.order,
**({"condition_label": bp.condition_label} if bp.condition_label else {}),
**(
{"condition_label": bp.condition_label}
if bp.condition_label
else {}
),
}
for bp in sorted(b.pokemon, key=lambda p: p.order)
],
@@ -167,9 +169,7 @@ async def export_game_bosses(
@router.get("/pokemon")
async def export_pokemon(session: AsyncSession = Depends(get_session)):
"""Export all pokemon in seed JSON format."""
result = await session.execute(
select(Pokemon).order_by(Pokemon.pokeapi_id)
)
result = await session.execute(select(Pokemon).order_by(Pokemon.pokeapi_id))
pokemon_list = result.scalars().all()
return [
{

View File

@@ -40,7 +40,9 @@ async def _get_game_or_404(session: AsyncSession, game_id: int) -> Game:
async def _get_version_group_id(session: AsyncSession, game_id: int) -> int:
game = await _get_game_or_404(session, game_id)
if game.version_group_id is None:
raise HTTPException(status_code=400, detail="Game has no version group assigned")
raise HTTPException(
status_code=400, detail="Game has no version group assigned"
)
return game.version_group_id
@@ -68,16 +70,18 @@ async def list_games_by_region(session: AsyncSession = Depends(get_session)):
for region in regions_data:
region_games = games_by_region.get(region["name"], [])
defaults = region["genlocke_defaults"]
response.append({
"name": region["name"],
"generation": region["generation"],
"order": region["order"],
"genlocke_defaults": {
"true_genlocke": defaults["true"],
"normal_genlocke": defaults["normal"],
},
"games": region_games,
})
response.append(
{
"name": region["name"],
"generation": region["generation"],
"order": region["order"],
"genlocke_defaults": {
"true_genlocke": defaults["true"],
"normal_genlocke": defaults["normal"],
},
"games": region_games,
}
)
return response
@@ -89,9 +93,7 @@ async def get_game(game_id: int, session: AsyncSession = Depends(get_session)):
# Load routes via version_group_id
result = await session.execute(
select(Route)
.where(Route.version_group_id == vg_id)
.order_by(Route.order)
select(Route).where(Route.version_group_id == vg_id).order_by(Route.order)
)
routes = result.scalars().all()
@@ -149,10 +151,13 @@ async def list_game_routes(
def route_to_dict(route: Route) -> dict:
# Only show encounter methods for the requested game
methods = sorted({
re.encounter_method for re in route.route_encounters
if re.game_id == game_id
})
methods = sorted(
{
re.encounter_method
for re in route.route_encounters
if re.game_id == game_id
}
)
return {
"id": route.id,
"name": route.name,
@@ -193,14 +198,12 @@ async def list_game_routes(
@router.post("", response_model=GameResponse, status_code=201)
async def create_game(
data: GameCreate, session: AsyncSession = Depends(get_session)
):
existing = await session.execute(
select(Game).where(Game.slug == data.slug)
)
async def create_game(data: GameCreate, session: AsyncSession = Depends(get_session)):
existing = await session.execute(select(Game).where(Game.slug == data.slug))
if existing.scalar_one_or_none() is not None:
raise HTTPException(status_code=409, detail="Game with this slug already exists")
raise HTTPException(
status_code=409, detail="Game with this slug already exists"
)
game = Game(**data.model_dump())
session.add(game)
@@ -223,7 +226,9 @@ async def update_game(
select(Game).where(Game.slug == update_data["slug"], Game.id != game_id)
)
if existing.scalar_one_or_none() is not None:
raise HTTPException(status_code=409, detail="Game with this slug already exists")
raise HTTPException(
status_code=409, detail="Game with this slug already exists"
)
for field, value in update_data.items():
setattr(game, field, value)
@@ -234,9 +239,7 @@ async def update_game(
@router.delete("/{game_id}", status_code=204)
async def delete_game(
game_id: int, session: AsyncSession = Depends(get_session)
):
async def delete_game(game_id: int, session: AsyncSession = Depends(get_session)):
result = await session.execute(
select(Game).where(Game.id == game_id).options(selectinload(Game.runs))
)
@@ -393,7 +396,9 @@ async def bulk_import_routes(
try:
route_name_to_id = await upsert_routes(session, vg_id, routes_data)
except Exception as e:
raise HTTPException(status_code=400, detail=f"Failed to import routes: {e}")
raise HTTPException(
status_code=400, detail=f"Failed to import routes: {e}"
) from e
# Upsert encounters for each route
encounter_count = 0
@@ -406,8 +411,11 @@ async def bulk_import_routes(
if item.encounters:
try:
count = await upsert_route_encounters(
session, route_id, [e.model_dump() for e in item.encounters],
dex_to_id, game_id,
session,
route_id,
[e.model_dump() for e in item.encounters],
dex_to_id,
game_id,
)
encounter_count += count
except Exception as e:
@@ -422,8 +430,11 @@ async def bulk_import_routes(
if child.encounters:
try:
count = await upsert_route_encounters(
session, child_id, [e.model_dump() for e in child.encounters],
dex_to_id, game_id,
session,
child_id,
[e.model_dump() for e in child.encounters],
dex_to_id,
game_id,
)
encounter_count += count
except Exception as e:

View File

@@ -1,6 +1,8 @@
from fastapi import APIRouter, Depends, HTTPException
from pydantic import BaseModel
from sqlalchemy import delete as sa_delete, func, select, update as sa_update
from sqlalchemy import delete as sa_delete
from sqlalchemy import func, select
from sqlalchemy import update as sa_update
from sqlalchemy.ext.asyncio import AsyncSession
from sqlalchemy.orm import selectinload
@@ -9,9 +11,9 @@ from app.models.encounter import Encounter
from app.models.evolution import Evolution
from app.models.game import Game
from app.models.genlocke import Genlocke, GenlockeLeg
from app.models.genlocke_transfer import GenlockeTransfer
from app.models.nuzlocke_run import NuzlockeRun
from app.models.pokemon import Pokemon
from app.models.genlocke_transfer import GenlockeTransfer
from app.models.route import Route
from app.schemas.genlocke import (
AddLegRequest,
@@ -74,9 +76,7 @@ async def list_genlockes(session: AsyncSession = Depends(get_session)):
@router.get("/{genlocke_id}", response_model=GenlockeDetailResponse)
async def get_genlocke(
genlocke_id: int, session: AsyncSession = Depends(get_session)
):
async def get_genlocke(genlocke_id: int, session: AsyncSession = Depends(get_session)):
result = await session.execute(
select(Genlocke)
.where(Genlocke.id == genlocke_id)
@@ -112,7 +112,9 @@ async def get_genlocke(
legs_completed = 0
for leg in genlocke.legs:
run_status = leg.run.status if leg.run else None
enc_count, death_count = stats_by_run.get(leg.run_id, (0, 0)) if leg.run_id else (0, 0)
enc_count, death_count = (
stats_by_run.get(leg.run_id, (0, 0)) if leg.run_id else (0, 0)
)
total_encounters += enc_count
total_deaths += death_count
if run_status == "completed":
@@ -254,7 +256,9 @@ async def get_genlocke_graveyard(
)
)
deadliest = max(deaths_per_leg, key=lambda s: s.death_count) if deaths_per_leg else None
deadliest = (
max(deaths_per_leg, key=lambda s: s.death_count) if deaths_per_leg else None
)
return GenlockeGraveyardResponse(
entries=entries,
@@ -285,9 +289,7 @@ async def get_genlocke_lineages(
# Query all transfers for this genlocke
transfer_result = await session.execute(
select(GenlockeTransfer).where(
GenlockeTransfer.genlocke_id == genlocke_id
)
select(GenlockeTransfer).where(GenlockeTransfer.genlocke_id == genlocke_id)
)
transfers = transfer_result.scalars().all()
@@ -302,7 +304,11 @@ async def get_genlocke_lineages(
backward.add(t.target_encounter_id)
# Find roots: sources that are NOT targets
roots = [t.source_encounter_id for t in transfers if t.source_encounter_id not in backward]
roots = [
t.source_encounter_id
for t in transfers
if t.source_encounter_id not in backward
]
# Deduplicate while preserving order
seen_roots: set[int] = set()
unique_roots: list[int] = []
@@ -421,7 +427,7 @@ async def get_genlocke_lineages(
)
# Sort by first leg order, then by encounter ID
lineages.sort(key=lambda l: (l.legs[0].leg_order, l.legs[0].encounter_id))
lineages.sort(key=lambda lin: (lin.legs[0].leg_order, lin.legs[0].encounter_id))
return GenlockeLineageResponse(
lineages=lineages,
@@ -440,15 +446,11 @@ async def create_genlocke(
raise HTTPException(status_code=400, detail="Name is required")
# Validate all game_ids exist
result = await session.execute(
select(Game).where(Game.id.in_(data.game_ids))
)
result = await session.execute(select(Game).where(Game.id.in_(data.game_ids)))
found_games = {g.id: g for g in result.scalars().all()}
missing = [gid for gid in data.game_ids if gid not in found_games]
if missing:
raise HTTPException(
status_code=404, detail=f"Games not found: {missing}"
)
raise HTTPException(status_code=404, detail=f"Games not found: {missing}")
# Create genlocke
genlocke = Genlocke(
@@ -578,9 +580,7 @@ async def advance_leg(
raise HTTPException(status_code=404, detail="Genlocke not found")
if genlocke.status != "active":
raise HTTPException(
status_code=400, detail="Genlocke is not active"
)
raise HTTPException(status_code=400, detail="Genlocke is not active")
# Find the current leg
current_leg = None
@@ -596,9 +596,7 @@ async def advance_leg(
# Verify current leg's run is completed
if current_leg.run_id is None:
raise HTTPException(
status_code=400, detail="Current leg has no run"
)
raise HTTPException(status_code=400, detail="Current leg has no run")
current_run = await session.get(NuzlockeRun, current_leg.run_id)
if current_run is None or current_run.status != "completed":
raise HTTPException(
@@ -606,14 +604,10 @@ async def advance_leg(
)
if next_leg is None:
raise HTTPException(
status_code=400, detail="No next leg to advance to"
)
raise HTTPException(status_code=400, detail="No next leg to advance to")
if next_leg.run_id is not None:
raise HTTPException(
status_code=400, detail="Next leg already has a run"
)
raise HTTPException(status_code=400, detail="Next leg already has a run")
# Compute retired Pokemon families if retireHoF is enabled
if genlocke.genlocke_rules.get("retireHoF", False):
@@ -807,10 +801,12 @@ async def get_retired_families(
for leg in legs:
ids = leg.retired_pokemon_ids or []
cumulative.update(ids)
by_leg.append(RetiredLegResponse(
leg_order=leg.leg_order,
retired_pokemon_ids=ids,
))
by_leg.append(
RetiredLegResponse(
leg_order=leg.leg_order,
retired_pokemon_ids=ids,
)
)
return RetiredFamiliesResponse(
retired_pokemon_ids=sorted(cumulative),
@@ -837,12 +833,15 @@ async def update_genlocke(
update_data = data.model_dump(exclude_unset=True)
if "status" in update_data:
if update_data["status"] not in ("active", "completed", "failed"):
raise HTTPException(
status_code=400,
detail="Status must be one of: active, completed, failed",
)
if "status" in update_data and update_data["status"] not in (
"active",
"completed",
"failed",
):
raise HTTPException(
status_code=400,
detail="Status must be one of: active, completed, failed",
)
for field, value in update_data.items():
setattr(genlocke, field, value)
@@ -871,8 +870,7 @@ async def delete_genlocke(
# Delete legs explicitly to avoid ORM cascade issues
# (genlocke_id is non-nullable, so SQLAlchemy can't nullify it)
await session.execute(
sa_delete(GenlockeLeg)
.where(GenlockeLeg.genlocke_id == genlocke_id)
sa_delete(GenlockeLeg).where(GenlockeLeg.genlocke_id == genlocke_id)
)
await session.delete(genlocke)

View File

@@ -8,7 +8,6 @@ from app.models.evolution import Evolution
from app.models.pokemon import Pokemon
from app.models.route import Route
from app.models.route_encounter import RouteEncounter
from app.models.game import Game
from app.schemas.pokemon import (
BulkImportItem,
BulkImportResult,
@@ -40,9 +39,7 @@ async def list_pokemon(
# Build base query with optional search filter
base_query = select(Pokemon)
if search:
base_query = base_query.where(
func.lower(Pokemon.name).contains(search.lower())
)
base_query = base_query.where(func.lower(Pokemon.name).contains(search.lower()))
if type:
base_query = base_query.where(Pokemon.types.any(type))
@@ -51,7 +48,11 @@ async def list_pokemon(
total = (await session.execute(count_query)).scalar() or 0
# Get paginated items
items_query = base_query.order_by(Pokemon.national_dex, Pokemon.name).offset(offset).limit(limit)
items_query = (
base_query.order_by(Pokemon.national_dex, Pokemon.name)
.offset(offset)
.limit(limit)
)
result = await session.execute(items_query)
items = result.scalars().all()
@@ -156,9 +157,7 @@ async def get_pokemon_families(
@router.get("/pokemon/{pokemon_id}", response_model=PokemonResponse)
async def get_pokemon(
pokemon_id: int, session: AsyncSession = Depends(get_session)
):
async def get_pokemon(pokemon_id: int, session: AsyncSession = Depends(get_session)):
pokemon = await session.get(Pokemon, pokemon_id)
if pokemon is None:
raise HTTPException(status_code=404, detail="Pokemon not found")
@@ -258,7 +257,8 @@ async def get_pokemon_evolution_chain(
# Filter evolutions to only those in the family
family_evo_ids = [
evo.id for evo in evolutions
evo.id
for evo in evolutions
if evo.from_pokemon_id in family and evo.to_pokemon_id in family
]
@@ -294,9 +294,7 @@ async def get_pokemon_evolutions(
.options(joinedload(Evolution.to_pokemon))
)
if region is not None:
query = query.where(
or_(Evolution.region.is_(None), Evolution.region == region)
)
query = query.where(or_(Evolution.region.is_(None), Evolution.region == region))
result = await session.execute(query)
evolutions = result.scalars().unique().all()
@@ -309,7 +307,8 @@ async def get_pokemon_evolutions(
}
if regional_keys:
evolutions = [
e for e in evolutions
e
for e in evolutions
if e.region is not None or (e.trigger, e.item) not in regional_keys
]
@@ -349,9 +348,7 @@ async def update_pokemon(
@router.delete("/pokemon/{pokemon_id}", status_code=204)
async def delete_pokemon(
pokemon_id: int, session: AsyncSession = Depends(get_session)
):
async def delete_pokemon(pokemon_id: int, session: AsyncSession = Depends(get_session)):
result = await session.execute(
select(Pokemon)
.where(Pokemon.id == pokemon_id)

View File

@@ -1,6 +1,17 @@
from fastapi import APIRouter
from app.api import bosses, encounters, evolutions, export, games, genlockes, health, pokemon, runs, stats
from app.api import (
bosses,
encounters,
evolutions,
export,
games,
genlockes,
health,
pokemon,
runs,
stats,
)
api_router = APIRouter()
api_router.include_router(health.router)

View File

@@ -1,4 +1,4 @@
from datetime import datetime, timezone
from datetime import UTC, datetime
from fastapi import APIRouter, Depends, HTTPException, Response
from sqlalchemy import func, select
@@ -9,18 +9,22 @@ from app.core.database import get_session
from app.models.boss_result import BossResult
from app.models.encounter import Encounter
from app.models.game import Game
from app.models.genlocke import Genlocke, GenlockeLeg
from app.models.genlocke import GenlockeLeg
from app.models.genlocke_transfer import GenlockeTransfer
from app.models.nuzlocke_run import NuzlockeRun
from app.schemas.run import RunCreate, RunDetailResponse, RunGenlockeContext, RunResponse, RunUpdate
from app.schemas.run import (
RunCreate,
RunDetailResponse,
RunGenlockeContext,
RunResponse,
RunUpdate,
)
router = APIRouter()
@router.post("", response_model=RunResponse, status_code=201)
async def create_run(
data: RunCreate, session: AsyncSession = Depends(get_session)
):
async def create_run(data: RunCreate, session: AsyncSession = Depends(get_session)):
# Validate game exists
game = await session.get(Game, data.game_id)
if game is None:
@@ -53,12 +57,9 @@ async def get_run(run_id: int, session: AsyncSession = Depends(get_session)):
.where(NuzlockeRun.id == run_id)
.options(
joinedload(NuzlockeRun.game),
selectinload(NuzlockeRun.encounters)
.joinedload(Encounter.pokemon),
selectinload(NuzlockeRun.encounters)
.joinedload(Encounter.current_pokemon),
selectinload(NuzlockeRun.encounters)
.joinedload(Encounter.route),
selectinload(NuzlockeRun.encounters).joinedload(Encounter.pokemon),
selectinload(NuzlockeRun.encounters).joinedload(Encounter.current_pokemon),
selectinload(NuzlockeRun.encounters).joinedload(Encounter.route),
)
)
run = result.scalar_one_or_none()
@@ -134,7 +135,10 @@ async def update_run(
update_data = data.model_dump(exclude_unset=True)
# Validate hof_encounter_ids if provided
if "hof_encounter_ids" in update_data and update_data["hof_encounter_ids"] is not None:
if (
"hof_encounter_ids" in update_data
and update_data["hof_encounter_ids"] is not None
):
hof_ids = update_data["hof_encounter_ids"]
if len(hof_ids) > 6:
raise HTTPException(
@@ -156,7 +160,8 @@ async def update_run(
detail=f"Encounters not found in this run: {missing}",
)
not_alive = [
eid for eid, e in found.items()
eid
for eid, e in found.items()
if e.status != "caught" or e.faint_level is not None
]
if not_alive:
@@ -168,13 +173,15 @@ async def update_run(
# Auto-set completed_at when ending a run
if "status" in update_data and update_data["status"] in ("completed", "failed"):
if run.status != "active":
raise HTTPException(
status_code=400, detail="Only active runs can be ended"
)
update_data["completed_at"] = datetime.now(timezone.utc)
raise HTTPException(status_code=400, detail="Only active runs can be ended")
update_data["completed_at"] = datetime.now(UTC)
# Block reactivating a completed/failed run that belongs to a genlocke
if "status" in update_data and update_data["status"] == "active" and run.status != "active":
if (
"status" in update_data
and update_data["status"] == "active"
and run.status != "active"
):
leg_result = await session.execute(
select(GenlockeLeg).where(GenlockeLeg.run_id == run_id)
)
@@ -215,9 +222,7 @@ async def update_run(
@router.delete("/{run_id}", status_code=204)
async def delete_run(
run_id: int, session: AsyncSession = Depends(get_session)
):
async def delete_run(run_id: int, session: AsyncSession = Depends(get_session)):
run = await session.get(NuzlockeRun, run_id)
if run is None:
raise HTTPException(status_code=404, detail="Run not found")

View File

@@ -84,8 +84,12 @@ async def get_stats(session: AsyncSession = Depends(get_session)):
fainted_count = enc.fainted
missed_count = enc.missed
catch_rate = round(caught_count / total_encounters, 4) if total_encounters > 0 else None
avg_encounters_per_run = round(total_encounters / total_runs, 1) if total_runs > 0 else None
catch_rate = (
round(caught_count / total_encounters, 4) if total_encounters > 0 else None
)
avg_encounters_per_run = (
round(total_encounters / total_runs, 1) if total_runs > 0 else None
)
# --- Top caught pokemon (top 10) ---
top_caught_q = await session.execute(
@@ -102,7 +106,9 @@ async def get_stats(session: AsyncSession = Depends(get_session)):
.limit(10)
)
top_caught_pokemon = [
PokemonRanking(pokemon_id=r.id, name=r.name, sprite_url=r.sprite_url, count=r.count)
PokemonRanking(
pokemon_id=r.id, name=r.name, sprite_url=r.sprite_url, count=r.count
)
for r in top_caught_q.all()
]
@@ -120,7 +126,9 @@ async def get_stats(session: AsyncSession = Depends(get_session)):
.limit(10)
)
top_encountered_pokemon = [
PokemonRanking(pokemon_id=r.id, name=r.name, sprite_url=r.sprite_url, count=r.count)
PokemonRanking(
pokemon_id=r.id, name=r.name, sprite_url=r.sprite_url, count=r.count
)
for r in top_enc_q.all()
]
@@ -149,8 +157,7 @@ async def get_stats(session: AsyncSession = Depends(get_session)):
.limit(5)
)
top_death_causes = [
DeathCause(cause=r.death_cause, count=r.count)
for r in death_causes_q.all()
DeathCause(cause=r.death_cause, count=r.count) for r in death_causes_q.all()
]
# Average levels
@@ -179,8 +186,7 @@ async def get_stats(session: AsyncSession = Depends(get_session)):
.order_by(func.count().desc())
)
type_distribution = [
TypeCount(type=r.type_name, count=r.count)
for r in type_q.all()
TypeCount(type=r.type_name, count=r.count) for r in type_q.all()
]
return StatsResponse(

View File

@@ -7,7 +7,9 @@ from app.core.database import Base
class BossBattle(Base):
__tablename__ = "boss_battles"
__table_args__ = (
UniqueConstraint("version_group_id", "order", name="uq_boss_battles_version_group_order"),
UniqueConstraint(
"version_group_id", "order", name="uq_boss_battles_version_group_order"
),
)
id: Mapped[int] = mapped_column(primary_key=True)
@@ -15,8 +17,12 @@ class BossBattle(Base):
ForeignKey("version_groups.id"), index=True
)
name: Mapped[str] = mapped_column(String(100))
boss_type: Mapped[str] = mapped_column(String(20)) # gym_leader, elite_four, champion, rival, evil_team, other
specialty_type: Mapped[str | None] = mapped_column(String(20), default=None) # pokemon type specialty (e.g. rock, water)
boss_type: Mapped[str] = mapped_column(
String(20)
) # gym_leader, elite_four, champion, rival, evil_team, other
specialty_type: Mapped[str | None] = mapped_column(
String(20), default=None
) # pokemon type specialty (e.g. rock, water)
badge_name: Mapped[str | None] = mapped_column(String(100))
badge_image_url: Mapped[str | None] = mapped_column(String(500))
level_cap: Mapped[int] = mapped_column(SmallInteger)
@@ -28,13 +34,13 @@ class BossBattle(Base):
section: Mapped[str | None] = mapped_column(String(100), default=None)
sprite_url: Mapped[str | None] = mapped_column(String(500))
version_group: Mapped["VersionGroup"] = relationship(
back_populates="boss_battles"
)
version_group: Mapped["VersionGroup"] = relationship(back_populates="boss_battles")
after_route: Mapped["Route | None"] = relationship()
pokemon: Mapped[list["BossPokemon"]] = relationship(
back_populates="boss_battle", cascade="all, delete-orphan"
)
def __repr__(self) -> str:
return f"<BossBattle(id={self.id}, name='{self.name}', type='{self.boss_type}')>"
return (
f"<BossBattle(id={self.id}, name='{self.name}', type='{self.boss_type}')>"
)

View File

@@ -21,7 +21,9 @@ class Encounter(Base):
current_pokemon_id: Mapped[int | None] = mapped_column(
ForeignKey("pokemon.id"), index=True
)
is_shiny: Mapped[bool] = mapped_column(Boolean, default=False, server_default=text("false"))
is_shiny: Mapped[bool] = mapped_column(
Boolean, default=False, server_default=text("false")
)
caught_at: Mapped[datetime] = mapped_column(
DateTime(timezone=True), server_default=func.now()
)

View File

@@ -14,7 +14,9 @@ class Evolution(Base):
min_level: Mapped[int | None] = mapped_column(SmallInteger)
item: Mapped[str | None] = mapped_column(String(50)) # e.g. thunder-stone
held_item: Mapped[str | None] = mapped_column(String(50))
condition: Mapped[str | None] = mapped_column(String(200)) # catch-all for other conditions
condition: Mapped[str | None] = mapped_column(
String(200)
) # catch-all for other conditions
region: Mapped[str | None] = mapped_column(String(30))
from_pokemon: Mapped["Pokemon"] = relationship(foreign_keys=[from_pokemon_id])

View File

@@ -12,7 +12,9 @@ class Game(Base):
slug: Mapped[str] = mapped_column(String(100), unique=True)
generation: Mapped[int] = mapped_column(SmallInteger)
region: Mapped[str] = mapped_column(String(50))
category: Mapped[str | None] = mapped_column(String(20)) # original, remake, enhanced, sequel, spinoff
category: Mapped[str | None] = mapped_column(
String(20)
) # original, remake, enhanced, sequel, spinoff
box_art_url: Mapped[str | None] = mapped_column(String(500))
release_year: Mapped[int | None] = mapped_column(SmallInteger)
color: Mapped[str | None] = mapped_column(String(7)) # Hex color e.g. #FF0000
@@ -20,9 +22,7 @@ class Game(Base):
ForeignKey("version_groups.id"), index=True
)
version_group: Mapped["VersionGroup | None"] = relationship(
back_populates="games"
)
version_group: Mapped["VersionGroup | None"] = relationship(back_populates="games")
runs: Mapped[list["NuzlockeRun"]] = relationship(back_populates="game")
def __repr__(self) -> str:

View File

@@ -13,7 +13,9 @@ class Genlocke(Base):
id: Mapped[int] = mapped_column(primary_key=True)
name: Mapped[str] = mapped_column(String(100))
status: Mapped[str] = mapped_column(String(20), index=True) # active, completed, failed
status: Mapped[str] = mapped_column(
String(20), index=True
) # active, completed, failed
genlocke_rules: Mapped[dict] = mapped_column(JSONB, default=dict)
nuzlocke_rules: Mapped[dict] = mapped_column(JSONB, default=dict)
created_at: Mapped[datetime] = mapped_column(

View File

@@ -13,7 +13,9 @@ class NuzlockeRun(Base):
id: Mapped[int] = mapped_column(primary_key=True)
game_id: Mapped[int] = mapped_column(ForeignKey("games.id"), index=True)
name: Mapped[str] = mapped_column(String(100))
status: Mapped[str] = mapped_column(String(20), index=True) # active, completed, failed
status: Mapped[str] = mapped_column(
String(20), index=True
) # active, completed, failed
rules: Mapped[dict] = mapped_column(JSONB, default=dict)
started_at: Mapped[datetime] = mapped_column(
DateTime(timezone=True), server_default=func.now()
@@ -26,4 +28,6 @@ class NuzlockeRun(Base):
boss_results: Mapped[list["BossResult"]] = relationship(back_populates="run")
def __repr__(self) -> str:
return f"<NuzlockeRun(id={self.id}, name='{self.name}', status='{self.status}')>"
return (
f"<NuzlockeRun(id={self.id}, name='{self.name}', status='{self.status}')>"
)

View File

@@ -7,7 +7,9 @@ from app.core.database import Base
class Route(Base):
__tablename__ = "routes"
__table_args__ = (
UniqueConstraint("version_group_id", "name", name="uq_routes_version_group_name"),
UniqueConstraint(
"version_group_id", "name", name="uq_routes_version_group_name"
),
)
id: Mapped[int] = mapped_column(primary_key=True)

View File

@@ -8,8 +8,11 @@ class RouteEncounter(Base):
__tablename__ = "route_encounters"
__table_args__ = (
UniqueConstraint(
"route_id", "pokemon_id", "encounter_method", "game_id",
name="uq_route_pokemon_method_game"
"route_id",
"pokemon_id",
"encounter_method",
"game_id",
name="uq_route_pokemon_method_game",
),
)

View File

@@ -14,7 +14,6 @@ from app.schemas.encounter import (
EncounterResponse,
EncounterUpdate,
)
from app.schemas.genlocke import GenlockeCreate, GenlockeResponse, GenlockeLegResponse
from app.schemas.game import (
GameCreate,
GameDetailResponse,
@@ -25,6 +24,7 @@ from app.schemas.game import (
RouteResponse,
RouteUpdate,
)
from app.schemas.genlocke import GenlockeCreate, GenlockeLegResponse, GenlockeResponse
from app.schemas.pokemon import (
BulkImportItem,
BulkImportResult,
@@ -37,7 +37,13 @@ from app.schemas.pokemon import (
RouteEncounterResponse,
RouteEncounterUpdate,
)
from app.schemas.run import RunCreate, RunDetailResponse, RunGenlockeContext, RunResponse, RunUpdate
from app.schemas.run import (
RunCreate,
RunDetailResponse,
RunGenlockeContext,
RunResponse,
RunUpdate,
)
__all__ = [
"BossBattleCreate",

View File

@@ -6,7 +6,7 @@ Usage:
import asyncio
import random
from datetime import datetime, timedelta, timezone
from datetime import UTC, datetime, timedelta
from sqlalchemy import delete, select
from sqlalchemy.ext.asyncio import AsyncSession
@@ -16,18 +16,52 @@ from app.models.encounter import Encounter
from app.models.evolution import Evolution
from app.models.game import Game
from app.models.nuzlocke_run import NuzlockeRun
from app.models.pokemon import Pokemon
from app.models.route import Route
random.seed(42) # reproducible data
# --- Nicknames pool ---
NICKNAMES = [
"Blaze", "Thunder", "Shadow", "Luna", "Spike", "Rex", "Cinder", "Misty",
"Rocky", "Breeze", "Fang", "Nova", "Scout", "Atlas", "Pepper", "Storm",
"Bandit", "Echo", "Maple", "Titan", "Ziggy", "Bolt", "Rusty", "Pearl",
"Ivy", "Ghost", "Sunny", "Dash", "Ember", "Frost", "Jade", "Onyx",
"Willow", "Tank", "Pip", "Mochi", "Salem", "Patches", "Bean", "Rocket",
"Blaze",
"Thunder",
"Shadow",
"Luna",
"Spike",
"Rex",
"Cinder",
"Misty",
"Rocky",
"Breeze",
"Fang",
"Nova",
"Scout",
"Atlas",
"Pepper",
"Storm",
"Bandit",
"Echo",
"Maple",
"Titan",
"Ziggy",
"Bolt",
"Rusty",
"Pearl",
"Ivy",
"Ghost",
"Sunny",
"Dash",
"Ember",
"Frost",
"Jade",
"Onyx",
"Willow",
"Tank",
"Pip",
"Mochi",
"Salem",
"Patches",
"Bean",
"Rocket",
]
DEATH_CAUSES = [
@@ -129,20 +163,18 @@ async def get_leaf_routes(session: AsyncSession, game_id: int) -> list[Route]:
"""Get routes that can have encounters (no children)."""
# Get all routes for the game
result = await session.execute(
select(Route)
.where(Route.game_id == game_id)
.order_by(Route.order)
select(Route).where(Route.game_id == game_id).order_by(Route.order)
)
all_routes = result.scalars().all()
parent_ids = {r.parent_route_id for r in all_routes if r.parent_route_id is not None}
parent_ids = {
r.parent_route_id for r in all_routes if r.parent_route_id is not None
}
leaf_routes = [r for r in all_routes if r.id not in parent_ids]
return leaf_routes
async def get_encounterables(
session: AsyncSession, game_id: int
) -> list[int]:
async def get_encounterables(session: AsyncSession, game_id: int) -> list[int]:
"""Get pokemon IDs that appear in route encounters for this game."""
from app.models.route_encounter import RouteEncounter
@@ -157,16 +189,16 @@ async def get_encounterables(
async def get_evolution_map(session: AsyncSession) -> dict[int, list[int]]:
"""Return {from_pokemon_id: [to_pokemon_id, ...]} for all evolutions."""
result = await session.execute(select(Evolution.from_pokemon_id, Evolution.to_pokemon_id))
result = await session.execute(
select(Evolution.from_pokemon_id, Evolution.to_pokemon_id)
)
evo_map: dict[int, list[int]] = {}
for from_id, to_id in result:
evo_map.setdefault(from_id, []).append(to_id)
return evo_map
def pick_routes_for_run(
leaf_routes: list[Route], progress: float
) -> list[Route]:
def pick_routes_for_run(leaf_routes: list[Route], progress: float) -> list[Route]:
"""Pick a subset of leaf routes respecting one-per-group.
For routes with a parent, only one sibling per parent_route_id is chosen.
@@ -257,74 +289,73 @@ async def inject():
"""Clear existing runs and inject test data."""
print("Injecting test data...")
async with async_session() as session:
async with session.begin():
# Clear existing runs and encounters
await session.execute(delete(Encounter))
await session.execute(delete(NuzlockeRun))
print("Cleared existing runs and encounters")
async with async_session() as session, session.begin():
# Clear existing runs and encounters
await session.execute(delete(Encounter))
await session.execute(delete(NuzlockeRun))
print("Cleared existing runs and encounters")
evo_map = await get_evolution_map(session)
now = datetime.now(timezone.utc)
evo_map = await get_evolution_map(session)
now = datetime.now(UTC)
total_runs = 0
total_encounters = 0
total_runs = 0
total_encounters = 0
for run_def in RUN_DEFS:
game = await get_game_by_slug(session, run_def["game_slug"])
if game is None:
print(f" Warning: game '{run_def['game_slug']}' not found, skipping")
continue
for run_def in RUN_DEFS:
game = await get_game_by_slug(session, run_def["game_slug"])
if game is None:
print(f" Warning: game '{run_def['game_slug']}' not found, skipping")
continue
# Build rules
rules = {**DEFAULT_RULES, **run_def["rules"]}
# Build rules
rules = {**DEFAULT_RULES, **run_def["rules"]}
# Compute dates
started_at = now - timedelta(days=run_def["started_days_ago"])
completed_at = None
if run_def["ended_days_ago"] is not None:
completed_at = now - timedelta(days=run_def["ended_days_ago"])
# Compute dates
started_at = now - timedelta(days=run_def["started_days_ago"])
completed_at = None
if run_def["ended_days_ago"] is not None:
completed_at = now - timedelta(days=run_def["ended_days_ago"])
run = NuzlockeRun(
game_id=game.id,
name=run_def["name"],
status=run_def["status"],
rules=rules,
started_at=started_at,
completed_at=completed_at,
)
session.add(run)
await session.flush() # get run.id
run = NuzlockeRun(
game_id=game.id,
name=run_def["name"],
status=run_def["status"],
rules=rules,
started_at=started_at,
completed_at=completed_at,
)
session.add(run)
await session.flush() # get run.id
# Get routes and pokemon for this game
leaf_routes = await get_leaf_routes(session, game.id)
pokemon_ids = await get_encounterables(session, game.id)
if not leaf_routes or not pokemon_ids:
print(f" {run_def['name']}: no routes or pokemon, skipping encounters")
total_runs += 1
continue
chosen_routes = pick_routes_for_run(leaf_routes, run_def["progress"])
used_pokemon: set[int] = set()
run_encounters = 0
for i, route in enumerate(chosen_routes):
enc = generate_encounter(
run.id, route, pokemon_ids, evo_map, used_pokemon, i
)
session.add(enc)
run_encounters += 1
# Get routes and pokemon for this game
leaf_routes = await get_leaf_routes(session, game.id)
pokemon_ids = await get_encounterables(session, game.id)
if not leaf_routes or not pokemon_ids:
print(f" {run_def['name']}: no routes or pokemon, skipping encounters")
total_runs += 1
total_encounters += run_encounters
continue
print(
f" {run_def['name']} ({game.name}, {run_def['status']}): "
f"{run_encounters} encounters across {len(chosen_routes)} routes"
chosen_routes = pick_routes_for_run(leaf_routes, run_def["progress"])
used_pokemon: set[int] = set()
run_encounters = 0
for i, route in enumerate(chosen_routes):
enc = generate_encounter(
run.id, route, pokemon_ids, evo_map, used_pokemon, i
)
session.add(enc)
run_encounters += 1
print(f"\nCreated {total_runs} runs with {total_encounters} total encounters")
total_runs += 1
total_encounters += run_encounters
print(
f" {run_def['name']} ({game.name}, {run_def['status']}): "
f"{run_encounters} encounters across {len(chosen_routes)} routes"
)
print(f"\nCreated {total_runs} runs with {total_encounters} total encounters")
print("Test data injection complete!")

View File

@@ -21,15 +21,18 @@ async def upsert_version_groups(
"""Upsert version group records, return {slug: id} mapping."""
for vg_slug, vg_info in vg_data.items():
vg_name = " / ".join(
g["name"].replace("Pokemon ", "")
for g in vg_info["games"].values()
g["name"].replace("Pokemon ", "") for g in vg_info["games"].values()
)
stmt = insert(VersionGroup).values(
name=vg_name,
slug=vg_slug,
).on_conflict_do_update(
index_elements=["slug"],
set_={"name": vg_name},
stmt = (
insert(VersionGroup)
.values(
name=vg_name,
slug=vg_slug,
)
.on_conflict_do_update(
index_elements=["slug"],
set_={"name": vg_name},
)
)
await session.execute(stmt)
@@ -69,9 +72,13 @@ async def upsert_games(
values["version_group_id"] = vg_id
update_set["version_group_id"] = vg_id
stmt = insert(Game).values(**values).on_conflict_do_update(
index_elements=["slug"],
set_=update_set,
stmt = (
insert(Game)
.values(**values)
.on_conflict_do_update(
index_elements=["slug"],
set_=update_set,
)
)
await session.execute(stmt)
@@ -81,23 +88,29 @@ async def upsert_games(
return {row.slug: row.id for row in result}
async def upsert_pokemon(session: AsyncSession, pokemon_list: list[dict]) -> dict[int, int]:
async def upsert_pokemon(
session: AsyncSession, pokemon_list: list[dict]
) -> dict[int, int]:
"""Upsert pokemon records, return {pokeapi_id: id} mapping."""
for poke in pokemon_list:
stmt = insert(Pokemon).values(
pokeapi_id=poke["pokeapi_id"],
national_dex=poke["national_dex"],
name=poke["name"],
types=poke["types"],
sprite_url=poke.get("sprite_url"),
).on_conflict_do_update(
index_elements=["pokeapi_id"],
set_={
"national_dex": poke["national_dex"],
"name": poke["name"],
"types": poke["types"],
"sprite_url": poke.get("sprite_url"),
},
stmt = (
insert(Pokemon)
.values(
pokeapi_id=poke["pokeapi_id"],
national_dex=poke["national_dex"],
name=poke["name"],
types=poke["types"],
sprite_url=poke.get("sprite_url"),
)
.on_conflict_do_update(
index_elements=["pokeapi_id"],
set_={
"national_dex": poke["national_dex"],
"name": poke["name"],
"types": poke["types"],
"sprite_url": poke.get("sprite_url"),
},
)
)
await session.execute(stmt)
@@ -119,14 +132,18 @@ async def upsert_routes(
"""
# First pass: upsert all parent routes (without parent_route_id)
for route in routes:
stmt = insert(Route).values(
name=route["name"],
version_group_id=version_group_id,
order=route["order"],
parent_route_id=None, # Parent routes have no parent
).on_conflict_do_update(
constraint="uq_routes_version_group_name",
set_={"order": route["order"], "parent_route_id": None},
stmt = (
insert(Route)
.values(
name=route["name"],
version_group_id=version_group_id,
order=route["order"],
parent_route_id=None, # Parent routes have no parent
)
.on_conflict_do_update(
constraint="uq_routes_version_group_name",
set_={"order": route["order"], "parent_route_id": None},
)
)
await session.execute(stmt)
@@ -146,19 +163,23 @@ async def upsert_routes(
parent_id = name_to_id[route["name"]]
for child in children:
stmt = insert(Route).values(
name=child["name"],
version_group_id=version_group_id,
order=child["order"],
parent_route_id=parent_id,
pinwheel_zone=child.get("pinwheel_zone"),
).on_conflict_do_update(
constraint="uq_routes_version_group_name",
set_={
"order": child["order"],
"parent_route_id": parent_id,
"pinwheel_zone": child.get("pinwheel_zone"),
},
stmt = (
insert(Route)
.values(
name=child["name"],
version_group_id=version_group_id,
order=child["order"],
parent_route_id=parent_id,
pinwheel_zone=child.get("pinwheel_zone"),
)
.on_conflict_do_update(
constraint="uq_routes_version_group_name",
set_={
"order": child["order"],
"parent_route_id": parent_id,
"pinwheel_zone": child.get("pinwheel_zone"),
},
)
)
await session.execute(stmt)
@@ -186,21 +207,25 @@ async def upsert_route_encounters(
print(f" Warning: no pokemon_id for pokeapi_id {enc['pokeapi_id']}")
continue
stmt = insert(RouteEncounter).values(
route_id=route_id,
pokemon_id=pokemon_id,
game_id=game_id,
encounter_method=enc["method"],
encounter_rate=enc["encounter_rate"],
min_level=enc["min_level"],
max_level=enc["max_level"],
).on_conflict_do_update(
constraint="uq_route_pokemon_method_game",
set_={
"encounter_rate": enc["encounter_rate"],
"min_level": enc["min_level"],
"max_level": enc["max_level"],
},
stmt = (
insert(RouteEncounter)
.values(
route_id=route_id,
pokemon_id=pokemon_id,
game_id=game_id,
encounter_method=enc["method"],
encounter_rate=enc["encounter_rate"],
min_level=enc["min_level"],
max_level=enc["max_level"],
)
.on_conflict_do_update(
constraint="uq_route_pokemon_method_game",
set_={
"encounter_rate": enc["encounter_rate"],
"min_level": enc["min_level"],
"max_level": enc["max_level"],
},
)
)
await session.execute(stmt)
count += 1
@@ -224,37 +249,44 @@ async def upsert_bosses(
if after_route_name and route_name_to_id:
after_route_id = route_name_to_id.get(after_route_name)
if after_route_id is None:
print(f" Warning: route '{after_route_name}' not found for boss '{boss['name']}'")
print(
f" Warning: route '{after_route_name}' not found for boss '{boss['name']}'"
)
# Upsert the boss battle on (version_group_id, order) conflict
stmt = insert(BossBattle).values(
version_group_id=version_group_id,
name=boss["name"],
boss_type=boss["boss_type"],
specialty_type=boss.get("specialty_type"),
badge_name=boss.get("badge_name"),
badge_image_url=boss.get("badge_image_url"),
level_cap=boss["level_cap"],
order=boss["order"],
after_route_id=after_route_id,
location=boss["location"],
section=boss.get("section"),
sprite_url=boss.get("sprite_url"),
).on_conflict_do_update(
constraint="uq_boss_battles_version_group_order",
set_={
"name": boss["name"],
"boss_type": boss["boss_type"],
"specialty_type": boss.get("specialty_type"),
"badge_name": boss.get("badge_name"),
"badge_image_url": boss.get("badge_image_url"),
"level_cap": boss["level_cap"],
"after_route_id": after_route_id,
"location": boss["location"],
"section": boss.get("section"),
"sprite_url": boss.get("sprite_url"),
},
).returning(BossBattle.id)
stmt = (
insert(BossBattle)
.values(
version_group_id=version_group_id,
name=boss["name"],
boss_type=boss["boss_type"],
specialty_type=boss.get("specialty_type"),
badge_name=boss.get("badge_name"),
badge_image_url=boss.get("badge_image_url"),
level_cap=boss["level_cap"],
order=boss["order"],
after_route_id=after_route_id,
location=boss["location"],
section=boss.get("section"),
sprite_url=boss.get("sprite_url"),
)
.on_conflict_do_update(
constraint="uq_boss_battles_version_group_order",
set_={
"name": boss["name"],
"boss_type": boss["boss_type"],
"specialty_type": boss.get("specialty_type"),
"badge_name": boss.get("badge_name"),
"badge_image_url": boss.get("badge_image_url"),
"level_cap": boss["level_cap"],
"after_route_id": after_route_id,
"location": boss["location"],
"section": boss.get("section"),
"sprite_url": boss.get("sprite_url"),
},
)
.returning(BossBattle.id)
)
result = await session.execute(stmt)
boss_id = result.scalar_one()
@@ -267,13 +299,15 @@ async def upsert_bosses(
if pokemon_id is None:
print(f" Warning: no pokemon_id for pokeapi_id {bp['pokeapi_id']}")
continue
session.add(BossPokemon(
boss_battle_id=boss_id,
pokemon_id=pokemon_id,
level=bp["level"],
order=bp["order"],
condition_label=bp.get("condition_label"),
))
session.add(
BossPokemon(
boss_battle_id=boss_id,
pokemon_id=pokemon_id,
level=bp["level"],
order=bp["order"],
condition_label=bp.get("condition_label"),
)
)
count += 1

View File

@@ -42,130 +42,139 @@ async def seed():
"""Run the full seed process."""
print("Starting seed...")
async with async_session() as session:
async with session.begin():
# 1. Upsert version groups
with open(VG_JSON) as f:
vg_data = json.load(f)
vg_slug_to_id = await upsert_version_groups(session, vg_data)
print(f"Version Groups: {len(vg_slug_to_id)} upserted")
async with async_session() as session, session.begin():
# 1. Upsert version groups
with open(VG_JSON) as f:
vg_data = json.load(f)
vg_slug_to_id = await upsert_version_groups(session, vg_data)
print(f"Version Groups: {len(vg_slug_to_id)} upserted")
# Build game_slug -> vg_id mapping
game_slug_to_vg_id: dict[str, int] = {}
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
for game_slug in vg_info["games"]:
game_slug_to_vg_id[game_slug] = vg_id
# Build game_slug -> vg_id mapping
game_slug_to_vg_id: dict[str, int] = {}
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
for game_slug in vg_info["games"]:
game_slug_to_vg_id[game_slug] = vg_id
# 2. Upsert games (with version_group_id)
games_data = load_json("games.json")
slug_to_id = await upsert_games(session, games_data, game_slug_to_vg_id)
print(f"Games: {len(slug_to_id)} upserted")
# 2. Upsert games (with version_group_id)
games_data = load_json("games.json")
slug_to_id = await upsert_games(session, games_data, game_slug_to_vg_id)
print(f"Games: {len(slug_to_id)} upserted")
# 3. Upsert Pokemon
pokemon_data = load_json("pokemon.json")
dex_to_id = await upsert_pokemon(session, pokemon_data)
print(f"Pokemon: {len(dex_to_id)} upserted")
# 3. Upsert Pokemon
pokemon_data = load_json("pokemon.json")
dex_to_id = await upsert_pokemon(session, pokemon_data)
print(f"Pokemon: {len(dex_to_id)} upserted")
# 4. Per version group: upsert routes once, then encounters per game
total_routes = 0
total_encounters = 0
route_maps_by_vg: dict[int, dict[str, int]] = {}
# 4. Per version group: upsert routes once, then encounters per game
total_routes = 0
total_encounters = 0
route_maps_by_vg: dict[int, dict[str, int]] = {}
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
game_slugs = list(vg_info["games"].keys())
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
game_slugs = list(vg_info["games"].keys())
# Use the first game's route JSON for the shared route structure
first_game_slug = game_slugs[0]
routes_file = DATA_DIR / f"{first_game_slug}.json"
if not routes_file.exists():
print(f" {vg_slug}: no route data ({first_game_slug}.json), skipping")
# Use the first game's route JSON for the shared route structure
first_game_slug = game_slugs[0]
routes_file = DATA_DIR / f"{first_game_slug}.json"
if not routes_file.exists():
print(f" {vg_slug}: no route data ({first_game_slug}.json), skipping")
continue
routes_data = load_json(f"{first_game_slug}.json")
if not routes_data:
print(f" {vg_slug}: empty route data, skipping")
continue
# Upsert routes once per version group
route_map = await upsert_routes(session, vg_id, routes_data)
route_maps_by_vg[vg_id] = route_map
total_routes += len(route_map)
print(f" {vg_slug}: {len(route_map)} routes")
# Upsert encounters per game (each game may have different encounters)
for game_slug in game_slugs:
game_id = slug_to_id.get(game_slug)
if game_id is None:
print(f" Warning: game '{game_slug}' not found, skipping")
continue
routes_data = load_json(f"{first_game_slug}.json")
if not routes_data:
print(f" {vg_slug}: empty route data, skipping")
game_routes_file = DATA_DIR / f"{game_slug}.json"
if not game_routes_file.exists():
continue
# Upsert routes once per version group
route_map = await upsert_routes(session, vg_id, routes_data)
route_maps_by_vg[vg_id] = route_map
total_routes += len(route_map)
print(f" {vg_slug}: {len(route_map)} routes")
# Upsert encounters per game (each game may have different encounters)
for game_slug in game_slugs:
game_id = slug_to_id.get(game_slug)
if game_id is None:
print(f" Warning: game '{game_slug}' not found, skipping")
game_routes_data = load_json(f"{game_slug}.json")
for route in game_routes_data:
route_id = route_map.get(route["name"])
if route_id is None:
print(f" Warning: route '{route['name']}' not found")
continue
game_routes_file = DATA_DIR / f"{game_slug}.json"
if not game_routes_file.exists():
continue
# Parent routes may have empty encounters
if route["encounters"]:
enc_count = await upsert_route_encounters(
session,
route_id,
route["encounters"],
dex_to_id,
game_id,
)
total_encounters += enc_count
game_routes_data = load_json(f"{game_slug}.json")
for route in game_routes_data:
route_id = route_map.get(route["name"])
if route_id is None:
print(f" Warning: route '{route['name']}' not found")
# Handle child routes
for child in route.get("children", []):
child_id = route_map.get(child["name"])
if child_id is None:
print(
f" Warning: child route '{child['name']}' not found"
)
continue
# Parent routes may have empty encounters
if route["encounters"]:
enc_count = await upsert_route_encounters(
session, route_id, route["encounters"],
dex_to_id, game_id,
)
total_encounters += enc_count
enc_count = await upsert_route_encounters(
session,
child_id,
child["encounters"],
dex_to_id,
game_id,
)
total_encounters += enc_count
# Handle child routes
for child in route.get("children", []):
child_id = route_map.get(child["name"])
if child_id is None:
print(f" Warning: child route '{child['name']}' not found")
continue
print(f" {game_slug}: encounters loaded")
enc_count = await upsert_route_encounters(
session, child_id, child["encounters"],
dex_to_id, game_id,
)
total_encounters += enc_count
print(f"\nTotal routes: {total_routes}")
print(f"Total encounters: {total_encounters}")
print(f" {game_slug}: encounters loaded")
# 5. Per version group: upsert bosses
total_bosses = 0
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
first_game_slug = list(vg_info["games"].keys())[0]
bosses_file = DATA_DIR / f"{first_game_slug}-bosses.json"
if not bosses_file.exists():
continue
print(f"\nTotal routes: {total_routes}")
print(f"Total encounters: {total_encounters}")
bosses_data = load_json(f"{first_game_slug}-bosses.json")
if not bosses_data:
continue
# 5. Per version group: upsert bosses
total_bosses = 0
for vg_slug, vg_info in vg_data.items():
vg_id = vg_slug_to_id[vg_slug]
first_game_slug = list(vg_info["games"].keys())[0]
bosses_file = DATA_DIR / f"{first_game_slug}-bosses.json"
if not bosses_file.exists():
continue
route_name_to_id = route_maps_by_vg.get(vg_id, {})
boss_count = await upsert_bosses(
session, vg_id, bosses_data, dex_to_id, route_name_to_id
)
total_bosses += boss_count
print(f" {vg_slug}: {boss_count} bosses")
bosses_data = load_json(f"{first_game_slug}-bosses.json")
if not bosses_data:
continue
print(f"Total bosses: {total_bosses}")
route_name_to_id = route_maps_by_vg.get(vg_id, {})
boss_count = await upsert_bosses(session, vg_id, bosses_data, dex_to_id, route_name_to_id)
total_bosses += boss_count
print(f" {vg_slug}: {boss_count} bosses")
print(f"Total bosses: {total_bosses}")
# 6. Upsert evolutions
evolutions_path = DATA_DIR / "evolutions.json"
if evolutions_path.exists():
evolutions_data = load_json("evolutions.json")
evo_count = await upsert_evolutions(session, evolutions_data, dex_to_id)
print(f"Evolutions: {evo_count} upserted")
else:
print("No evolutions.json found, skipping evolutions")
# 6. Upsert evolutions
evolutions_path = DATA_DIR / "evolutions.json"
if evolutions_path.exists():
evolutions_data = load_json("evolutions.json")
evo_count = await upsert_evolutions(session, evolutions_data, dex_to_id)
print(f"Evolutions: {evo_count} upserted")
else:
print("No evolutions.json found, skipping evolutions")
print("Seed complete!")
@@ -180,7 +189,9 @@ async def verify():
games_count = (await session.execute(select(func.count(Game.id)))).scalar()
pokemon_count = (await session.execute(select(func.count(Pokemon.id)))).scalar()
routes_count = (await session.execute(select(func.count(Route.id)))).scalar()
enc_count = (await session.execute(select(func.count(RouteEncounter.id)))).scalar()
enc_count = (
await session.execute(select(func.count(RouteEncounter.id)))
).scalar()
boss_count = (await session.execute(select(func.count(BossBattle.id)))).scalar()
print(f"Version Groups: {vg_count}")
@@ -328,7 +339,7 @@ async def _export_routes(session: AsyncSession, vg_data: dict):
games_by_slug = {g.slug: g for g in game_result.scalars().all()}
exported = 0
for vg_slug, vg_info in vg_data.items():
for _vg_slug, vg_info in vg_data.items():
for game_slug in vg_info["games"]:
game = games_by_slug.get(game_slug)
if game is None or game.version_group_id is None:
@@ -356,11 +367,9 @@ async def _export_routes(session: AsyncSession, vg_data: dict):
if r.parent_route_id is not None:
children_by_parent.setdefault(r.parent_route_id, []).append(r)
def format_encounters(route: Route) -> list[dict]:
def format_encounters(route: Route, _game: Game = game) -> list[dict]:
game_encounters = [
enc
for enc in route.route_encounters
if enc.game_id == game.id
enc for enc in route.route_encounters if enc.game_id == _game.id
]
return [
{
@@ -384,17 +393,20 @@ async def _export_routes(session: AsyncSession, vg_data: dict):
data["pinwheel_zone"] = route.pinwheel_zone
return data
def format_route(route: Route) -> dict:
def format_route(
route: Route,
_children_by_parent: dict[int, list[Route]] = children_by_parent,
) -> dict:
data: dict = {
"name": route.name,
"order": route.order,
"encounters": format_encounters(route),
}
children = children_by_parent.get(route.id, [])
children = _children_by_parent.get(route.id, [])
if children:
data["children"] = [
format_child(c)
for c in sorted(children, key=lambda r: r.order)
for c in sorted(children, key=lambda route: route.order)
]
return data
@@ -444,7 +456,9 @@ def _download_image(
if filename not in downloaded:
output_dir.mkdir(parents=True, exist_ok=True)
req = urllib.request.Request(url, headers={"User-Agent": "nuzlocke-tracker/1.0"})
req = urllib.request.Request(
url, headers={"User-Agent": "nuzlocke-tracker/1.0"}
)
try:
with urllib.request.urlopen(req, timeout=30) as resp:
dest.write_bytes(resp.read())
@@ -496,37 +510,45 @@ async def _export_bosses(session: AsyncSession, vg_data: dict):
if badge_image_url and b.badge_name:
badge_slug = _slugify(b.badge_name)
badge_image_url = _download_image(
badge_image_url, badge_dir, badge_slug, downloaded_badges,
badge_image_url,
badge_dir,
badge_slug,
downloaded_badges,
)
if sprite_url:
sprite_slug = _slugify(b.name)
sprite_url = _download_image(
sprite_url, sprite_dir, sprite_slug, downloaded_sprites,
sprite_url,
sprite_dir,
sprite_slug,
downloaded_sprites,
)
data.append({
"name": b.name,
"boss_type": b.boss_type,
"specialty_type": b.specialty_type,
"badge_name": b.badge_name,
"badge_image_url": badge_image_url,
"level_cap": b.level_cap,
"order": b.order,
"after_route_name": b.after_route.name if b.after_route else None,
"location": b.location,
"section": b.section,
"sprite_url": sprite_url,
"pokemon": [
{
"pokeapi_id": bp.pokemon.pokeapi_id,
"pokemon_name": bp.pokemon.name,
"level": bp.level,
"order": bp.order,
}
for bp in sorted(b.pokemon, key=lambda p: p.order)
],
})
data.append(
{
"name": b.name,
"boss_type": b.boss_type,
"specialty_type": b.specialty_type,
"badge_name": b.badge_name,
"badge_image_url": badge_image_url,
"level_cap": b.level_cap,
"order": b.order,
"after_route_name": b.after_route.name if b.after_route else None,
"location": b.location,
"section": b.section,
"sprite_url": sprite_url,
"pokemon": [
{
"pokeapi_id": bp.pokemon.pokeapi_id,
"pokemon_name": bp.pokemon.name,
"level": bp.level,
"order": bp.order,
}
for bp in sorted(b.pokemon, key=lambda p: p.order)
],
}
)
_write_json(f"{first_game_slug}-bosses.json", data)
exported += 1