many changes
This commit is contained in:
parent
6d1c618a3a
commit
e3797f32ca
337
.gitea/workflows/build-release.yml
Normal file
337
.gitea/workflows/build-release.yml
Normal file
@ -0,0 +1,337 @@
|
|||||||
|
name: Build & Release
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags:
|
||||||
|
- 'v*'
|
||||||
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
version:
|
||||||
|
description: 'Version to build (e.g., v1.0.0)'
|
||||||
|
required: false
|
||||||
|
|
||||||
|
env:
|
||||||
|
RUST_BACKTRACE: 1
|
||||||
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test:
|
||||||
|
name: Run Tests
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Bun
|
||||||
|
uses: oven-sh/setup-bun@v1
|
||||||
|
|
||||||
|
- name: Setup Rust
|
||||||
|
uses: actions-rs/toolchain@v1
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
override: true
|
||||||
|
|
||||||
|
- name: Cache Rust dependencies
|
||||||
|
uses: actions/cache@v3
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/bin/
|
||||||
|
~/.cargo/registry/index/
|
||||||
|
~/.cargo/registry/cache/
|
||||||
|
~/.cargo/git/db/
|
||||||
|
target/
|
||||||
|
key: ${{ runner.os }}-cargo-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: bun install
|
||||||
|
|
||||||
|
- name: Run frontend tests
|
||||||
|
run: bun run test --run
|
||||||
|
continue-on-error: false
|
||||||
|
|
||||||
|
- name: Run Rust tests
|
||||||
|
run: bun run test:rust
|
||||||
|
continue-on-error: false
|
||||||
|
|
||||||
|
- name: Check TypeScript
|
||||||
|
run: bun run check
|
||||||
|
continue-on-error: false
|
||||||
|
|
||||||
|
build-linux:
|
||||||
|
name: Build Linux
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
needs: test
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Bun
|
||||||
|
uses: oven-sh/setup-bun@v1
|
||||||
|
|
||||||
|
- name: Setup Rust
|
||||||
|
uses: actions-rs/toolchain@v1
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
override: true
|
||||||
|
|
||||||
|
- name: Install system dependencies
|
||||||
|
run: |
|
||||||
|
sudo apt-get update
|
||||||
|
sudo apt-get install -y \
|
||||||
|
libwebkit2gtk-4.1-dev \
|
||||||
|
build-essential \
|
||||||
|
curl \
|
||||||
|
wget \
|
||||||
|
file \
|
||||||
|
libssl-dev \
|
||||||
|
libgtk-3-dev \
|
||||||
|
libayatana-appindicator3-dev \
|
||||||
|
librsvg2-dev
|
||||||
|
|
||||||
|
- name: Cache Rust dependencies
|
||||||
|
uses: actions/cache@v3
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/bin/
|
||||||
|
~/.cargo/registry/index/
|
||||||
|
~/.cargo/registry/cache/
|
||||||
|
~/.cargo/git/db/
|
||||||
|
target/
|
||||||
|
key: ${{ runner.os }}-cargo-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: bun install
|
||||||
|
|
||||||
|
- name: Build for Linux
|
||||||
|
run: bun run tauri build
|
||||||
|
env:
|
||||||
|
TAURI_SKIP_UPDATER: true
|
||||||
|
|
||||||
|
- name: Prepare Linux artifacts
|
||||||
|
run: |
|
||||||
|
mkdir -p dist/linux
|
||||||
|
# Copy AppImage
|
||||||
|
if [ -f "src-tauri/target/release/bundle/appimage/jellytau_"*.AppImage ]; then
|
||||||
|
cp src-tauri/target/release/bundle/appimage/jellytau_*.AppImage dist/linux/
|
||||||
|
fi
|
||||||
|
# Copy .deb if built
|
||||||
|
if [ -f "src-tauri/target/release/bundle/deb/jellytau_"*.deb ]; then
|
||||||
|
cp src-tauri/target/release/bundle/deb/jellytau_*.deb dist/linux/
|
||||||
|
fi
|
||||||
|
ls -lah dist/linux/
|
||||||
|
|
||||||
|
- name: Upload Linux build artifact
|
||||||
|
uses: actions/upload-artifact@v3
|
||||||
|
with:
|
||||||
|
name: jellytau-linux
|
||||||
|
path: dist/linux/
|
||||||
|
retention-days: 30
|
||||||
|
|
||||||
|
build-android:
|
||||||
|
name: Build Android
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
needs: test
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Bun
|
||||||
|
uses: oven-sh/setup-bun@v1
|
||||||
|
|
||||||
|
- name: Setup Java
|
||||||
|
uses: actions/setup-java@v3
|
||||||
|
with:
|
||||||
|
distribution: 'temurin'
|
||||||
|
java-version: '17'
|
||||||
|
|
||||||
|
- name: Setup Android SDK
|
||||||
|
uses: android-actions/setup-android@v2
|
||||||
|
with:
|
||||||
|
api-level: 33
|
||||||
|
|
||||||
|
- name: Setup Rust
|
||||||
|
uses: actions-rs/toolchain@v1
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
override: true
|
||||||
|
|
||||||
|
- name: Add Android targets
|
||||||
|
run: |
|
||||||
|
rustup target add aarch64-linux-android
|
||||||
|
rustup target add armv7-linux-androideabi
|
||||||
|
rustup target add x86_64-linux-android
|
||||||
|
|
||||||
|
- name: Install Android NDK
|
||||||
|
run: |
|
||||||
|
sdkmanager "ndk;25.1.8937393"
|
||||||
|
|
||||||
|
- name: Cache Rust dependencies
|
||||||
|
uses: actions/cache@v3
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/bin/
|
||||||
|
~/.cargo/registry/index/
|
||||||
|
~/.cargo/registry/cache/
|
||||||
|
~/.cargo/git/db/
|
||||||
|
target/
|
||||||
|
key: ${{ runner.os }}-cargo-android-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-android-
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: bun install
|
||||||
|
|
||||||
|
- name: Build for Android
|
||||||
|
run: bun run tauri android build
|
||||||
|
env:
|
||||||
|
ANDROID_NDK_HOME: ${{ android.ndk-home }}
|
||||||
|
ANDROID_SDK_ROOT: ${{ android.sdk-root }}
|
||||||
|
ANDROID_HOME: ${{ android.sdk-root }}
|
||||||
|
|
||||||
|
- name: Prepare Android artifacts
|
||||||
|
run: |
|
||||||
|
mkdir -p dist/android
|
||||||
|
# Copy APK
|
||||||
|
if [ -f "src-tauri/gen/android/app/build/outputs/apk/release/app-release.apk" ]; then
|
||||||
|
cp src-tauri/gen/android/app/build/outputs/apk/release/app-release.apk dist/android/jellytau-release.apk
|
||||||
|
fi
|
||||||
|
# Copy AAB (Android App Bundle) if built
|
||||||
|
if [ -f "src-tauri/gen/android/app/build/outputs/bundle/release/app-release.aab" ]; then
|
||||||
|
cp src-tauri/gen/android/app/build/outputs/bundle/release/app-release.aab dist/android/jellytau-release.aab
|
||||||
|
fi
|
||||||
|
ls -lah dist/android/
|
||||||
|
|
||||||
|
- name: Upload Android build artifact
|
||||||
|
uses: actions/upload-artifact@v3
|
||||||
|
with:
|
||||||
|
name: jellytau-android
|
||||||
|
path: dist/android/
|
||||||
|
retention-days: 30
|
||||||
|
|
||||||
|
create-release:
|
||||||
|
name: Create Release
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
needs: [build-linux, build-android]
|
||||||
|
if: startsWith(github.ref, 'refs/tags/v')
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Get version from tag
|
||||||
|
id: tag_name
|
||||||
|
run: |
|
||||||
|
echo "VERSION=${GITHUB_REF#refs/tags/}" >> $GITHUB_OUTPUT
|
||||||
|
echo "RELEASE_NAME=JellyTau ${GITHUB_REF#refs/tags/}" >> $GITHUB_OUTPUT
|
||||||
|
|
||||||
|
- name: Download Linux artifacts
|
||||||
|
uses: actions/download-artifact@v3
|
||||||
|
with:
|
||||||
|
name: jellytau-linux
|
||||||
|
path: artifacts/linux/
|
||||||
|
|
||||||
|
- name: Download Android artifacts
|
||||||
|
uses: actions/download-artifact@v3
|
||||||
|
with:
|
||||||
|
name: jellytau-android
|
||||||
|
path: artifacts/android/
|
||||||
|
|
||||||
|
- name: Prepare release notes
|
||||||
|
id: release_notes
|
||||||
|
run: |
|
||||||
|
VERSION="${{ steps.tag_name.outputs.VERSION }}"
|
||||||
|
echo "## 📱 JellyTau $VERSION Release" > release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "### 📦 Downloads" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "#### Linux" >> release_notes.md
|
||||||
|
echo "- **AppImage** - Run directly on most Linux distributions" >> release_notes.md
|
||||||
|
echo "- **DEB** - Install via `sudo dpkg -i jellytau_*.deb` (Ubuntu/Debian)" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "#### Android" >> release_notes.md
|
||||||
|
echo "- **APK** - Install via `adb install jellytau-release.apk` or sideload via file manager" >> release_notes.md
|
||||||
|
echo "- **AAB** - Upload to Google Play Console or testing platforms" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "### ✨ What's New" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "See [CHANGELOG.md](CHANGELOG.md) for detailed changes." >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "### 🔧 Installation" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "#### Linux (AppImage)" >> release_notes.md
|
||||||
|
echo "\`\`\`bash" >> release_notes.md
|
||||||
|
echo "chmod +x jellytau_*.AppImage" >> release_notes.md
|
||||||
|
echo "./jellytau_*.AppImage" >> release_notes.md
|
||||||
|
echo "\`\`\`" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "#### Linux (DEB)" >> release_notes.md
|
||||||
|
echo "\`\`\`bash" >> release_notes.md
|
||||||
|
echo "sudo dpkg -i jellytau_*.deb" >> release_notes.md
|
||||||
|
echo "jellytau" >> release_notes.md
|
||||||
|
echo "\`\`\`" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "#### Android" >> release_notes.md
|
||||||
|
echo "- Sideload: Download APK and install via file manager or ADB" >> release_notes.md
|
||||||
|
echo "- Play Store: Coming soon" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "### 🐛 Known Issues" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "See [GitHub Issues](../../issues) for reported bugs." >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "### 📝 Requirements" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "**Linux:**" >> release_notes.md
|
||||||
|
echo "- 64-bit Linux system" >> release_notes.md
|
||||||
|
echo "- GLIBC 2.29+" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "**Android:**" >> release_notes.md
|
||||||
|
echo "- Android 8.0 or higher" >> release_notes.md
|
||||||
|
echo "- 50MB free storage" >> release_notes.md
|
||||||
|
echo "" >> release_notes.md
|
||||||
|
echo "---" >> release_notes.md
|
||||||
|
echo "Built with Tauri, SvelteKit, and Rust 🦀" >> release_notes.md
|
||||||
|
|
||||||
|
- name: Create GitHub Release
|
||||||
|
uses: softprops/action-gh-release@v1
|
||||||
|
if: startsWith(github.ref, 'refs/tags/')
|
||||||
|
with:
|
||||||
|
name: ${{ steps.tag_name.outputs.RELEASE_NAME }}
|
||||||
|
body_path: release_notes.md
|
||||||
|
files: |
|
||||||
|
artifacts/linux/*
|
||||||
|
artifacts/android/*
|
||||||
|
draft: false
|
||||||
|
prerelease: ${{ contains(steps.tag_name.outputs.VERSION, 'rc') || contains(steps.tag_name.outputs.VERSION, 'beta') || contains(steps.tag_name.outputs.VERSION, 'alpha') }}
|
||||||
|
env:
|
||||||
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
|
||||||
|
- name: Upload to Gitea Releases
|
||||||
|
run: |
|
||||||
|
VERSION="${{ steps.tag_name.outputs.VERSION }}"
|
||||||
|
|
||||||
|
echo "📦 Release artifacts prepared for $VERSION"
|
||||||
|
echo ""
|
||||||
|
echo "Linux:"
|
||||||
|
ls -lh artifacts/linux/ || echo "No Linux artifacts"
|
||||||
|
echo ""
|
||||||
|
echo "Android:"
|
||||||
|
ls -lh artifacts/android/ || echo "No Android artifacts"
|
||||||
|
echo ""
|
||||||
|
echo "✅ Release $VERSION is ready!"
|
||||||
|
echo "📄 Release notes saved to release_notes.md"
|
||||||
|
|
||||||
|
- name: Publish release notes
|
||||||
|
run: |
|
||||||
|
echo "## 🎉 Release Published"
|
||||||
|
echo ""
|
||||||
|
echo "**Version:** ${{ steps.tag_name.outputs.VERSION }}"
|
||||||
|
echo "**Tag:** ${{ github.ref }}"
|
||||||
|
echo ""
|
||||||
|
echo "Artifacts:"
|
||||||
|
echo "- Linux artifacts in: artifacts/linux/"
|
||||||
|
echo "- Android artifacts in: artifacts/android/"
|
||||||
|
echo ""
|
||||||
|
echo "Visit the Release page to download files."
|
||||||
142
.gitea/workflows/traceability-check.yml
Normal file
142
.gitea/workflows/traceability-check.yml
Normal file
@ -0,0 +1,142 @@
|
|||||||
|
name: Traceability Validation
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
- main
|
||||||
|
- develop
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
- main
|
||||||
|
- develop
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
validate-traces:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
name: Check Requirement Traces
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
fetch-depth: 0
|
||||||
|
|
||||||
|
- name: Setup Bun
|
||||||
|
uses: oven-sh/setup-bun@v1
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: bun install
|
||||||
|
|
||||||
|
- name: Extract traces
|
||||||
|
run: |
|
||||||
|
echo "🔍 Extracting requirement traces..."
|
||||||
|
bun run traces:json > traces-report.json
|
||||||
|
|
||||||
|
- name: Validate traces
|
||||||
|
run: |
|
||||||
|
set -e
|
||||||
|
|
||||||
|
echo "📊 Validating requirement traceability..."
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Parse JSON
|
||||||
|
TOTAL_TRACES=$(jq '.totalTraces' traces-report.json)
|
||||||
|
UR=$(jq '.byType.UR | length' traces-report.json)
|
||||||
|
IR=$(jq '.byType.IR | length' traces-report.json)
|
||||||
|
DR=$(jq '.byType.DR | length' traces-report.json)
|
||||||
|
JA=$(jq '.byType.JA | length' traces-report.json)
|
||||||
|
|
||||||
|
# Print coverage report
|
||||||
|
echo "✅ TRACES Found: $TOTAL_TRACES"
|
||||||
|
echo ""
|
||||||
|
echo "📋 Coverage Summary:"
|
||||||
|
echo " User Requirements (UR): $UR / 39 ($(( UR * 100 / 39 ))%)"
|
||||||
|
echo " Integration Requirements (IR): $IR / 24 ($(( IR * 100 / 24 ))%)"
|
||||||
|
echo " Development Requirements (DR): $DR / 48 ($(( DR * 100 / 48 ))%)"
|
||||||
|
echo " Jellyfin API Requirements (JA): $JA / 3 ($(( JA * 100 / 3 ))%)"
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
COVERED=$((UR + IR + DR + JA))
|
||||||
|
TOTAL_REQS=114
|
||||||
|
COVERAGE=$((COVERED * 100 / TOTAL_REQS))
|
||||||
|
|
||||||
|
echo "📈 Overall Coverage: $COVERED / $TOTAL_REQS ($COVERAGE%)"
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Check minimum threshold
|
||||||
|
MIN_THRESHOLD=50
|
||||||
|
if [ "$COVERAGE" -lt "$MIN_THRESHOLD" ]; then
|
||||||
|
echo "❌ ERROR: Coverage ($COVERAGE%) is below minimum threshold ($MIN_THRESHOLD%)"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "✅ Coverage is acceptable ($COVERAGE% >= $MIN_THRESHOLD%)"
|
||||||
|
|
||||||
|
- name: Check modified files
|
||||||
|
if: github.event_name == 'pull_request'
|
||||||
|
run: |
|
||||||
|
echo "🔍 Checking modified files for traces..."
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Get changed files
|
||||||
|
CHANGED=$(git diff --name-only origin/${{ github.base_ref }}...HEAD | grep -E '\.(ts|tsx|svelte|rs)$' || echo "")
|
||||||
|
|
||||||
|
if [ -z "$CHANGED" ]; then
|
||||||
|
echo "✅ No TypeScript/Rust files changed"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "📝 Changed files:"
|
||||||
|
echo "$CHANGED" | sed 's/^/ /'
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Check each file
|
||||||
|
MISSING_TRACES=0
|
||||||
|
while IFS= read -r file; do
|
||||||
|
# Skip test files
|
||||||
|
if [[ "$file" == *".test."* ]]; then
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -f "$file" ]; then
|
||||||
|
if ! grep -q "TRACES:" "$file"; then
|
||||||
|
echo "⚠️ Missing TRACES: $file"
|
||||||
|
MISSING_TRACES=$((MISSING_TRACES + 1))
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done <<< "$CHANGED"
|
||||||
|
|
||||||
|
if [ "$MISSING_TRACES" -gt 0 ]; then
|
||||||
|
echo ""
|
||||||
|
echo "📝 Recommendation: Add TRACES comments to new/modified code"
|
||||||
|
echo " Format: // TRACES: UR-001, UR-002 | DR-003"
|
||||||
|
echo ""
|
||||||
|
echo "💡 For more info, see: scripts/README.md"
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: Generate full report
|
||||||
|
if: always()
|
||||||
|
run: |
|
||||||
|
echo "📄 Generating full traceability report..."
|
||||||
|
bun run traces:markdown
|
||||||
|
|
||||||
|
- name: Display report summary
|
||||||
|
if: always()
|
||||||
|
run: |
|
||||||
|
echo ""
|
||||||
|
echo "📊 Full Report Generated"
|
||||||
|
echo "📁 Location: docs/TRACEABILITY.md"
|
||||||
|
echo ""
|
||||||
|
head -50 docs/TRACEABILITY.md || true
|
||||||
|
|
||||||
|
- name: Save artifacts
|
||||||
|
if: always()
|
||||||
|
uses: actions/upload-artifact@v3
|
||||||
|
with:
|
||||||
|
name: traceability-reports
|
||||||
|
path: |
|
||||||
|
traces-report.json
|
||||||
|
docs/TRACEABILITY.md
|
||||||
|
retention-days: 30
|
||||||
173
.gitea/workflows/traceability.yml
Normal file
173
.gitea/workflows/traceability.yml
Normal file
@ -0,0 +1,173 @@
|
|||||||
|
name: Requirement Traceability Check
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
- main
|
||||||
|
- develop
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
- main
|
||||||
|
- develop
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
traceability:
|
||||||
|
name: Validate Requirement Traces
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout code
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Bun
|
||||||
|
uses: oven-sh/setup-bun@v1
|
||||||
|
with:
|
||||||
|
bun-version: latest
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: bun install
|
||||||
|
|
||||||
|
- name: Extract requirement traces
|
||||||
|
run: bun run traces:json > traces.json
|
||||||
|
|
||||||
|
- name: Validate trace format
|
||||||
|
run: |
|
||||||
|
if ! jq empty traces.json 2>/dev/null; then
|
||||||
|
echo "❌ Invalid traces.json format"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
echo "✅ Traces JSON is valid"
|
||||||
|
|
||||||
|
- name: Check requirement coverage
|
||||||
|
run: |
|
||||||
|
set -e
|
||||||
|
|
||||||
|
# Extract coverage stats
|
||||||
|
TOTAL_TRACES=$(jq '.totalTraces' traces.json)
|
||||||
|
UR_COUNT=$(jq '.byType.UR | length' traces.json)
|
||||||
|
IR_COUNT=$(jq '.byType.IR | length' traces.json)
|
||||||
|
DR_COUNT=$(jq '.byType.DR | length' traces.json)
|
||||||
|
JA_COUNT=$(jq '.byType.JA | length' traces.json)
|
||||||
|
|
||||||
|
echo "## 📊 Requirement Traceability Report"
|
||||||
|
echo ""
|
||||||
|
echo "**Total TRACES Found:** $TOTAL_TRACES"
|
||||||
|
echo ""
|
||||||
|
echo "### Requirements Covered:"
|
||||||
|
echo "- User Requirements (UR): $UR_COUNT / 39 ($(( UR_COUNT * 100 / 39 ))%)"
|
||||||
|
echo "- Integration Requirements (IR): $IR_COUNT / 24 ($(( IR_COUNT * 100 / 24 ))%)"
|
||||||
|
echo "- Development Requirements (DR): $DR_COUNT / 48 ($(( DR_COUNT * 100 / 48 ))%)"
|
||||||
|
echo "- Jellyfin API Requirements (JA): $JA_COUNT / 3 ($(( JA_COUNT * 100 / 3 ))%)"
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Set minimum coverage threshold (50%)
|
||||||
|
TOTAL_REQS=114
|
||||||
|
MIN_COVERAGE=$((TOTAL_REQS / 2))
|
||||||
|
COVERED=$((UR_COUNT + IR_COUNT + DR_COUNT + JA_COUNT))
|
||||||
|
COVERAGE_PERCENT=$((COVERED * 100 / TOTAL_REQS))
|
||||||
|
|
||||||
|
echo "**Overall Coverage:** $COVERED / $TOTAL_REQS ($COVERAGE_PERCENT%)"
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
if [ "$COVERED" -lt "$MIN_COVERAGE" ]; then
|
||||||
|
echo "❌ Coverage below minimum threshold ($COVERAGE_PERCENT% < 50%)"
|
||||||
|
exit 1
|
||||||
|
else
|
||||||
|
echo "✅ Coverage meets minimum threshold ($COVERAGE_PERCENT% >= 50%)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: Check for new untraced code
|
||||||
|
run: |
|
||||||
|
set -e
|
||||||
|
|
||||||
|
# Find files modified in this PR/push
|
||||||
|
if [ "${{ github.event_name }}" = "pull_request" ]; then
|
||||||
|
CHANGED_FILES=$(git diff --name-only origin/${{ github.base_ref }}...HEAD | grep -E '\.(ts|tsx|svelte|rs)$' || true)
|
||||||
|
else
|
||||||
|
CHANGED_FILES=$(git diff --name-only HEAD~1 | grep -E '\.(ts|tsx|svelte|rs)$' || true)
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -z "$CHANGED_FILES" ]; then
|
||||||
|
echo "✅ No source files changed"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "### Files Changed:"
|
||||||
|
echo "$CHANGED_FILES" | sed 's/^/- /'
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Check if changed files have TRACES
|
||||||
|
UNTRACED_FILES=""
|
||||||
|
while IFS= read -r file; do
|
||||||
|
if [ -f "$file" ]; then
|
||||||
|
# Skip test files and generated code
|
||||||
|
if [[ "$file" == *".test."* ]] || [[ "$file" == *"node_modules"* ]]; then
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check if file has TRACES comments
|
||||||
|
if ! grep -q "TRACES:" "$file" 2>/dev/null; then
|
||||||
|
UNTRACED_FILES+="$file"$'\n'
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done <<< "$CHANGED_FILES"
|
||||||
|
|
||||||
|
if [ -n "$UNTRACED_FILES" ]; then
|
||||||
|
echo "⚠️ New files without TRACES:"
|
||||||
|
echo "$UNTRACED_FILES" | sed 's/^/ - /'
|
||||||
|
echo ""
|
||||||
|
echo "💡 Add TRACES comments to link code to requirements:"
|
||||||
|
echo " // TRACES: UR-001, UR-002 | DR-003"
|
||||||
|
else
|
||||||
|
echo "✅ All changed files have TRACES comments"
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: Generate traceability report
|
||||||
|
if: always()
|
||||||
|
run: bun run traces:markdown
|
||||||
|
|
||||||
|
- name: Upload traceability report
|
||||||
|
if: always()
|
||||||
|
uses: actions/upload-artifact@v3
|
||||||
|
with:
|
||||||
|
name: traceability-report
|
||||||
|
path: docs/TRACEABILITY.md
|
||||||
|
retention-days: 30
|
||||||
|
|
||||||
|
- name: Comment PR with coverage report
|
||||||
|
if: github.event_name == 'pull_request'
|
||||||
|
uses: actions/github-script@v7
|
||||||
|
with:
|
||||||
|
script: |
|
||||||
|
const fs = require('fs');
|
||||||
|
const traces = JSON.parse(fs.readFileSync('traces.json', 'utf8'));
|
||||||
|
|
||||||
|
const urCount = traces.byType.UR.length;
|
||||||
|
const irCount = traces.byType.IR.length;
|
||||||
|
const drCount = traces.byType.DR.length;
|
||||||
|
const jaCount = traces.byType.JA.length;
|
||||||
|
const total = urCount + irCount + drCount + jaCount;
|
||||||
|
const coverage = Math.round((total / 114) * 100);
|
||||||
|
|
||||||
|
const comment = `## 📊 Requirement Traceability Report
|
||||||
|
|
||||||
|
**Coverage:** ${coverage}% (${total}/114 requirements traced)
|
||||||
|
|
||||||
|
### By Type:
|
||||||
|
- **User Requirements (UR):** ${urCount}/39 (${Math.round(urCount/39*100)}%)
|
||||||
|
- **Integration Requirements (IR):** ${irCount}/24 (${Math.round(irCount/24*100)}%)
|
||||||
|
- **Development Requirements (DR):** ${drCount}/48 (${Math.round(drCount/48*100)}%)
|
||||||
|
- **Jellyfin API (JA):** ${jaCount}/3 (${Math.round(jaCount/3*100)}%)
|
||||||
|
|
||||||
|
**Total Traces:** ${traces.totalTraces}
|
||||||
|
|
||||||
|
[View full report](artifacts) | [Format Guide](https://github.com/yourusername/jellytau/blob/master/scripts/README.md#extract-tracests)`;
|
||||||
|
|
||||||
|
github.rest.issues.createComment({
|
||||||
|
issue_number: context.issue.number,
|
||||||
|
owner: context.repo.owner,
|
||||||
|
repo: context.repo.repo,
|
||||||
|
body: comment
|
||||||
|
});
|
||||||
347
docs/BUILD_RELEASE.md
Normal file
347
docs/BUILD_RELEASE.md
Normal file
@ -0,0 +1,347 @@
|
|||||||
|
# Build & Release Workflow
|
||||||
|
|
||||||
|
This document explains the automated build and release process for JellyTau.
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
The CI/CD pipeline automatically:
|
||||||
|
1. ✅ Runs all tests (frontend + Rust)
|
||||||
|
2. ✅ Builds Linux binaries (AppImage + DEB)
|
||||||
|
3. ✅ Builds Android APK and AAB
|
||||||
|
4. ✅ Creates releases with artifacts
|
||||||
|
5. ✅ Tags releases with version numbers
|
||||||
|
|
||||||
|
## Workflow Triggers
|
||||||
|
|
||||||
|
### Automatic Trigger
|
||||||
|
When you push a version tag:
|
||||||
|
```bash
|
||||||
|
git tag v1.0.0
|
||||||
|
git push origin v1.0.0
|
||||||
|
```
|
||||||
|
|
||||||
|
The workflow automatically:
|
||||||
|
1. Runs tests
|
||||||
|
2. Builds both platforms
|
||||||
|
3. Creates a GitHub release with artifacts
|
||||||
|
4. Tags it as release/prerelease based on version
|
||||||
|
|
||||||
|
### Manual Trigger
|
||||||
|
In Gitea Actions UI:
|
||||||
|
1. Go to **Actions** tab
|
||||||
|
2. Click **Build & Release** workflow
|
||||||
|
3. Click **Run workflow**
|
||||||
|
4. Optionally specify a version
|
||||||
|
5. Workflow runs without creating a release
|
||||||
|
|
||||||
|
## Version Tagging
|
||||||
|
|
||||||
|
### Format
|
||||||
|
Version tags follow semantic versioning: `v{MAJOR}.{MINOR}.{PATCH}`
|
||||||
|
|
||||||
|
Examples:
|
||||||
|
- `v1.0.0` - Release version
|
||||||
|
- `v1.0.0-rc1` - Release candidate (marked as prerelease)
|
||||||
|
- `v1.0.0-beta` - Beta version (marked as prerelease)
|
||||||
|
- `v0.1.0-alpha` - Alpha version (marked as prerelease)
|
||||||
|
|
||||||
|
### Creating a Release
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Create and push a version tag
|
||||||
|
git tag v1.0.0 -m "Release version 1.0.0"
|
||||||
|
git push origin v1.0.0
|
||||||
|
|
||||||
|
# Or create from main branch
|
||||||
|
git tag -a v1.0.0 -m "Release version 1.0.0" main
|
||||||
|
git push origin v1.0.0
|
||||||
|
```
|
||||||
|
|
||||||
|
### Release Status
|
||||||
|
|
||||||
|
Versions containing `rc`, `beta`, or `alpha` are marked as **prerelease**:
|
||||||
|
```bash
|
||||||
|
git tag v1.0.0-rc1 # ⚠️ Prerelease
|
||||||
|
git tag v1.0.0-beta # ⚠️ Prerelease
|
||||||
|
git tag v1.0.0-alpha # ⚠️ Prerelease
|
||||||
|
git tag v1.0.0 # ✅ Full release
|
||||||
|
```
|
||||||
|
|
||||||
|
## Workflow Steps
|
||||||
|
|
||||||
|
### 1. Test Phase
|
||||||
|
Runs on all tags and manual triggers:
|
||||||
|
- Frontend tests (`vitest`)
|
||||||
|
- Rust tests (`cargo test`)
|
||||||
|
- TypeScript type checking
|
||||||
|
|
||||||
|
**Failure:** Stops workflow, no build/release
|
||||||
|
|
||||||
|
### 2. Build Linux Phase
|
||||||
|
Runs after tests pass:
|
||||||
|
- Installs system dependencies
|
||||||
|
- Builds with Tauri
|
||||||
|
- Generates:
|
||||||
|
- **AppImage** - Universal Linux binary
|
||||||
|
- **DEB** - Debian/Ubuntu package
|
||||||
|
|
||||||
|
**Output:** `artifacts/linux/`
|
||||||
|
|
||||||
|
### 3. Build Android Phase
|
||||||
|
Runs in parallel with Linux build:
|
||||||
|
- Installs Android SDK/NDK
|
||||||
|
- Configures Rust for Android targets
|
||||||
|
- Builds with Tauri
|
||||||
|
- Generates:
|
||||||
|
- **APK** - Android app package (installable)
|
||||||
|
- **AAB** - Android App Bundle (for Play Store)
|
||||||
|
|
||||||
|
**Output:** `artifacts/android/`
|
||||||
|
|
||||||
|
### 4. Create Release Phase
|
||||||
|
Runs after both builds succeed (only on version tags):
|
||||||
|
- Prepares release notes
|
||||||
|
- Downloads build artifacts
|
||||||
|
- Creates GitHub/Gitea release
|
||||||
|
- Uploads all artifacts
|
||||||
|
- Tags as prerelease if applicable
|
||||||
|
|
||||||
|
## Artifacts
|
||||||
|
|
||||||
|
### Linux Artifacts
|
||||||
|
|
||||||
|
#### AppImage
|
||||||
|
- **File:** `jellytau_*.AppImage`
|
||||||
|
- **Size:** ~100-150 MB
|
||||||
|
- **Use:** Run directly on any Linux distro
|
||||||
|
- **Installation:**
|
||||||
|
```bash
|
||||||
|
chmod +x jellytau_*.AppImage
|
||||||
|
./jellytau_*.AppImage
|
||||||
|
```
|
||||||
|
|
||||||
|
#### DEB Package
|
||||||
|
- **File:** `jellytau_*.deb`
|
||||||
|
- **Size:** ~80-120 MB
|
||||||
|
- **Use:** Install on Debian/Ubuntu/similar
|
||||||
|
- **Installation:**
|
||||||
|
```bash
|
||||||
|
sudo dpkg -i jellytau_*.deb
|
||||||
|
jellytau
|
||||||
|
```
|
||||||
|
|
||||||
|
### Android Artifacts
|
||||||
|
|
||||||
|
#### APK
|
||||||
|
- **File:** `jellytau-release.apk`
|
||||||
|
- **Size:** ~60-100 MB
|
||||||
|
- **Use:** Direct installation on Android devices
|
||||||
|
- **Installation:**
|
||||||
|
```bash
|
||||||
|
adb install jellytau-release.apk
|
||||||
|
# Or sideload via file manager
|
||||||
|
```
|
||||||
|
|
||||||
|
#### AAB (Android App Bundle)
|
||||||
|
- **File:** `jellytau-release.aab`
|
||||||
|
- **Size:** ~50-90 MB
|
||||||
|
- **Use:** Upload to Google Play Console
|
||||||
|
- **Note:** Cannot be installed directly; for Play Store distribution
|
||||||
|
|
||||||
|
## Release Notes
|
||||||
|
|
||||||
|
Release notes are automatically generated with:
|
||||||
|
- Version number
|
||||||
|
- Download links
|
||||||
|
- Installation instructions
|
||||||
|
- System requirements
|
||||||
|
- Known issues link
|
||||||
|
- Changelog reference
|
||||||
|
|
||||||
|
## Build Matrix
|
||||||
|
|
||||||
|
| Platform | OS | Architecture | Format |
|
||||||
|
|----------|----|----|--------|
|
||||||
|
| **Linux** | Any | x86_64 | AppImage, DEB |
|
||||||
|
| **Android** | 8.0+ | arm64, armv7, x86_64 | APK, AAB |
|
||||||
|
|
||||||
|
## Troubleshooting
|
||||||
|
|
||||||
|
### Build Fails During Test Phase
|
||||||
|
1. Check test output in Gitea Actions
|
||||||
|
2. Run tests locally: `bun run test` and `bun run test:rust`
|
||||||
|
3. Fix failing tests
|
||||||
|
4. Create new tag with fixed code
|
||||||
|
|
||||||
|
### Linux Build Fails
|
||||||
|
1. Check system dependencies installed
|
||||||
|
2. Verify Tauri configuration
|
||||||
|
3. Check cargo dependencies
|
||||||
|
4. Clear cache: Delete `.cargo` and `target/` directories
|
||||||
|
|
||||||
|
### Android Build Fails
|
||||||
|
1. Check Android SDK/NDK setup
|
||||||
|
2. Verify Java 17 is installed
|
||||||
|
3. Check Rust Android targets: `rustup target list`
|
||||||
|
4. Clear cache and rebuild
|
||||||
|
|
||||||
|
### Release Not Created
|
||||||
|
1. Tag must start with `v` (e.g., `v1.0.0`)
|
||||||
|
2. Tests must pass
|
||||||
|
3. Both builds must succeed
|
||||||
|
4. Check workflow logs for errors
|
||||||
|
|
||||||
|
## GitHub Release vs Gitea
|
||||||
|
|
||||||
|
The workflow uses GitHub Actions SDK but is designed for Gitea. For Gitea-native releases:
|
||||||
|
|
||||||
|
1. Workflow creates artifacts
|
||||||
|
2. Artifacts are available in Actions artifacts
|
||||||
|
3. Download and manually create Gitea release, or
|
||||||
|
4. Set up Gitea API integration to auto-publish
|
||||||
|
|
||||||
|
## Customization
|
||||||
|
|
||||||
|
### Change Release Notes Template
|
||||||
|
|
||||||
|
Edit `.gitea/workflows/build-release.yml`, section `Prepare release notes`:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
- name: Prepare release notes
|
||||||
|
id: release_notes
|
||||||
|
run: |
|
||||||
|
# Add your custom release notes format here
|
||||||
|
echo "Custom notes" > release_notes.md
|
||||||
|
```
|
||||||
|
|
||||||
|
### Add New Platforms
|
||||||
|
|
||||||
|
To add macOS or Windows builds:
|
||||||
|
|
||||||
|
1. Add new `build-{platform}` job
|
||||||
|
2. Set appropriate `runs-on` runner
|
||||||
|
3. Add platform-specific dependencies
|
||||||
|
4. Update artifact upload
|
||||||
|
5. Include in `needs: [build-linux, build-android, build-{platform}]`
|
||||||
|
|
||||||
|
### Change Build Targets
|
||||||
|
|
||||||
|
Modify Tauri configuration or add targets:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
- name: Build for Linux
|
||||||
|
run: |
|
||||||
|
# Add target specification
|
||||||
|
bun run tauri build -- --target x86_64-unknown-linux-gnu
|
||||||
|
```
|
||||||
|
|
||||||
|
## Monitoring
|
||||||
|
|
||||||
|
### Check Status
|
||||||
|
1. Go to **Actions** tab in Gitea
|
||||||
|
2. View **Build & Release** workflow runs
|
||||||
|
3. Click specific run to see logs
|
||||||
|
|
||||||
|
### Notifications
|
||||||
|
Set up notifications for:
|
||||||
|
- Build failures
|
||||||
|
- Release creation
|
||||||
|
- Tag pushes
|
||||||
|
|
||||||
|
## Performance
|
||||||
|
|
||||||
|
### Build Times (Approximate)
|
||||||
|
- Test phase: 5-10 minutes
|
||||||
|
- Linux build: 10-15 minutes
|
||||||
|
- Android build: 15-20 minutes
|
||||||
|
- Total: 30-45 minutes
|
||||||
|
|
||||||
|
### Caching
|
||||||
|
Workflow caches:
|
||||||
|
- Rust dependencies (cargo)
|
||||||
|
- Bun node_modules
|
||||||
|
- Android SDK components
|
||||||
|
|
||||||
|
## Security
|
||||||
|
|
||||||
|
### Secrets
|
||||||
|
The workflow uses:
|
||||||
|
- `GITHUB_TOKEN` - Built-in, no setup needed
|
||||||
|
- No credentials needed for Gitea
|
||||||
|
|
||||||
|
### Verification
|
||||||
|
To verify build integrity:
|
||||||
|
1. Download artifacts
|
||||||
|
2. Verify signatures (if implemented)
|
||||||
|
3. Check file hashes
|
||||||
|
4. Test on target platform
|
||||||
|
|
||||||
|
## Best Practices
|
||||||
|
|
||||||
|
### Versioning
|
||||||
|
1. Follow semantic versioning: `v{MAJOR}.{MINOR}.{PATCH}`
|
||||||
|
2. Tag releases in git
|
||||||
|
3. Update CHANGELOG.md before tagging
|
||||||
|
4. Include release notes in tag message
|
||||||
|
|
||||||
|
### Testing Before Release
|
||||||
|
```bash
|
||||||
|
# Local testing before release
|
||||||
|
bun run test # Frontend tests
|
||||||
|
bun run test:rust # Rust tests
|
||||||
|
bun run check # Type checking
|
||||||
|
bun run tauri build # Local build test
|
||||||
|
```
|
||||||
|
|
||||||
|
### Documentation
|
||||||
|
1. Update [CHANGELOG.md](../CHANGELOG.md) with changes
|
||||||
|
2. Update [README.md](../README.md) with new features
|
||||||
|
3. Document breaking changes
|
||||||
|
4. Add migration guide if needed
|
||||||
|
|
||||||
|
## Example Release Workflow
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# 1. Update version in relevant files (package.json, Cargo.toml, etc.)
|
||||||
|
vim package.json
|
||||||
|
vim src-tauri/tauri.conf.json
|
||||||
|
|
||||||
|
# 2. Update CHANGELOG
|
||||||
|
vim CHANGELOG.md
|
||||||
|
|
||||||
|
# 3. Commit changes
|
||||||
|
git add .
|
||||||
|
git commit -m "Bump version to v1.0.0"
|
||||||
|
|
||||||
|
# 4. Create annotated tag
|
||||||
|
git tag -a v1.0.0 -m "Release version 1.0.0
|
||||||
|
|
||||||
|
Features:
|
||||||
|
- Feature 1
|
||||||
|
- Feature 2
|
||||||
|
|
||||||
|
Fixes:
|
||||||
|
- Fix 1
|
||||||
|
- Fix 2"
|
||||||
|
|
||||||
|
# 5. Push tag to trigger workflow
|
||||||
|
git push origin v1.0.0
|
||||||
|
|
||||||
|
# 6. Monitor workflow in Gitea Actions
|
||||||
|
# Wait for tests → Linux build → Android build → Release
|
||||||
|
|
||||||
|
# 7. Download artifacts and test
|
||||||
|
# Visit release page and verify downloads
|
||||||
|
```
|
||||||
|
|
||||||
|
## References
|
||||||
|
|
||||||
|
- [Tauri Documentation](https://tauri.app/)
|
||||||
|
- [Semantic Versioning](https://semver.org/)
|
||||||
|
- [GitHub Release Best Practices](https://docs.github.com/en/repositories/releasing-projects-on-github/about-releases)
|
||||||
|
- [Android App Bundle](https://developer.android.com/guide/app-bundle)
|
||||||
|
- [AppImage Documentation](https://docs.appimage.org/)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
**Last Updated:** 2026-02-13
|
||||||
1327
docs/TRACEABILITY.md
Normal file
1327
docs/TRACEABILITY.md
Normal file
File diff suppressed because it is too large
Load Diff
288
docs/TRACEABILITY_CI.md
Normal file
288
docs/TRACEABILITY_CI.md
Normal file
@ -0,0 +1,288 @@
|
|||||||
|
# Requirement Traceability CI/CD Pipeline
|
||||||
|
|
||||||
|
This document explains the automated requirement traceability validation system for JellyTau.
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
The CI/CD pipeline automatically validates that code changes are properly traced to requirements. This ensures:
|
||||||
|
- ✅ Requirements are implemented with clear traceability
|
||||||
|
- ✅ No requirement coverage regressions
|
||||||
|
- ✅ Code changes are linked to specific requirements
|
||||||
|
- ✅ Quality metrics are tracked over time
|
||||||
|
|
||||||
|
## Gitea Actions Workflows
|
||||||
|
|
||||||
|
Two workflows are configured in `.gitea/workflows/`:
|
||||||
|
|
||||||
|
### 1. `traceability-check.yml` (Primary - Recommended)
|
||||||
|
Gitea-native workflow with:
|
||||||
|
- ✅ Automatic trace extraction
|
||||||
|
- ✅ Coverage validation against minimum threshold (50%)
|
||||||
|
- ✅ Modified file checking
|
||||||
|
- ✅ Artifact preservation
|
||||||
|
- ✅ Summary reports
|
||||||
|
|
||||||
|
**Runs on:** Every push and pull request
|
||||||
|
|
||||||
|
### 2. `traceability.yml` (Alternative)
|
||||||
|
GitHub-compatible workflow with additional features:
|
||||||
|
- Pull request comments with coverage stats
|
||||||
|
- GitHub-specific integrations
|
||||||
|
|
||||||
|
## What Gets Validated
|
||||||
|
|
||||||
|
### 1. Trace Extraction
|
||||||
|
```bash
|
||||||
|
bun run traces:json > traces-report.json
|
||||||
|
```
|
||||||
|
Extracts all TRACES comments from:
|
||||||
|
- TypeScript files (`src/**/*.ts`)
|
||||||
|
- Svelte components (`src/**/*.svelte`)
|
||||||
|
- Rust code (`src-tauri/src/**/*.rs`)
|
||||||
|
- Test files
|
||||||
|
|
||||||
|
### 2. Coverage Thresholds
|
||||||
|
The workflow checks:
|
||||||
|
- **Minimum overall coverage:** 50% (57+ requirements traced)
|
||||||
|
- **Requirements by type:**
|
||||||
|
- UR (User): 23+ of 39
|
||||||
|
- IR (Integration): 5+ of 24
|
||||||
|
- DR (Development): 28+ of 48
|
||||||
|
- JA (Jellyfin API): 0+ of 3
|
||||||
|
|
||||||
|
If coverage drops below threshold, the workflow **fails** and blocks merge.
|
||||||
|
|
||||||
|
### 3. Modified File Checking
|
||||||
|
On pull requests, the workflow:
|
||||||
|
1. Detects all changed TypeScript/Svelte/Rust files
|
||||||
|
2. Warns if new/modified files lack TRACES comments
|
||||||
|
3. Suggests the TRACES format for missing comments
|
||||||
|
|
||||||
|
## How to Add Traces to New Code
|
||||||
|
|
||||||
|
When you add new code or modify existing code, include TRACES comments:
|
||||||
|
|
||||||
|
### TypeScript/Svelte Example
|
||||||
|
```typescript
|
||||||
|
// TRACES: UR-005, UR-026 | DR-029
|
||||||
|
export function handlePlayback() {
|
||||||
|
// Implementation...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Rust Example
|
||||||
|
```rust
|
||||||
|
/// TRACES: UR-005 | DR-001
|
||||||
|
pub fn player_state_changed(state: PlayerState) {
|
||||||
|
// Implementation...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Test Example
|
||||||
|
```rust
|
||||||
|
// TRACES: UR-005 | DR-001 | UT-026, UT-027
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
// Tests...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## TRACES Format
|
||||||
|
|
||||||
|
```
|
||||||
|
TRACES: [UR-###, ...] | [IR-###, ...] | [DR-###, ...] | [JA-###, ...]
|
||||||
|
```
|
||||||
|
|
||||||
|
- `UR-###` - User Requirements (features users see)
|
||||||
|
- `IR-###` - Integration Requirements (API/platform integration)
|
||||||
|
- `DR-###` - Development Requirements (internal architecture)
|
||||||
|
- `JA-###` - Jellyfin API Requirements (Jellyfin API usage)
|
||||||
|
|
||||||
|
**Examples:**
|
||||||
|
- `// TRACES: UR-005` - Single requirement
|
||||||
|
- `// TRACES: UR-005, UR-026` - Multiple of same type
|
||||||
|
- `// TRACES: UR-005 | DR-029` - Multiple types
|
||||||
|
- `// TRACES: UR-005, UR-026 | DR-001, DR-029 | UT-001` - Complex
|
||||||
|
|
||||||
|
## Workflow Behavior
|
||||||
|
|
||||||
|
### On Push to Main Branch
|
||||||
|
1. ✅ Extracts all traces from code
|
||||||
|
2. ✅ Validates coverage is >= 50%
|
||||||
|
3. ✅ Generates full traceability report
|
||||||
|
4. ✅ Saves report as artifact
|
||||||
|
|
||||||
|
### On Pull Request
|
||||||
|
1. ✅ Extracts all traces
|
||||||
|
2. ✅ Validates coverage >= 50%
|
||||||
|
3. ✅ Checks modified files for TRACES
|
||||||
|
4. ✅ Warns if new code lacks TRACES
|
||||||
|
5. ✅ Suggests proper format
|
||||||
|
6. ✅ Generates report artifact
|
||||||
|
|
||||||
|
### Failure Scenarios
|
||||||
|
The workflow **fails** (blocks merge) if:
|
||||||
|
- Coverage drops below 50%
|
||||||
|
- JSON extraction fails
|
||||||
|
- Invalid trace format
|
||||||
|
|
||||||
|
The workflow **warns** (but doesn't block) if:
|
||||||
|
- New files lack TRACES comments
|
||||||
|
- Coverage drops (but still above threshold)
|
||||||
|
|
||||||
|
## Viewing Reports
|
||||||
|
|
||||||
|
### In Gitea Actions UI
|
||||||
|
1. Go to **Actions** tab
|
||||||
|
2. Click the **Traceability Validation** workflow run
|
||||||
|
3. Download **traceability-reports** artifact
|
||||||
|
4. View:
|
||||||
|
- `traces-report.json` - Raw trace data
|
||||||
|
- `docs/TRACEABILITY.md` - Formatted report
|
||||||
|
|
||||||
|
### Locally
|
||||||
|
```bash
|
||||||
|
# Extract current traces
|
||||||
|
bun run traces:json | jq '.byType'
|
||||||
|
|
||||||
|
# Generate full report
|
||||||
|
bun run traces:markdown
|
||||||
|
cat docs/TRACEABILITY.md
|
||||||
|
```
|
||||||
|
|
||||||
|
## Coverage Goals
|
||||||
|
|
||||||
|
### Current Status
|
||||||
|
- Overall: 51% (56/114)
|
||||||
|
- UR: 59% (23/39)
|
||||||
|
- IR: 21% (5/24)
|
||||||
|
- DR: 58% (28/48)
|
||||||
|
- JA: 0% (0/3)
|
||||||
|
|
||||||
|
### Targets
|
||||||
|
- **Short term** (Sprint): Maintain ≥50% overall
|
||||||
|
- **Medium term** (Month): Reach 70% overall coverage
|
||||||
|
- **Long term** (Release): Reach 90% coverage with focus on:
|
||||||
|
- IR requirements (API clients)
|
||||||
|
- JA requirements (Jellyfin API endpoints)
|
||||||
|
- Remaining UR/DR requirements
|
||||||
|
|
||||||
|
## Improving Coverage
|
||||||
|
|
||||||
|
### For Missing User Requirements (UR)
|
||||||
|
1. Review [README.md](../README.md) for unimplemented features
|
||||||
|
2. Add TRACES to code that implements them
|
||||||
|
3. Focus on high-priority features (High/Medium priority)
|
||||||
|
|
||||||
|
### For Missing Integration Requirements (IR)
|
||||||
|
1. Add TRACES to Jellyfin API client methods
|
||||||
|
2. Add TRACES to platform-specific backends (Android/Linux)
|
||||||
|
3. Link to corresponding Jellyfin API endpoints
|
||||||
|
|
||||||
|
### For Missing Development Requirements (DR)
|
||||||
|
1. Add TRACES to UI components in `src/lib/components/`
|
||||||
|
2. Add TRACES to composables in `src/lib/composables/`
|
||||||
|
3. Add TRACES to player backend in `src-tauri/src/player/`
|
||||||
|
|
||||||
|
### For Jellyfin API Requirements (JA)
|
||||||
|
1. Add TRACES to Jellyfin API wrapper methods
|
||||||
|
2. Document which endpoints map to which requirements
|
||||||
|
3. Link to Jellyfin API documentation
|
||||||
|
|
||||||
|
## Example PR Checklist
|
||||||
|
|
||||||
|
When submitting a pull request:
|
||||||
|
|
||||||
|
- [ ] All new code has TRACES comments linking to requirements
|
||||||
|
- [ ] TRACES format is correct: `// TRACES: UR-001 | DR-002`
|
||||||
|
- [ ] Workflow passes (coverage ≥ 50%)
|
||||||
|
- [ ] No coverage regressions
|
||||||
|
- [ ] Artifact traceability report was generated
|
||||||
|
|
||||||
|
## Troubleshooting
|
||||||
|
|
||||||
|
### "Coverage below minimum threshold"
|
||||||
|
**Problem:** Workflow fails with coverage < 50%
|
||||||
|
|
||||||
|
**Solution:**
|
||||||
|
1. Run `bun run traces:json` locally
|
||||||
|
2. Check which requirements are traced
|
||||||
|
3. Add TRACES to untraced code sections
|
||||||
|
4. Re-run extraction to verify
|
||||||
|
|
||||||
|
### "New files without TRACES"
|
||||||
|
**Problem:** Workflow warns about new files lacking TRACES
|
||||||
|
|
||||||
|
**Solution:**
|
||||||
|
1. Add TRACES comments to all new code
|
||||||
|
2. Format: `// TRACES: UR-001 | DR-002`
|
||||||
|
3. Map code to specific requirements from README.md
|
||||||
|
4. Re-push
|
||||||
|
|
||||||
|
### "Invalid JSON format"
|
||||||
|
**Problem:** Trace extraction produces invalid JSON
|
||||||
|
|
||||||
|
**Solution:**
|
||||||
|
1. Check for malformed TRACES comments
|
||||||
|
2. Run locally: `bun run traces:json`
|
||||||
|
3. Look for parsing errors
|
||||||
|
4. Fix and retry
|
||||||
|
|
||||||
|
## Integration with Development
|
||||||
|
|
||||||
|
### Before Committing
|
||||||
|
```bash
|
||||||
|
# Check your traces
|
||||||
|
bun run traces:json | jq '.byType'
|
||||||
|
|
||||||
|
# Regenerate report
|
||||||
|
bun run traces:markdown
|
||||||
|
|
||||||
|
# Verify traces syntax
|
||||||
|
grep "TRACES:" src/**/*.ts src/**/*.rs
|
||||||
|
```
|
||||||
|
|
||||||
|
### In Your IDE
|
||||||
|
Add a file watcher to regenerate traces on save:
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"fileWatcher.watchPatterns": [
|
||||||
|
"src/**/*.ts",
|
||||||
|
"src/**/*.svelte",
|
||||||
|
"src-tauri/src/**/*.rs"
|
||||||
|
],
|
||||||
|
"fileWatcher.command": "bun run traces:markdown"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Git Hooks
|
||||||
|
Add a pre-push hook to validate traces:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
#!/bin/bash
|
||||||
|
# .git/hooks/pre-push
|
||||||
|
bun run traces:json > /dev/null
|
||||||
|
if [ $? -ne 0 ]; then
|
||||||
|
echo "❌ Invalid TRACES format"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
```
|
||||||
|
|
||||||
|
## References
|
||||||
|
|
||||||
|
- [Extract Traces Script](../scripts/README.md#extract-tracests)
|
||||||
|
- [Requirements Specification](../README.md#requirements-specification)
|
||||||
|
- [Traceability Matrix](./TRACEABILITY.md)
|
||||||
|
- [Gitea Actions Documentation](https://docs.gitea.io/en-us/actions/)
|
||||||
|
|
||||||
|
## Support
|
||||||
|
|
||||||
|
For issues or questions:
|
||||||
|
1. Check this document
|
||||||
|
2. Review example traces in `src/lib/stores/`
|
||||||
|
3. Check existing TRACES comments for format
|
||||||
|
4. Review workflow logs in Gitea Actions
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
**Last Updated:** 2026-02-13
|
||||||
@ -23,7 +23,10 @@
|
|||||||
"android:check": "./scripts/check-android.sh",
|
"android:check": "./scripts/check-android.sh",
|
||||||
"android:logs": "./scripts/logcat.sh",
|
"android:logs": "./scripts/logcat.sh",
|
||||||
"clean": "./scripts/clean.sh",
|
"clean": "./scripts/clean.sh",
|
||||||
"tauri": "tauri"
|
"tauri": "tauri",
|
||||||
|
"traces": "bun run scripts/extract-traces.ts",
|
||||||
|
"traces:json": "bun run scripts/extract-traces.ts --format json",
|
||||||
|
"traces:markdown": "bun run scripts/extract-traces.ts --format markdown > docs/TRACEABILITY.md"
|
||||||
},
|
},
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
|||||||
@ -60,6 +60,42 @@ View Android logcat filtered for the app.
|
|||||||
./scripts/logcat.sh
|
./scripts/logcat.sh
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Traceability & Documentation
|
||||||
|
|
||||||
|
### `extract-traces.ts`
|
||||||
|
Extract requirement IDs (TRACES) from source code and generate a traceability matrix mapping requirements to implementation locations.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
bun run traces # Generate markdown report
|
||||||
|
bun run traces:json # Generate JSON report
|
||||||
|
bun run traces:markdown # Save to docs/TRACEABILITY.md
|
||||||
|
```
|
||||||
|
|
||||||
|
The script scans all TypeScript, Svelte, and Rust files looking for `TRACES:` comments and generates a comprehensive mapping of:
|
||||||
|
- Which code files implement which requirements
|
||||||
|
- Line numbers and code context
|
||||||
|
- Coverage summary by requirement type (UR, IR, DR, JA)
|
||||||
|
|
||||||
|
Example TRACES comment in code:
|
||||||
|
```typescript
|
||||||
|
// TRACES: UR-005, UR-026 | DR-029
|
||||||
|
function handlePlayback() { ... }
|
||||||
|
```
|
||||||
|
|
||||||
|
See [docs/TRACEABILITY.md](../docs/TRACEABILITY.md) for the latest generated mapping.
|
||||||
|
|
||||||
|
### CI/CD Validation
|
||||||
|
|
||||||
|
The traceability system is integrated with Gitea Actions CI/CD:
|
||||||
|
- Automatically validates TRACES on every push and pull request
|
||||||
|
- Enforces minimum 50% coverage threshold
|
||||||
|
- Warns if new code lacks TRACES comments
|
||||||
|
- Generates traceability reports automatically
|
||||||
|
|
||||||
|
For details, see:
|
||||||
|
- [Traceability CI Guide](../docs/TRACEABILITY_CI.md) - Full CI/CD documentation
|
||||||
|
- [TRACES Quick Reference](../TRACES_QUICK_REF.md) - Quick guide for adding TRACES
|
||||||
|
|
||||||
## Utility Scripts
|
## Utility Scripts
|
||||||
|
|
||||||
### `clean.sh`
|
### `clean.sh`
|
||||||
|
|||||||
@ -10,9 +10,7 @@
|
|||||||
|
|
||||||
import * as fs from "fs";
|
import * as fs from "fs";
|
||||||
import * as path from "path";
|
import * as path from "path";
|
||||||
|
import { execSync } from "child_process";
|
||||||
// Use built-in Bun.glob
|
|
||||||
const glob = (pattern: string) => new Bun.Glob(pattern);
|
|
||||||
|
|
||||||
interface TraceEntry {
|
interface TraceEntry {
|
||||||
file: string;
|
file: string;
|
||||||
@ -46,21 +44,54 @@ function extractRequirementIds(tracesString: string): string[] {
|
|||||||
return matches.map((m) => `${m[1]}-${m[2]}`);
|
return matches.map((m) => `${m[1]}-${m[2]}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
function getContext(content: string, lineNum: number): string {
|
function getAllSourceFiles(): string[] {
|
||||||
const lines = content.split("\n");
|
const baseDir = "/home/dtourolle/Development/JellyTau";
|
||||||
const contextStart = Math.max(0, lineNum - 3);
|
const patterns = ["src", "src-tauri/src"];
|
||||||
const contextEnd = Math.min(lines.length, lineNum + 1);
|
const files: string[] = [];
|
||||||
const contextLines = lines.slice(contextStart, contextEnd);
|
|
||||||
return contextLines.join("\n").trim();
|
function walkDir(dir: string) {
|
||||||
|
try {
|
||||||
|
const entries = fs.readdirSync(dir, { withFileTypes: true });
|
||||||
|
for (const entry of entries) {
|
||||||
|
const fullPath = path.join(dir, entry.name);
|
||||||
|
const relativePath = path.relative(baseDir, fullPath);
|
||||||
|
|
||||||
|
// Skip node_modules, target, build
|
||||||
|
if (
|
||||||
|
relativePath.includes("node_modules") ||
|
||||||
|
relativePath.includes("target") ||
|
||||||
|
relativePath.includes("build") ||
|
||||||
|
relativePath.includes(".git")
|
||||||
|
) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (entry.isDirectory()) {
|
||||||
|
walkDir(fullPath);
|
||||||
|
} else if (
|
||||||
|
entry.name.endsWith(".ts") ||
|
||||||
|
entry.name.endsWith(".svelte") ||
|
||||||
|
entry.name.endsWith(".rs")
|
||||||
|
) {
|
||||||
|
files.push(fullPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
// Skip directories we can't read
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const pattern of patterns) {
|
||||||
|
const dir = path.join(baseDir, pattern);
|
||||||
|
if (fs.existsSync(dir)) {
|
||||||
|
walkDir(dir);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return files;
|
||||||
}
|
}
|
||||||
|
|
||||||
async function extractTraces(): Promise<TracesData> {
|
function extractTraces(): TracesData {
|
||||||
const patterns = [
|
|
||||||
"src/**/*.ts",
|
|
||||||
"src/**/*.svelte",
|
|
||||||
"src-tauri/src/**/*.rs",
|
|
||||||
];
|
|
||||||
|
|
||||||
const requirementMap: RequirementMapping = {};
|
const requirementMap: RequirementMapping = {};
|
||||||
const byType: Record<string, Set<string>> = {
|
const byType: Record<string, Set<string>> = {
|
||||||
UR: new Set(),
|
UR: new Set(),
|
||||||
@ -70,25 +101,15 @@ async function extractTraces(): Promise<TracesData> {
|
|||||||
};
|
};
|
||||||
|
|
||||||
let totalTraces = 0;
|
let totalTraces = 0;
|
||||||
const processedFiles = new Set<string>();
|
const baseDir = "/home/dtourolle/Development/JellyTau";
|
||||||
|
|
||||||
for (const pattern of patterns) {
|
const files = getAllSourceFiles();
|
||||||
const globber = glob(pattern);
|
|
||||||
const files = [];
|
|
||||||
for await (const file of globber.scan({
|
|
||||||
cwd: "/home/dtourolle/Development/JellyTau",
|
|
||||||
})) {
|
|
||||||
files.push(file);
|
|
||||||
}
|
|
||||||
|
|
||||||
for (const file of files) {
|
|
||||||
if (processedFiles.has(file)) continue;
|
|
||||||
processedFiles.add(file);
|
|
||||||
|
|
||||||
|
for (const fullPath of files) {
|
||||||
try {
|
try {
|
||||||
const fullPath = `/home/dtourolle/Development/JellyTau/${file}`;
|
|
||||||
const content = fs.readFileSync(fullPath, "utf-8");
|
const content = fs.readFileSync(fullPath, "utf-8");
|
||||||
const lines = content.split("\n");
|
const lines = content.split("\n");
|
||||||
|
const relativePath = path.relative(baseDir, fullPath);
|
||||||
|
|
||||||
let match;
|
let match;
|
||||||
TRACES_PATTERN.lastIndex = 0;
|
TRACES_PATTERN.lastIndex = 0;
|
||||||
@ -115,15 +136,19 @@ async function extractTraces(): Promise<TracesData> {
|
|||||||
line.includes("pub struct ") ||
|
line.includes("pub struct ") ||
|
||||||
line.includes("impl ") ||
|
line.includes("impl ") ||
|
||||||
line.includes("async function ") ||
|
line.includes("async function ") ||
|
||||||
line.includes("class ")
|
line.includes("class ") ||
|
||||||
|
line.includes("export type ")
|
||||||
) {
|
) {
|
||||||
context = line.trim();
|
context = line
|
||||||
|
.trim()
|
||||||
|
.replace(/^\s*\/\/\s*/, "")
|
||||||
|
.replace(/^\s*\/\*\*\s*/, "");
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const entry: TraceEntry = {
|
const entry: TraceEntry = {
|
||||||
file: file.replace(/^\//, ""),
|
file: relativePath,
|
||||||
line: lineNum + 1,
|
line: lineNum + 1,
|
||||||
context,
|
context,
|
||||||
requirements: reqIds,
|
requirements: reqIds,
|
||||||
@ -145,14 +170,13 @@ async function extractTraces(): Promise<TracesData> {
|
|||||||
totalTraces++;
|
totalTraces++;
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error(`Error processing ${file}:`, error);
|
// Skip files we can't read
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return {
|
return {
|
||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
totalFiles: processedFiles.size,
|
totalFiles: files.length,
|
||||||
totalTraces,
|
totalTraces,
|
||||||
requirements: requirementMap,
|
requirements: requirementMap,
|
||||||
byType: {
|
byType: {
|
||||||
@ -224,7 +248,8 @@ ${data.byType.JA.join(", ")}
|
|||||||
for (const entry of entries) {
|
for (const entry of entries) {
|
||||||
md += `- **File:** [\`${entry.file}\`](${entry.file}#L${entry.line})\n`;
|
md += `- **File:** [\`${entry.file}\`](${entry.file}#L${entry.line})\n`;
|
||||||
md += ` - **Line:** ${entry.line}\n`;
|
md += ` - **Line:** ${entry.line}\n`;
|
||||||
md += ` - **Context:** \`${entry.context.substring(0, 80)}...\`\n`;
|
const contextPreview = entry.context.substring(0, 70);
|
||||||
|
md += ` - **Context:** \`${contextPreview}${entry.context.length > 70 ? "..." : ""}\`\n`;
|
||||||
}
|
}
|
||||||
md += "\n";
|
md += "\n";
|
||||||
}
|
}
|
||||||
@ -242,8 +267,8 @@ const format = args.includes("--format")
|
|||||||
? args[args.indexOf("--format") + 1]
|
? args[args.indexOf("--format") + 1]
|
||||||
: "markdown";
|
: "markdown";
|
||||||
|
|
||||||
console.error("Extracting TRACES from codebase...");
|
console.error("🔍 Extracting TRACES from codebase...");
|
||||||
const data = await extractTraces();
|
const data = extractTraces();
|
||||||
|
|
||||||
if (format === "json") {
|
if (format === "json") {
|
||||||
console.log(generateJson(data));
|
console.log(generateJson(data));
|
||||||
@ -251,4 +276,6 @@ if (format === "json") {
|
|||||||
console.log(generateMarkdown(data));
|
console.log(generateMarkdown(data));
|
||||||
}
|
}
|
||||||
|
|
||||||
console.error(`\n✅ Complete! Found ${data.totalTraces} TRACES across ${data.totalFiles} files`);
|
console.error(
|
||||||
|
`\n✅ Complete! Found ${data.totalTraces} TRACES across ${data.totalFiles} files`
|
||||||
|
);
|
||||||
|
|||||||
128
src-tauri/src/commands/device.rs
Normal file
128
src-tauri/src/commands/device.rs
Normal file
@ -0,0 +1,128 @@
|
|||||||
|
//! Device identification commands
|
||||||
|
//!
|
||||||
|
//! Handles persistent device ID generation and retrieval for Jellyfin server communication.
|
||||||
|
//! TRACES: UR-009 | DR-011
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use log::info;
|
||||||
|
use tauri::State;
|
||||||
|
use uuid::Uuid;
|
||||||
|
|
||||||
|
use crate::commands::storage::DatabaseWrapper;
|
||||||
|
use crate::storage::db_service::{DatabaseService, Query, QueryParam};
|
||||||
|
|
||||||
|
/// Get or create the device ID.
|
||||||
|
/// Device ID is a UUID v4 that persists across app restarts.
|
||||||
|
/// On first call, generates and stores a new UUID.
|
||||||
|
/// On subsequent calls, retrieves the stored UUID.
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
/// - `Ok(String)` - The device ID (UUID v4)
|
||||||
|
/// - `Err(String)` - If database operation fails
|
||||||
|
///
|
||||||
|
/// TRACES: UR-009 | DR-011
|
||||||
|
#[tauri::command]
|
||||||
|
pub async fn device_get_id(db: State<'_, DatabaseWrapper>) -> Result<String, String> {
|
||||||
|
let db_service = {
|
||||||
|
let database = db.0.lock().map_err(|e| e.to_string())?;
|
||||||
|
Arc::new(database.service())
|
||||||
|
};
|
||||||
|
|
||||||
|
// Try to get existing device ID from database
|
||||||
|
let query = Query::with_params(
|
||||||
|
"SELECT value FROM app_settings WHERE key = ?",
|
||||||
|
vec![QueryParam::String("device_id".to_string())],
|
||||||
|
);
|
||||||
|
|
||||||
|
let existing_id: Option<String> = db_service
|
||||||
|
.query_one(query, |row| row.get(0))
|
||||||
|
.await
|
||||||
|
.ok()
|
||||||
|
.flatten();
|
||||||
|
|
||||||
|
if let Some(device_id) = existing_id {
|
||||||
|
info!("[Device] Retrieved existing device ID");
|
||||||
|
return Ok(device_id);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Generate new device ID
|
||||||
|
let device_id = Uuid::new_v4().to_string();
|
||||||
|
|
||||||
|
// Store it in database
|
||||||
|
let insert_query = Query::with_params(
|
||||||
|
"INSERT INTO app_settings (key, value) VALUES (?, ?)",
|
||||||
|
vec![
|
||||||
|
QueryParam::String("device_id".to_string()),
|
||||||
|
QueryParam::String(device_id.clone()),
|
||||||
|
],
|
||||||
|
);
|
||||||
|
|
||||||
|
db_service
|
||||||
|
.execute(insert_query)
|
||||||
|
.await
|
||||||
|
.map_err(|e| e.to_string())?;
|
||||||
|
|
||||||
|
info!("[Device] Generated and stored new device ID");
|
||||||
|
Ok(device_id)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Set the device ID (primarily for testing or recovery).
|
||||||
|
/// Overwrites any existing device ID.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
/// * `device_id` - The device ID to store (should be UUID v4 format)
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
/// - `Ok(())` - If device ID was stored successfully
|
||||||
|
/// - `Err(String)` - If database operation fails
|
||||||
|
///
|
||||||
|
/// TRACES: UR-009 | DR-011
|
||||||
|
#[tauri::command]
|
||||||
|
pub async fn device_set_id(device_id: String, db: State<'_, DatabaseWrapper>) -> Result<(), String> {
|
||||||
|
let db_service = {
|
||||||
|
let database = db.0.lock().map_err(|e| e.to_string())?;
|
||||||
|
Arc::new(database.service())
|
||||||
|
};
|
||||||
|
|
||||||
|
let query = Query::with_params(
|
||||||
|
"INSERT OR REPLACE INTO app_settings (key, value) VALUES (?, ?)",
|
||||||
|
vec![
|
||||||
|
QueryParam::String("device_id".to_string()),
|
||||||
|
QueryParam::String(device_id),
|
||||||
|
],
|
||||||
|
);
|
||||||
|
|
||||||
|
db_service.execute(query).await.map_err(|e| e.to_string())?;
|
||||||
|
|
||||||
|
info!("[Device] Device ID set");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_device_id_is_valid_uuid() {
|
||||||
|
let id = Uuid::new_v4().to_string();
|
||||||
|
// Should parse as UUID
|
||||||
|
let parsed = Uuid::parse_str(&id);
|
||||||
|
assert!(parsed.is_ok(), "Device ID should be a valid UUID");
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_device_id_format() {
|
||||||
|
let id = Uuid::new_v4().to_string();
|
||||||
|
// UUID v4 format: xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx (36 chars with hyphens)
|
||||||
|
assert_eq!(id.len(), 36, "Device ID should be 36 characters");
|
||||||
|
assert!(id.contains('-'), "Device ID should contain hyphens");
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_device_ids_are_unique() {
|
||||||
|
let id1 = Uuid::new_v4().to_string();
|
||||||
|
let id2 = Uuid::new_v4().to_string();
|
||||||
|
assert_ne!(id1, id2, "Generated device IDs should be unique");
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -1530,6 +1530,7 @@ pub fn get_album_affinity_status(
|
|||||||
Ok(statuses)
|
Ok(statuses)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-011, UR-018 | DR-015, DR-018 | UT-042, UT-043
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -1,6 +1,10 @@
|
|||||||
|
// Tauri commands exposed to frontend
|
||||||
|
// TRACES: UR-002, UR-003, UR-004, UR-005, UR-009, UR-011, UR-012, UR-017, UR-019, UR-025 |
|
||||||
|
// DR-015, DR-017, DR-021, DR-028
|
||||||
pub mod auth;
|
pub mod auth;
|
||||||
pub mod connectivity;
|
pub mod connectivity;
|
||||||
pub mod conversions;
|
pub mod conversions;
|
||||||
|
pub mod device;
|
||||||
pub mod download;
|
pub mod download;
|
||||||
pub mod offline;
|
pub mod offline;
|
||||||
pub mod playback_mode;
|
pub mod playback_mode;
|
||||||
@ -14,6 +18,7 @@ pub mod sync;
|
|||||||
pub use auth::*;
|
pub use auth::*;
|
||||||
pub use connectivity::*;
|
pub use connectivity::*;
|
||||||
pub use conversions::*;
|
pub use conversions::*;
|
||||||
|
pub use device::*;
|
||||||
pub use download::*;
|
pub use download::*;
|
||||||
pub use offline::*;
|
pub use offline::*;
|
||||||
pub use playback_mode::*;
|
pub use playback_mode::*;
|
||||||
|
|||||||
@ -130,6 +130,7 @@ pub async fn offline_search(
|
|||||||
.map_err(|e| e.to_string())
|
.map_err(|e| e.to_string())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-002, UR-011 | DR-017 | UT-044
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -367,6 +367,33 @@ pub fn repository_get_image_url(
|
|||||||
Ok(repo.as_ref().get_image_url(&item_id, image_type, options))
|
Ok(repo.as_ref().get_image_url(&item_id, image_type, options))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get subtitle URL for a media item
|
||||||
|
#[tauri::command]
|
||||||
|
pub fn repository_get_subtitle_url(
|
||||||
|
manager: State<'_, RepositoryManagerWrapper>,
|
||||||
|
handle: String,
|
||||||
|
item_id: String,
|
||||||
|
media_source_id: String,
|
||||||
|
stream_index: i32,
|
||||||
|
format: String,
|
||||||
|
) -> Result<String, String> {
|
||||||
|
let repo = manager.0.get(&handle).ok_or("Repository not found")?;
|
||||||
|
Ok(repo.as_ref().get_subtitle_url(&item_id, &media_source_id, stream_index, &format))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get video download URL with quality preset
|
||||||
|
#[tauri::command]
|
||||||
|
pub fn repository_get_video_download_url(
|
||||||
|
manager: State<'_, RepositoryManagerWrapper>,
|
||||||
|
handle: String,
|
||||||
|
item_id: String,
|
||||||
|
quality: String,
|
||||||
|
media_source_id: Option<String>,
|
||||||
|
) -> Result<String, String> {
|
||||||
|
let repo = manager.0.get(&handle).ok_or("Repository not found")?;
|
||||||
|
Ok(repo.as_ref().get_video_download_url(&item_id, &quality, media_source_id.as_deref()))
|
||||||
|
}
|
||||||
|
|
||||||
/// Mark an item as favorite
|
/// Mark an item as favorite
|
||||||
#[tauri::command]
|
#[tauri::command]
|
||||||
pub async fn repository_mark_favorite(
|
pub async fn repository_mark_favorite(
|
||||||
|
|||||||
@ -2,6 +2,7 @@
|
|||||||
//!
|
//!
|
||||||
//! The sync queue stores mutations (favorites, playback progress, etc.)
|
//! The sync queue stores mutations (favorites, playback progress, etc.)
|
||||||
//! that need to be synced to the Jellyfin server when connectivity is restored.
|
//! that need to be synced to the Jellyfin server when connectivity is restored.
|
||||||
|
//! TRACES: UR-002, UR-017, UR-025 | DR-014
|
||||||
|
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|||||||
@ -65,6 +65,8 @@ use commands::{
|
|||||||
auth_initialize, auth_connect_to_server, auth_login, auth_verify_session,
|
auth_initialize, auth_connect_to_server, auth_login, auth_verify_session,
|
||||||
auth_logout, auth_get_session, auth_set_session, auth_start_verification,
|
auth_logout, auth_get_session, auth_set_session, auth_start_verification,
|
||||||
auth_stop_verification, auth_reauthenticate,
|
auth_stop_verification, auth_reauthenticate,
|
||||||
|
// Device commands
|
||||||
|
device_get_id, device_set_id,
|
||||||
// Connectivity commands
|
// Connectivity commands
|
||||||
connectivity_check_server, connectivity_set_server_url, connectivity_get_status,
|
connectivity_check_server, connectivity_set_server_url, connectivity_get_status,
|
||||||
connectivity_start_monitoring, connectivity_stop_monitoring,
|
connectivity_start_monitoring, connectivity_stop_monitoring,
|
||||||
@ -642,6 +644,9 @@ pub fn run() {
|
|||||||
auth_start_verification,
|
auth_start_verification,
|
||||||
auth_stop_verification,
|
auth_stop_verification,
|
||||||
auth_reauthenticate,
|
auth_reauthenticate,
|
||||||
|
// Device commands
|
||||||
|
device_get_id,
|
||||||
|
device_set_id,
|
||||||
// Connectivity commands
|
// Connectivity commands
|
||||||
connectivity_check_server,
|
connectivity_check_server,
|
||||||
connectivity_set_server_url,
|
connectivity_set_server_url,
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Autoplay decision logic
|
||||||
|
// TRACES: UR-023, UR-026 | DR-047, DR-048, DR-029
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use crate::repository::types::MediaItem;
|
use crate::repository::types::MediaItem;
|
||||||
|
|
||||||
|
|||||||
@ -232,14 +232,13 @@ impl PlayerBackend for NullBackend {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-003, UR-004 | DR-004 | UT-026, UT-027, UT-028, UT-029, UT-030, UT-031, UT-032, UT-033
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
/// Test NullBackend volume default value
|
/// Test NullBackend volume default value
|
||||||
///
|
/// TRACES: UR-016 | DR-004 | UT-026
|
||||||
/// @req-test: UT-026 - NullBackend volume default value
|
|
||||||
/// @req-test: DR-004 - PlayerBackend trait
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_null_backend_volume_default() {
|
fn test_null_backend_volume_default() {
|
||||||
let backend = NullBackend::new();
|
let backend = NullBackend::new();
|
||||||
|
|||||||
@ -1,3 +1,7 @@
|
|||||||
|
// Player module - Complete playback control system
|
||||||
|
// TRACES: UR-003, UR-004, UR-005, UR-019, UR-023, UR-026 |
|
||||||
|
// IR-003, IR-004, IR-006, IR-008 |
|
||||||
|
// DR-001, DR-004, DR-005, DR-009, DR-028, DR-029, DR-047
|
||||||
pub mod autoplay;
|
pub mod autoplay;
|
||||||
pub mod backend;
|
pub mod backend;
|
||||||
pub mod events;
|
pub mod events;
|
||||||
|
|||||||
@ -4,6 +4,8 @@
|
|||||||
/// - Tokio runtime panics when spawning async tasks from std::thread
|
/// - Tokio runtime panics when spawning async tasks from std::thread
|
||||||
/// - Position update thread failures
|
/// - Position update thread failures
|
||||||
/// - Event emission issues
|
/// - Event emission issues
|
||||||
|
///
|
||||||
|
/// TRACES: UR-003, UR-004 | IR-003 | IT-003, IT-004
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
|
|||||||
@ -505,6 +505,7 @@ pub enum AddPosition {
|
|||||||
End,
|
End,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-005, UR-015 | DR-005 | UT-003, UT-004, UT-005
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -78,6 +78,7 @@ impl SleepTimerState {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-026 | DR-029 | UT-012
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -326,6 +326,27 @@ impl MediaRepository for HybridRepository {
|
|||||||
self.online.get_image_url(item_id, image_type, options)
|
self.online.get_image_url(item_id, image_type, options)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
media_source_id: &str,
|
||||||
|
stream_index: i32,
|
||||||
|
format: &str,
|
||||||
|
) -> String {
|
||||||
|
// Always use online URL for subtitles
|
||||||
|
self.online.get_subtitle_url(item_id, media_source_id, stream_index, format)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
quality: &str,
|
||||||
|
media_source_id: Option<&str>,
|
||||||
|
) -> String {
|
||||||
|
// Always use online URL for downloads
|
||||||
|
self.online.get_video_download_url(item_id, quality, media_source_id)
|
||||||
|
}
|
||||||
|
|
||||||
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError> {
|
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError> {
|
||||||
// Write operations go directly to server
|
// Write operations go directly to server
|
||||||
self.online.mark_favorite(item_id).await
|
self.online.mark_favorite(item_id).await
|
||||||
@ -497,6 +518,25 @@ mod tests {
|
|||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_media_source_id: &str,
|
||||||
|
_stream_index: i32,
|
||||||
|
_format: &str,
|
||||||
|
) -> String {
|
||||||
|
unimplemented!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_quality: &str,
|
||||||
|
_media_source_id: Option<&str>,
|
||||||
|
) -> String {
|
||||||
|
unimplemented!()
|
||||||
|
}
|
||||||
|
|
||||||
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
@ -603,6 +643,25 @@ mod tests {
|
|||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_media_source_id: &str,
|
||||||
|
_stream_index: i32,
|
||||||
|
_format: &str,
|
||||||
|
) -> String {
|
||||||
|
unimplemented!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_quality: &str,
|
||||||
|
_media_source_id: Option<&str>,
|
||||||
|
) -> String {
|
||||||
|
unimplemented!()
|
||||||
|
}
|
||||||
|
|
||||||
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
|
|||||||
@ -146,6 +146,23 @@ pub trait MediaRepository: Send + Sync {
|
|||||||
options: Option<ImageOptions>,
|
options: Option<ImageOptions>,
|
||||||
) -> String;
|
) -> String;
|
||||||
|
|
||||||
|
/// Get subtitle URL (synchronous - just constructs URL)
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
media_source_id: &str,
|
||||||
|
stream_index: i32,
|
||||||
|
format: &str,
|
||||||
|
) -> String;
|
||||||
|
|
||||||
|
/// Get video download URL (synchronous - just constructs URL)
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
quality: &str,
|
||||||
|
media_source_id: Option<&str>,
|
||||||
|
) -> String;
|
||||||
|
|
||||||
/// Mark item as favorite
|
/// Mark item as favorite
|
||||||
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError>;
|
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError>;
|
||||||
|
|
||||||
|
|||||||
@ -918,6 +918,27 @@ impl MediaRepository for OfflineRepository {
|
|||||||
format!("offline://{}/{}", item_id, type_str)
|
format!("offline://{}/{}", item_id, type_str)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_media_source_id: &str,
|
||||||
|
_stream_index: i32,
|
||||||
|
_format: &str,
|
||||||
|
) -> String {
|
||||||
|
// Subtitles not available offline
|
||||||
|
String::new()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
_item_id: &str,
|
||||||
|
_quality: &str,
|
||||||
|
_media_source_id: Option<&str>,
|
||||||
|
) -> String {
|
||||||
|
// Cannot download while offline
|
||||||
|
String::new()
|
||||||
|
}
|
||||||
|
|
||||||
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
async fn mark_favorite(&self, _item_id: &str) -> Result<(), RepoError> {
|
||||||
// Cannot update server while offline
|
// Cannot update server while offline
|
||||||
Err(RepoError::Offline)
|
Err(RepoError::Offline)
|
||||||
|
|||||||
@ -562,11 +562,15 @@ impl MediaRepository for OnlineRepository {
|
|||||||
let mut ungrouped = Vec::new();
|
let mut ungrouped = Vec::new();
|
||||||
|
|
||||||
for item in items {
|
for item in items {
|
||||||
if let Some(album_id) = &item.album_id {
|
// Use album_id if available, fall back to album_name for grouping
|
||||||
debug!("[get_recently_played_audio] Grouping item '{}' into album '{}'", item.name, album_id);
|
let group_key = item.album_id.clone()
|
||||||
album_map.entry(album_id.clone()).or_insert_with(Vec::new).push(item);
|
.or_else(|| item.album_name.clone());
|
||||||
|
|
||||||
|
if let Some(key) = group_key {
|
||||||
|
debug!("[get_recently_played_audio] Grouping item '{}' into album '{}'", item.name, key);
|
||||||
|
album_map.entry(key).or_insert_with(Vec::new).push(item);
|
||||||
} else {
|
} else {
|
||||||
debug!("[get_recently_played_audio] No album_id for item: '{}'", item.name);
|
debug!("[get_recently_played_audio] No album_id or album_name for item: '{}'", item.name);
|
||||||
ungrouped.push(item);
|
ungrouped.push(item);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1025,6 +1029,50 @@ impl MediaRepository for OnlineRepository {
|
|||||||
url
|
url
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
media_source_id: &str,
|
||||||
|
stream_index: i32,
|
||||||
|
format: &str,
|
||||||
|
) -> String {
|
||||||
|
format!(
|
||||||
|
"{}/Videos/{}/{}/Subtitles/{}/{}",
|
||||||
|
self.server_url,
|
||||||
|
item_id,
|
||||||
|
media_source_id,
|
||||||
|
stream_index,
|
||||||
|
format
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
quality: &str,
|
||||||
|
media_source_id: Option<&str>,
|
||||||
|
) -> String {
|
||||||
|
let mut url = format!("{}/Videos/{}/download", self.server_url, item_id);
|
||||||
|
let mut params = vec![format!("api_key={}", self.access_token)];
|
||||||
|
|
||||||
|
// Add quality parameter if not "original"
|
||||||
|
if quality != "original" {
|
||||||
|
params.push(format!("quality={}", quality));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add media source ID if provided
|
||||||
|
if let Some(source_id) = media_source_id {
|
||||||
|
params.push(format!("mediaSourceId={}", source_id));
|
||||||
|
}
|
||||||
|
|
||||||
|
if !params.is_empty() {
|
||||||
|
url.push('?');
|
||||||
|
url.push_str(¶ms.join("&"));
|
||||||
|
}
|
||||||
|
|
||||||
|
url
|
||||||
|
}
|
||||||
|
|
||||||
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError> {
|
async fn mark_favorite(&self, item_id: &str) -> Result<(), RepoError> {
|
||||||
let endpoint = format!("/Users/{}/FavoriteItems/{}", self.user_id, item_id);
|
let endpoint = format!("/Users/{}/FavoriteItems/{}", self.user_id, item_id);
|
||||||
self.post_json(&endpoint, &serde_json::json!({})).await
|
self.post_json(&endpoint, &serde_json::json!({})).await
|
||||||
|
|||||||
433
src-tauri/src/repository/online_integration_test.rs
Normal file
433
src-tauri/src/repository/online_integration_test.rs
Normal file
@ -0,0 +1,433 @@
|
|||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use crate::api::jellyfin::{
|
||||||
|
GetItemsOptions, ImageType, ImageOptions, SortOrder,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Mock for testing URL construction without a real server
|
||||||
|
struct MockOnlineRepository {
|
||||||
|
server_url: String,
|
||||||
|
access_token: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MockOnlineRepository {
|
||||||
|
fn new(server_url: &str, access_token: &str) -> Self {
|
||||||
|
Self {
|
||||||
|
server_url: server_url.to_string(),
|
||||||
|
access_token: access_token.to_string(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Test helper: construct image URL similar to backend
|
||||||
|
fn get_image_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
image_type: &str,
|
||||||
|
options: Option<&ImageOptions>,
|
||||||
|
) -> String {
|
||||||
|
let mut url = format!(
|
||||||
|
"{}/Items/{}/Images/{}",
|
||||||
|
self.server_url, item_id, image_type
|
||||||
|
);
|
||||||
|
|
||||||
|
let mut params = vec![("api_key", self.access_token.clone())];
|
||||||
|
|
||||||
|
if let Some(opts) = options {
|
||||||
|
if let Some(max_width) = opts.max_width {
|
||||||
|
params.push(("maxWidth", max_width.to_string()));
|
||||||
|
}
|
||||||
|
if let Some(max_height) = opts.max_height {
|
||||||
|
params.push(("maxHeight", max_height.to_string()));
|
||||||
|
}
|
||||||
|
if let Some(quality) = opts.quality {
|
||||||
|
params.push(("quality", quality.to_string()));
|
||||||
|
}
|
||||||
|
if let Some(tag) = &opts.tag {
|
||||||
|
params.push(("tag", tag.clone()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let query_string = params
|
||||||
|
.iter()
|
||||||
|
.map(|(k, v)| format!("{}={}", k, v))
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
.join("&");
|
||||||
|
|
||||||
|
if !query_string.is_empty() {
|
||||||
|
url.push('?');
|
||||||
|
url.push_str(&query_string);
|
||||||
|
}
|
||||||
|
|
||||||
|
url
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Test helper: construct subtitle URL
|
||||||
|
fn get_subtitle_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
media_source_id: &str,
|
||||||
|
stream_index: usize,
|
||||||
|
format: &str,
|
||||||
|
) -> String {
|
||||||
|
format!(
|
||||||
|
"{}/Videos/{}/Subtitles/{}/{}/subtitles.{}?api_key={}",
|
||||||
|
self.server_url,
|
||||||
|
item_id,
|
||||||
|
media_source_id,
|
||||||
|
stream_index,
|
||||||
|
format,
|
||||||
|
self.access_token
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Test helper: construct video download URL
|
||||||
|
fn get_video_download_url(
|
||||||
|
&self,
|
||||||
|
item_id: &str,
|
||||||
|
quality: &str,
|
||||||
|
) -> String {
|
||||||
|
let (max_width, bitrate) = match quality {
|
||||||
|
"1080p" => ("1920", "15000k"),
|
||||||
|
"720p" => ("1280", "8000k"),
|
||||||
|
"480p" => ("854", "3000k"),
|
||||||
|
_ => ("0", ""), // original
|
||||||
|
};
|
||||||
|
|
||||||
|
if quality == "original" {
|
||||||
|
format!("{}/Videos/{}/stream.mp4?api_key={}", self.server_url, item_id, self.access_token)
|
||||||
|
} else {
|
||||||
|
format!(
|
||||||
|
"{}/Videos/{}/stream.mp4?maxWidth={}&videoBitrate={}&api_key={}",
|
||||||
|
self.server_url, item_id, max_width, bitrate, self.access_token
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== Image URL Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_basic() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", None);
|
||||||
|
|
||||||
|
assert!(url.contains("https://jellyfin.example.com"));
|
||||||
|
assert!(url.contains("/Items/item123/Images/Primary"));
|
||||||
|
assert!(url.contains("api_key=token123"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_with_max_width() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: Some(300),
|
||||||
|
max_height: None,
|
||||||
|
quality: None,
|
||||||
|
tag: None,
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", Some(&options));
|
||||||
|
|
||||||
|
assert!(url.contains("maxWidth=300"));
|
||||||
|
assert!(url.contains("api_key=token123"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_with_all_options() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: Some(1920),
|
||||||
|
max_height: Some(1080),
|
||||||
|
quality: Some(90),
|
||||||
|
tag: Some("abc123".to_string()),
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item456", "Backdrop", Some(&options));
|
||||||
|
|
||||||
|
assert!(url.contains("/Items/item456/Images/Backdrop"));
|
||||||
|
assert!(url.contains("maxWidth=1920"));
|
||||||
|
assert!(url.contains("maxHeight=1080"));
|
||||||
|
assert!(url.contains("quality=90"));
|
||||||
|
assert!(url.contains("tag=abc123"));
|
||||||
|
assert!(url.contains("api_key=token123"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_different_image_types() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let image_types = vec!["Primary", "Backdrop", "Logo", "Thumb"];
|
||||||
|
|
||||||
|
for image_type in image_types {
|
||||||
|
let url = repo.get_image_url("item123", image_type, None);
|
||||||
|
assert!(url.contains(&format!("/Images/{}", image_type)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_credentials_included_in_backend() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "secret_token");
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", None);
|
||||||
|
|
||||||
|
// Credentials should be included in backend-generated URL
|
||||||
|
assert!(url.contains("api_key=secret_token"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_proper_encoding() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: Some(300),
|
||||||
|
max_height: None,
|
||||||
|
quality: None,
|
||||||
|
tag: Some("tag-with-special-chars".to_string()),
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", Some(&options));
|
||||||
|
|
||||||
|
// URL should be properly formatted
|
||||||
|
assert!(url.contains("?"));
|
||||||
|
assert!(url.contains("&") || !url.contains("&&")); // No double ampersands
|
||||||
|
assert!(!url.ends_with("&")); // No trailing ampersand
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== Subtitle URL Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_subtitle_url_vtt_format() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_subtitle_url("item123", "source456", 0, "vtt");
|
||||||
|
|
||||||
|
assert!(url.contains("Videos/item123"));
|
||||||
|
assert!(url.contains("Subtitles/source456/0"));
|
||||||
|
assert!(url.contains("subtitles.vtt"));
|
||||||
|
assert!(url.contains("api_key=token123"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_subtitle_url_srt_format() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_subtitle_url("item123", "source456", 1, "srt");
|
||||||
|
|
||||||
|
assert!(url.contains("Subtitles/source456/1"));
|
||||||
|
assert!(url.contains("subtitles.srt"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_subtitle_url_multiple_streams() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
for stream_index in 0..5 {
|
||||||
|
let url = repo.get_subtitle_url("item123", "source456", stream_index, "vtt");
|
||||||
|
assert!(url.contains(&format!("/{}/subtitles", stream_index)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_subtitle_url_different_media_sources() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let media_sources = vec!["src1", "src2", "src3"];
|
||||||
|
|
||||||
|
for media_source_id in media_sources {
|
||||||
|
let url = repo.get_subtitle_url("item123", media_source_id, 0, "vtt");
|
||||||
|
assert!(url.contains(&format!("Subtitles/{}/", media_source_id)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== Video Download URL Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_video_download_url_original_quality() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_video_download_url("item123", "original");
|
||||||
|
|
||||||
|
assert!(url.contains("Videos/item123/stream.mp4"));
|
||||||
|
assert!(url.contains("api_key=token123"));
|
||||||
|
assert!(!url.contains("maxWidth")); // Original should have no transcoding params
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_video_download_url_1080p() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_video_download_url("item123", "1080p");
|
||||||
|
|
||||||
|
assert!(url.contains("maxWidth=1920"));
|
||||||
|
assert!(url.contains("videoBitrate=15000k"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_video_download_url_720p() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_video_download_url("item123", "720p");
|
||||||
|
|
||||||
|
assert!(url.contains("maxWidth=1280"));
|
||||||
|
assert!(url.contains("videoBitrate=8000k"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_video_download_url_480p() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let url = repo.get_video_download_url("item123", "480p");
|
||||||
|
|
||||||
|
assert!(url.contains("maxWidth=854"));
|
||||||
|
assert!(url.contains("videoBitrate=3000k"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_video_download_url_quality_presets() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
let qualities = vec!["original", "1080p", "720p", "480p"];
|
||||||
|
|
||||||
|
for quality in qualities {
|
||||||
|
let url = repo.get_video_download_url("item123", quality);
|
||||||
|
assert!(url.contains("Videos/item123/stream.mp4"));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== Security Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_credentials_never_exposed_in_frontend() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "super_secret_token");
|
||||||
|
|
||||||
|
let image_url = repo.get_image_url("item123", "Primary", None);
|
||||||
|
let subtitle_url = repo.get_subtitle_url("item123", "src123", 0, "vtt");
|
||||||
|
let download_url = repo.get_video_download_url("item123", "720p");
|
||||||
|
|
||||||
|
// These URLs are constructed in BACKEND and returned to frontend
|
||||||
|
// Frontend never receives this token directly
|
||||||
|
assert!(image_url.contains("api_key=super_secret_token"));
|
||||||
|
assert!(subtitle_url.contains("api_key=super_secret_token"));
|
||||||
|
assert!(download_url.contains("api_key=super_secret_token"));
|
||||||
|
|
||||||
|
// In actual implementation, frontend would only get the URL string
|
||||||
|
// Frontend cannot construct its own URLs or extract the token
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_url_parameter_injection_prevention() {
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "token123");
|
||||||
|
|
||||||
|
// Try to inject parameters through item_id
|
||||||
|
let malicious_id = "item123&extraParam=malicious";
|
||||||
|
let url = repo.get_image_url(malicious_id, "Primary", None);
|
||||||
|
|
||||||
|
// URL should contain the full item_id, backend should handle escaping
|
||||||
|
assert!(url.contains(malicious_id));
|
||||||
|
// Backend should be responsible for proper URL encoding
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== URL Format Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_image_url_format_correctness() {
|
||||||
|
let repo = MockOnlineRepository::new("https://server.com", "token");
|
||||||
|
|
||||||
|
let url = repo.get_image_url("id123", "Primary", None);
|
||||||
|
|
||||||
|
// Should be valid format
|
||||||
|
assert!(url.starts_with("https://server.com"));
|
||||||
|
assert!(url.contains("/Items/id123/Images/Primary"));
|
||||||
|
assert!(url.contains("?api_key="));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_query_string_properly_separated() {
|
||||||
|
let repo = MockOnlineRepository::new("https://server.com", "token");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: Some(300),
|
||||||
|
max_height: Some(200),
|
||||||
|
quality: None,
|
||||||
|
tag: None,
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("id123", "Primary", Some(&options));
|
||||||
|
|
||||||
|
// Should have single ? separator
|
||||||
|
let question_marks = url.matches('?').count();
|
||||||
|
assert_eq!(question_marks, 1);
|
||||||
|
|
||||||
|
// Should have ampersands between params
|
||||||
|
assert!(url.contains("?"));
|
||||||
|
assert!(url.contains("&"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_special_characters_in_urls() {
|
||||||
|
let repo = MockOnlineRepository::new("https://server.com", "token_with_special-chars");
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item-with-special_chars", "Primary", None);
|
||||||
|
|
||||||
|
// Should handle special characters in token and id
|
||||||
|
assert!(url.contains("token_with_special-chars"));
|
||||||
|
assert!(url.contains("item-with-special_chars"));
|
||||||
|
}
|
||||||
|
|
||||||
|
// ===== Backend vs Frontend Responsibility Tests =====
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_backend_owns_url_construction() {
|
||||||
|
// This test documents that URL construction is ONLY in backend
|
||||||
|
let repo = MockOnlineRepository::new("https://jellyfin.example.com", "secret_token");
|
||||||
|
|
||||||
|
// Backend generates full URL with credentials
|
||||||
|
let url = repo.get_image_url("item123", "Primary", None);
|
||||||
|
|
||||||
|
// URL is complete and ready to use
|
||||||
|
assert!(url.starts_with("https://"));
|
||||||
|
assert!(url.contains("api_key="));
|
||||||
|
|
||||||
|
// Frontend never constructs URLs directly
|
||||||
|
// Frontend only receives pre-constructed URLs from backend
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_url_includes_all_necessary_parameters() {
|
||||||
|
let repo = MockOnlineRepository::new("https://server.com", "token");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: Some(300),
|
||||||
|
max_height: Some(200),
|
||||||
|
quality: Some(90),
|
||||||
|
tag: Some("abc".to_string()),
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", Some(&options));
|
||||||
|
|
||||||
|
// All provided options should be in URL
|
||||||
|
assert!(url.contains("maxWidth=300"));
|
||||||
|
assert!(url.contains("maxHeight=200"));
|
||||||
|
assert!(url.contains("quality=90"));
|
||||||
|
assert!(url.contains("tag=abc"));
|
||||||
|
assert!(url.contains("api_key=token"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_optional_parameters_omitted_when_not_provided() {
|
||||||
|
let repo = MockOnlineRepository::new("https://server.com", "token");
|
||||||
|
let options = ImageOptions {
|
||||||
|
max_width: None,
|
||||||
|
max_height: None,
|
||||||
|
quality: None,
|
||||||
|
tag: None,
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = repo.get_image_url("item123", "Primary", Some(&options));
|
||||||
|
|
||||||
|
// Should only have api_key
|
||||||
|
assert!(url.contains("api_key=token"));
|
||||||
|
assert!(!url.contains("maxWidth"));
|
||||||
|
assert!(!url.contains("maxHeight"));
|
||||||
|
assert!(!url.contains("quality"));
|
||||||
|
assert!(!url.contains("tag"));
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -285,6 +285,7 @@ fn convert_params(params: &[QueryParam]) -> Vec<rusqlite::types::Value> {
|
|||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-002, UR-012 | DR-012 | UT-014, UT-015, UT-016, UT-017, UT-018, UT-019, UT-020, UT-021, UT-022, UT-023, UT-025
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -156,6 +156,7 @@ impl Database {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TRACES: UR-002, UR-012, UR-019, UR-025 | DR-012 | UT-014, UT-015, UT-016, UT-017, UT-018, UT-019, UT-020, UT-021, UT-022, UT-023, UT-025
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|||||||
@ -17,6 +17,7 @@ pub const MIGRATIONS: &[(&str, &str)] = &[
|
|||||||
("012_download_source", MIGRATION_012),
|
("012_download_source", MIGRATION_012),
|
||||||
("013_downloads_item_status_index", MIGRATION_013),
|
("013_downloads_item_status_index", MIGRATION_013),
|
||||||
("014_series_audio_preferences", MIGRATION_014),
|
("014_series_audio_preferences", MIGRATION_014),
|
||||||
|
("015_device_id", MIGRATION_015),
|
||||||
];
|
];
|
||||||
|
|
||||||
/// Initial schema migration
|
/// Initial schema migration
|
||||||
@ -638,3 +639,20 @@ CREATE TABLE IF NOT EXISTS series_audio_preferences (
|
|||||||
CREATE INDEX IF NOT EXISTS idx_series_audio_prefs_user_series
|
CREATE INDEX IF NOT EXISTS idx_series_audio_prefs_user_series
|
||||||
ON series_audio_preferences(user_id, series_id);
|
ON series_audio_preferences(user_id, series_id);
|
||||||
"#;
|
"#;
|
||||||
|
|
||||||
|
/// Migration to add device ID storage
|
||||||
|
/// - Creates app_settings table for app-wide configuration (device ID, etc.)
|
||||||
|
/// - Device ID is generated once and persisted for Jellyfin server identification
|
||||||
|
const MIGRATION_015: &str = r#"
|
||||||
|
-- App-wide settings table for device ID and other app-level configuration
|
||||||
|
-- Device ID is a unique identifier for this app installation
|
||||||
|
-- Required for Jellyfin server communication and session tracking
|
||||||
|
CREATE TABLE IF NOT EXISTS app_settings (
|
||||||
|
key TEXT PRIMARY KEY,
|
||||||
|
value TEXT NOT NULL,
|
||||||
|
updated_at TEXT DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
-- Create index for efficient lookups (though key is already primary key)
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_app_settings_key ON app_settings(key);
|
||||||
|
"#;
|
||||||
|
|||||||
545
src/lib/api/backend-integration.test.ts
Normal file
545
src/lib/api/backend-integration.test.ts
Normal file
@ -0,0 +1,545 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
|
import { RepositoryClient } from "./repository-client";
|
||||||
|
|
||||||
|
vi.mock("@tauri-apps/api/core");
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Integration tests documenting Phase 1 & 2 refactoring:
|
||||||
|
* - Sorting moved to backend (no frontend compareFn)
|
||||||
|
* - Filtering moved to backend (no frontend iteration/matching)
|
||||||
|
* - URL construction moved to backend (async Tauri invoke)
|
||||||
|
* - Search moved to backend (backend search command)
|
||||||
|
*/
|
||||||
|
|
||||||
|
describe("Backend Integration - Refactored Business Logic", () => {
|
||||||
|
let client: RepositoryClient;
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
client = new RepositoryClient();
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Sorting Delegated to Backend", () => {
|
||||||
|
it("should pass sortBy to backend instead of frontend sorting", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [
|
||||||
|
{ id: "item1", name: "Album A" },
|
||||||
|
{ id: "item2", name: "Album B" },
|
||||||
|
{ id: "item3", name: "Album C" },
|
||||||
|
],
|
||||||
|
totalRecordCount: 3,
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
});
|
||||||
|
|
||||||
|
// Backend should have done the sorting
|
||||||
|
expect(result.items[0].name).toBe("Album A");
|
||||||
|
|
||||||
|
// Frontend doesn't have a compareFn
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_items", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
parentId: "library123",
|
||||||
|
options: {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
},
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support different sort fields via backend", async () => {
|
||||||
|
const sortFields = ["SortName", "Artist", "Album", "DatePlayed", "ProductionYear"];
|
||||||
|
|
||||||
|
for (const sortField of sortFields) {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.getItems("library123", {
|
||||||
|
sortBy: sortField,
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_items",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
sortBy: sortField,
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should pass sort order to backend", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Descending",
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_items",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
sortOrder: "Descending",
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should NOT include frontend compareFn (removed entirely)", async () => {
|
||||||
|
// Old code pattern:
|
||||||
|
// sortOptions: [{
|
||||||
|
// key: "title",
|
||||||
|
// label: "Title",
|
||||||
|
// compareFn: (a, b) => a.name.localeCompare(b.name) // ← REMOVED
|
||||||
|
// }]
|
||||||
|
|
||||||
|
// New code pattern:
|
||||||
|
// sortOptions: [{
|
||||||
|
// key: "SortName", // Jellyfin field name
|
||||||
|
// label: "Title"
|
||||||
|
// }]
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
{ key: "Artist", label: "Artist" },
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
// Verify no compareFn property exists
|
||||||
|
for (const option of config.sortOptions) {
|
||||||
|
expect((option as any).compareFn).toBeUndefined();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Filtering Delegated to Backend", () => {
|
||||||
|
it("should pass includeItemTypes to backend", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.getItems("library123", {
|
||||||
|
includeItemTypes: ["Audio", "MusicAlbum"],
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_items",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
includeItemTypes: ["Audio", "MusicAlbum"],
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should pass genres filter to backend", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.getItems("library123", {
|
||||||
|
genres: ["Rock", "Jazz"],
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_items",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
genres: ["Rock", "Jazz"],
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should NOT include frontend filtering logic", async () => {
|
||||||
|
// Old code pattern:
|
||||||
|
// let filtered = items.filter(item => {
|
||||||
|
// return searchFields.some(field => {
|
||||||
|
// const fieldValue = item[field]?.toLowerCase() ?? "";
|
||||||
|
// return fieldValue.includes(query.toLowerCase());
|
||||||
|
// });
|
||||||
|
// }); // ← REMOVED
|
||||||
|
|
||||||
|
// New code pattern:
|
||||||
|
// Use backend search instead
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [{ id: "item1", name: "Search Result" }],
|
||||||
|
totalRecordCount: 1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await client.search("query");
|
||||||
|
|
||||||
|
expect(result.items.length).toBeGreaterThan(0);
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_search",
|
||||||
|
expect.objectContaining({
|
||||||
|
query: "query",
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support pagination via backend", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 1000,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.getItems("library123", {
|
||||||
|
startIndex: 100,
|
||||||
|
limit: 50,
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_items",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
startIndex: 100,
|
||||||
|
limit: 50,
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Search Delegated to Backend", () => {
|
||||||
|
it("should use backend search command instead of frontend filtering", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [
|
||||||
|
{ id: "item1", name: "Found Item" },
|
||||||
|
{ id: "item2", name: "Another Found Item" },
|
||||||
|
],
|
||||||
|
totalRecordCount: 2,
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await client.search("query");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_search",
|
||||||
|
expect.objectContaining({
|
||||||
|
query: "query",
|
||||||
|
})
|
||||||
|
);
|
||||||
|
|
||||||
|
expect(result.items.length).toBe(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support search with item type filters", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
await client.search("query", {
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_search",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: expect.objectContaining({
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
}),
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should NOT do client-side search filtering", async () => {
|
||||||
|
// Old code pattern:
|
||||||
|
// const query = searchInput.toLowerCase();
|
||||||
|
// const results = items.filter(item =>
|
||||||
|
// config.searchFields.some(field =>
|
||||||
|
// item[field]?.toLowerCase()?.includes(query)
|
||||||
|
// )
|
||||||
|
// ); // ← REMOVED
|
||||||
|
|
||||||
|
// New code pattern:
|
||||||
|
// Call backend search directly
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [{ id: "item1" }],
|
||||||
|
totalRecordCount: 1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await client.search("search term");
|
||||||
|
|
||||||
|
// Backend did the filtering
|
||||||
|
expect(result.items).toBeDefined();
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_search", expect.any(Object));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("URL Construction Delegated to Backend", () => {
|
||||||
|
it("should get image URLs from backend (not construct in frontend)", async () => {
|
||||||
|
const backendUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(backendUrl);
|
||||||
|
|
||||||
|
const url = await client.getImageUrl("item123", "Primary");
|
||||||
|
|
||||||
|
// Backend constructed and returned the URL
|
||||||
|
expect(url).toBe(backendUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_image_url",
|
||||||
|
expect.objectContaining({
|
||||||
|
itemId: "item123",
|
||||||
|
imageType: "Primary",
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should NOT construct image URLs in frontend", async () => {
|
||||||
|
// Old code pattern:
|
||||||
|
// return `${serverUrl}/Items/${itemId}/Images/${imageType}?api_key=${token}&maxWidth=${options.maxWidth}`;
|
||||||
|
// ← REMOVED - NEVER construct URLs in frontend
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce("https://server.com/image");
|
||||||
|
|
||||||
|
const url = await client.getImageUrl("item123", "Primary", { maxWidth: 300 });
|
||||||
|
|
||||||
|
// URL came from backend, not constructed in frontend
|
||||||
|
expect(typeof url).toBe("string");
|
||||||
|
expect(url).toContain("http");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get video stream URLs from backend", async () => {
|
||||||
|
const backendUrl = "https://server.com/Videos/item123/stream.mp4?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(backendUrl);
|
||||||
|
|
||||||
|
const url = await client.getVideoStreamUrl("item123");
|
||||||
|
|
||||||
|
expect(url).toBe(backendUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_video_stream_url",
|
||||||
|
expect.any(Object)
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get subtitle URLs from backend", async () => {
|
||||||
|
const backendUrl = "https://server.com/Videos/item123/Subtitles/0/subtitles.vtt?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(backendUrl);
|
||||||
|
|
||||||
|
const url = await client.getSubtitleUrl("item123", "source456", 0);
|
||||||
|
|
||||||
|
expect(url).toBe(backendUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_subtitle_url",
|
||||||
|
expect.any(Object)
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get video download URLs from backend", async () => {
|
||||||
|
const backendUrl = "https://server.com/Videos/item123/stream.mp4?maxWidth=1280&api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(backendUrl);
|
||||||
|
|
||||||
|
const url = await client.getVideoDownloadUrl("item123", "720p");
|
||||||
|
|
||||||
|
expect(url).toBe(backendUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_video_download_url",
|
||||||
|
expect.any(Object)
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should never expose access token in frontend code", async () => {
|
||||||
|
// The access token is NEVER used in frontend URL construction
|
||||||
|
// It's only stored in backend for secure URL generation
|
||||||
|
|
||||||
|
// Frontend code NEVER has direct access to use the token
|
||||||
|
const client2 = new RepositoryClient();
|
||||||
|
// client2._accessToken is private and should never be accessed or used
|
||||||
|
|
||||||
|
// All token usage is in backend via Tauri commands
|
||||||
|
expect(invoke).toBeDefined();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Component Config Simplification", () => {
|
||||||
|
it("should have simplified MediaListConfig (no searchFields)", () => {
|
||||||
|
// Old type:
|
||||||
|
// interface MediaListConfig {
|
||||||
|
// searchFields: string[]; // ← REMOVED
|
||||||
|
// compareFn?: (a, b) => number; // ← REMOVED
|
||||||
|
// }
|
||||||
|
|
||||||
|
// New type:
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
// No compareFn
|
||||||
|
],
|
||||||
|
// No searchFields
|
||||||
|
};
|
||||||
|
|
||||||
|
// Verify no searchFields
|
||||||
|
expect((config as any).searchFields).toBeUndefined();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use Jellyfin field names in sort options", () => {
|
||||||
|
// Old:
|
||||||
|
// { key: "title", label: "Title", compareFn: ... }
|
||||||
|
|
||||||
|
// New:
|
||||||
|
// { key: "SortName", label: "Title" }
|
||||||
|
|
||||||
|
const sortOptions = [
|
||||||
|
{ key: "SortName", label: "A-Z" },
|
||||||
|
{ key: "Artist", label: "Artist" },
|
||||||
|
{ key: "Album", label: "Album" },
|
||||||
|
{ key: "DatePlayed", label: "Recent" },
|
||||||
|
];
|
||||||
|
|
||||||
|
for (const option of sortOptions) {
|
||||||
|
// Should be Jellyfin field names
|
||||||
|
expect(typeof option.key).toBe("string");
|
||||||
|
expect(option.key).toMatch(/^[A-Z]/); // Jellyfin fields start with capital
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Debounced Search Implementation", () => {
|
||||||
|
it("should debounce search without frontend filtering", async () => {
|
||||||
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
const mockSearch = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
(invoke as any).mockImplementation((cmd: string, args: any) => {
|
||||||
|
if (cmd === "repository_search") {
|
||||||
|
return mockSearch(args.query);
|
||||||
|
}
|
||||||
|
return Promise.resolve({ items: [], totalRecordCount: 0 });
|
||||||
|
});
|
||||||
|
|
||||||
|
// Simulate rapid search queries
|
||||||
|
await client.search("t");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
await client.search("te");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
await client.search("test");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// All calls go to backend (debouncing happens in component via $effect)
|
||||||
|
expect(invoke).toHaveBeenCalled();
|
||||||
|
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("End-to-End Data Flow", () => {
|
||||||
|
it("should support complete flow: load → sort → display", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [
|
||||||
|
{ id: "id1", name: "Album A", sortName: "A" },
|
||||||
|
{ id: "id2", name: "Album B", sortName: "B" },
|
||||||
|
],
|
||||||
|
totalRecordCount: 2,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Frontend requests items with sort
|
||||||
|
const result = await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
});
|
||||||
|
|
||||||
|
// Backend returned pre-sorted items
|
||||||
|
expect(result.items[0].sortName).toBe("A");
|
||||||
|
expect(result.items[1].sortName).toBe("B");
|
||||||
|
|
||||||
|
// Frontend just displays them
|
||||||
|
// No compareFn, no local sorting
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support complete flow: search → load images → display", async () => {
|
||||||
|
// 1. Frontend calls backend search
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [{ id: "item1", name: "Result", primaryImageTag: "tag1" }],
|
||||||
|
totalRecordCount: 1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const searchResult = await client.search("query");
|
||||||
|
expect(searchResult.items.length).toBe(1);
|
||||||
|
|
||||||
|
// 2. Frontend loads image URL from backend
|
||||||
|
(invoke as any).mockResolvedValueOnce("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
const imageUrl = await client.getImageUrl("item1", "Primary");
|
||||||
|
expect(imageUrl).toContain("http");
|
||||||
|
|
||||||
|
// 3. Frontend displays search results with images
|
||||||
|
// No client-side filtering, sorting, or URL construction
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Performance Characteristics", () => {
|
||||||
|
it("should reduce memory usage by not storing frontend sorting state", async () => {
|
||||||
|
// Old: Frontend stores items + sorting state + filtered results
|
||||||
|
// Old: Multiple copies of data (original, filtered, sorted)
|
||||||
|
|
||||||
|
// New: Backend returns already-sorted data
|
||||||
|
// New: Frontend just stores the result
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: Array.from({ length: 10000 }, (_, i) => ({
|
||||||
|
id: `id${i}`,
|
||||||
|
name: `Item ${i}`,
|
||||||
|
})),
|
||||||
|
totalRecordCount: 10000,
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
limit: 10000,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Backend handled sorting
|
||||||
|
expect(result.items.length).toBe(10000);
|
||||||
|
// Frontend just stores the result array
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should reduce CPU usage by avoiding client-side operations", async () => {
|
||||||
|
// Old pattern required:
|
||||||
|
// - Parsing all items into memory
|
||||||
|
// - Iterating to apply filters
|
||||||
|
// - Sorting algorithm (O(n log n) comparisons)
|
||||||
|
// - Updating multiple state variables
|
||||||
|
|
||||||
|
// New pattern:
|
||||||
|
(invoke as any).mockResolvedValueOnce({
|
||||||
|
items: [], // Backend already filtered/sorted
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Frontend just awaits backend result
|
||||||
|
const result = await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
});
|
||||||
|
|
||||||
|
// No client-side work
|
||||||
|
expect(result).toBeDefined();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
428
src/lib/api/repository-client.test.ts
Normal file
428
src/lib/api/repository-client.test.ts
Normal file
@ -0,0 +1,428 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
|
import { RepositoryClient } from "./repository-client";
|
||||||
|
|
||||||
|
vi.mock("@tauri-apps/api/core");
|
||||||
|
|
||||||
|
describe("RepositoryClient", () => {
|
||||||
|
let client: RepositoryClient;
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
client = new RepositoryClient();
|
||||||
|
vi.clearAllMocks();
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
vi.resetAllMocks();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Initialization", () => {
|
||||||
|
it("should initialize with no handle", () => {
|
||||||
|
expect(() => client.getHandle()).toThrow("Repository not initialized");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should create repository with invoke command", async () => {
|
||||||
|
const mockHandle = "test-handle-123";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockHandle);
|
||||||
|
|
||||||
|
const handle = await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
expect(handle).toBe(mockHandle);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_create", {
|
||||||
|
serverUrl: "https://server.com",
|
||||||
|
userId: "user1",
|
||||||
|
accessToken: "token123",
|
||||||
|
serverId: "server1",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should store handle after creation", async () => {
|
||||||
|
const mockHandle = "test-handle-456";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockHandle);
|
||||||
|
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
expect(client.getHandle()).toBe(mockHandle);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should destroy repository and clear handle", async () => {
|
||||||
|
const mockHandle = "test-handle-789";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockHandle);
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce(undefined);
|
||||||
|
await client.destroy();
|
||||||
|
|
||||||
|
expect(() => client.getHandle()).toThrow("Repository not initialized");
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_destroy", { handle: mockHandle });
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Image URL Methods", () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get image URL from backend", async () => {
|
||||||
|
const mockUrl = "https://server.com/Items/item123/Images/Primary?maxWidth=300&api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const imageUrl = await client.getImageUrl("item123", "Primary", { maxWidth: 300 });
|
||||||
|
|
||||||
|
expect(imageUrl).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_image_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
imageType: "Primary",
|
||||||
|
options: { maxWidth: 300 },
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use default image type if not provided", async () => {
|
||||||
|
const mockUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getImageUrl("item123");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_image_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
imageType: "Primary",
|
||||||
|
options: null,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should pass multiple image options to backend", async () => {
|
||||||
|
const mockUrl = "https://server.com/Items/item123/Images/Backdrop?maxWidth=1920&maxHeight=1080&quality=90&api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const options = {
|
||||||
|
maxWidth: 1920,
|
||||||
|
maxHeight: 1080,
|
||||||
|
quality: 90,
|
||||||
|
tag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
await client.getImageUrl("item123", "Backdrop", options);
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_image_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
imageType: "Backdrop",
|
||||||
|
options,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle different image types", async () => {
|
||||||
|
const mockUrl = "https://server.com/Items/item123/Images/Logo?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getImageUrl("item123", "Logo");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_image_url", {
|
||||||
|
handle: expect.any(String),
|
||||||
|
itemId: "item123",
|
||||||
|
imageType: "Logo",
|
||||||
|
options: null,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should throw error if not initialized before getImageUrl", async () => {
|
||||||
|
const newClient = new RepositoryClient();
|
||||||
|
|
||||||
|
await expect(newClient.getImageUrl("item123")).rejects.toThrow(
|
||||||
|
"Repository not initialized"
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Subtitle URL Methods", () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get subtitle URL from backend", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/Subtitles/1/subtitles.vtt?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const subtitleUrl = await client.getSubtitleUrl("item123", "source456", 0);
|
||||||
|
|
||||||
|
expect(subtitleUrl).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_subtitle_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
mediaSourceId: "source456",
|
||||||
|
streamIndex: 0,
|
||||||
|
format: "vtt",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use default format if not provided", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/Subtitles/0/subtitles.vtt?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getSubtitleUrl("item123", "source456", 0);
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_subtitle_url", {
|
||||||
|
handle: expect.any(String),
|
||||||
|
itemId: "item123",
|
||||||
|
mediaSourceId: "source456",
|
||||||
|
streamIndex: 0,
|
||||||
|
format: "vtt",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support custom subtitle formats", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/Subtitles/0/subtitles.srt?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getSubtitleUrl("item123", "source456", 1, "srt");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_subtitle_url", {
|
||||||
|
handle: expect.any(String),
|
||||||
|
itemId: "item123",
|
||||||
|
mediaSourceId: "source456",
|
||||||
|
streamIndex: 1,
|
||||||
|
format: "srt",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Video Download URL Methods", () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get video download URL from backend", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/stream.mp4?maxWidth=1920&api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const downloadUrl = await client.getVideoDownloadUrl("item123", "1080p");
|
||||||
|
|
||||||
|
expect(downloadUrl).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_video_download_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
quality: "1080p",
|
||||||
|
mediaSourceId: null,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use original quality by default", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/stream.mp4?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getVideoDownloadUrl("item123");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_video_download_url", {
|
||||||
|
handle: expect.any(String),
|
||||||
|
itemId: "item123",
|
||||||
|
quality: "original",
|
||||||
|
mediaSourceId: null,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support quality presets", async () => {
|
||||||
|
const qualities = ["original", "1080p", "720p", "480p"];
|
||||||
|
|
||||||
|
for (const quality of qualities) {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce(`https://server.com/stream.mp4?quality=${quality}`);
|
||||||
|
|
||||||
|
await client.getVideoDownloadUrl("item123", quality as any);
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_video_download_url",
|
||||||
|
expect.objectContaining({
|
||||||
|
quality,
|
||||||
|
})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support optional media source ID", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/stream.mp4?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
await client.getVideoDownloadUrl("item123", "720p", "source789");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_video_download_url", {
|
||||||
|
handle: expect.any(String),
|
||||||
|
itemId: "item123",
|
||||||
|
quality: "720p",
|
||||||
|
mediaSourceId: "source789",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Library Methods", () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get libraries from backend", async () => {
|
||||||
|
const mockLibraries = [
|
||||||
|
{ id: "lib1", name: "Music", collectionType: "music" },
|
||||||
|
{ id: "lib2", name: "Movies", collectionType: "movies" },
|
||||||
|
];
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockLibraries);
|
||||||
|
|
||||||
|
const libraries = await client.getLibraries();
|
||||||
|
|
||||||
|
expect(libraries).toEqual(mockLibraries);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_libraries", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get items with sorting parameters", async () => {
|
||||||
|
const mockResult = {
|
||||||
|
items: [
|
||||||
|
{ id: "item1", name: "Track 1", type: "Audio" },
|
||||||
|
{ id: "item2", name: "Track 2", type: "Audio" },
|
||||||
|
],
|
||||||
|
totalRecordCount: 2,
|
||||||
|
};
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockResult);
|
||||||
|
|
||||||
|
const result = await client.getItems("library123", {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
limit: 50,
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(result).toEqual(mockResult);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_items", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
parentId: "library123",
|
||||||
|
options: {
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
limit: 50,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should search with backend search command", async () => {
|
||||||
|
const mockResult = {
|
||||||
|
items: [
|
||||||
|
{ id: "item1", name: "Search Result 1", type: "Audio" },
|
||||||
|
],
|
||||||
|
totalRecordCount: 1,
|
||||||
|
};
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockResult);
|
||||||
|
|
||||||
|
const result = await client.search("query", {
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
limit: 100,
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(result).toEqual(mockResult);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_search", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
query: "query",
|
||||||
|
options: {
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
limit: 100,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Playback Methods", () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get audio stream URL", async () => {
|
||||||
|
const mockUrl = "https://server.com/Audio/item123/stream.mp3?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const url = await client.getAudioStreamUrl("item123");
|
||||||
|
|
||||||
|
expect(url).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_audio_stream_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get video stream URL", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/stream.mp4?api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const url = await client.getVideoStreamUrl("item123");
|
||||||
|
|
||||||
|
expect(url).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_video_stream_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
mediaSourceId: null,
|
||||||
|
startTimeSeconds: null,
|
||||||
|
audioStreamIndex: null,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should get video stream URL with options", async () => {
|
||||||
|
const mockUrl = "https://server.com/Videos/item123/stream.mp4?start=300&api_key=token";
|
||||||
|
(invoke as any).mockResolvedValueOnce(mockUrl);
|
||||||
|
|
||||||
|
const url = await client.getVideoStreamUrl("item123", "source456", 300, 0);
|
||||||
|
|
||||||
|
expect(url).toBe(mockUrl);
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_get_video_stream_url", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
mediaSourceId: "source456",
|
||||||
|
startTimeSeconds: 300,
|
||||||
|
audioStreamIndex: 0,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should report playback progress", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce(undefined);
|
||||||
|
|
||||||
|
await client.reportPlaybackProgress("item123", 5000000);
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith("repository_report_playback_progress", {
|
||||||
|
handle: "test-handle-123",
|
||||||
|
itemId: "item123",
|
||||||
|
positionTicks: 5000000,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Error Handling", () => {
|
||||||
|
it("should throw error if invoke fails", async () => {
|
||||||
|
(invoke as any).mockRejectedValueOnce(new Error("Network error"));
|
||||||
|
|
||||||
|
await expect(client.create("https://server.com", "user1", "token", "server1")).rejects.toThrow(
|
||||||
|
"Network error"
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle missing optional parameters", async () => {
|
||||||
|
(invoke as any).mockResolvedValueOnce("test-handle-123");
|
||||||
|
await client.create("https://server.com", "user1", "token123", "server1");
|
||||||
|
|
||||||
|
(invoke as any).mockResolvedValueOnce("");
|
||||||
|
|
||||||
|
await client.getImageUrl("item123");
|
||||||
|
|
||||||
|
expect(invoke).toHaveBeenCalledWith(
|
||||||
|
"repository_get_image_url",
|
||||||
|
expect.objectContaining({
|
||||||
|
options: null,
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@ -5,7 +5,6 @@
|
|||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
import type { QualityPreset } from "./quality-presets";
|
import type { QualityPreset } from "./quality-presets";
|
||||||
import { QUALITY_PRESETS } from "./quality-presets";
|
import { QUALITY_PRESETS } from "./quality-presets";
|
||||||
import { validateItemId, validateImageType, validateMediaSourceId, validateNumericParam, validateQueryParamValue } from "$lib/utils/validation";
|
|
||||||
import type {
|
import type {
|
||||||
Library,
|
Library,
|
||||||
MediaItem,
|
MediaItem,
|
||||||
@ -229,79 +228,40 @@ export class RepositoryClient {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get subtitle URL - constructs URL synchronously (no server call)
|
* Get subtitle URL from backend
|
||||||
|
* The Rust backend constructs and returns the URL with proper credentials handling
|
||||||
*/
|
*/
|
||||||
getSubtitleUrl(itemId: string, mediaSourceId: string, streamIndex: number, format: string = "vtt"): string {
|
async getSubtitleUrl(
|
||||||
if (!this._serverUrl || !this._accessToken) {
|
itemId: string,
|
||||||
throw new Error("Repository not initialized - call create() first");
|
mediaSourceId: string,
|
||||||
}
|
streamIndex: number,
|
||||||
|
format: string = "vtt"
|
||||||
// Validate inputs to prevent injection attacks
|
): Promise<string> {
|
||||||
validateItemId(itemId);
|
return invoke<string>("repository_get_subtitle_url", {
|
||||||
validateMediaSourceId(mediaSourceId);
|
handle: this.ensureHandle(),
|
||||||
const index = validateNumericParam(streamIndex, 0, 1000, "streamIndex");
|
itemId,
|
||||||
|
mediaSourceId,
|
||||||
// Validate format - only allow safe subtitle formats
|
streamIndex,
|
||||||
if (!/^[a-z]+$/.test(format)) {
|
format,
|
||||||
throw new Error("Invalid subtitle format");
|
});
|
||||||
}
|
|
||||||
|
|
||||||
return `${this._serverUrl}/Videos/${itemId}/${mediaSourceId}/Subtitles/${index}/Stream.${format}?api_key=${this._accessToken}`;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get video download URL with quality preset - constructs URL synchronously
|
* Get video download URL with quality preset from backend
|
||||||
* Used for offline downloads
|
* The Rust backend constructs and returns the URL with proper credentials handling
|
||||||
|
* Used for offline downloads and transcoding
|
||||||
*/
|
*/
|
||||||
getVideoDownloadUrl(
|
async getVideoDownloadUrl(
|
||||||
itemId: string,
|
itemId: string,
|
||||||
quality: QualityPreset = "original",
|
quality: QualityPreset = "original",
|
||||||
mediaSourceId?: string
|
mediaSourceId?: string
|
||||||
): string {
|
): Promise<string> {
|
||||||
if (!this._serverUrl || !this._accessToken) {
|
return invoke<string>("repository_get_video_download_url", {
|
||||||
throw new Error("Repository not initialized - call create() first");
|
handle: this.ensureHandle(),
|
||||||
}
|
itemId,
|
||||||
|
quality,
|
||||||
// Validate itemId and mediaSourceId
|
mediaSourceId: mediaSourceId ?? null,
|
||||||
validateItemId(itemId);
|
|
||||||
if (mediaSourceId) {
|
|
||||||
validateMediaSourceId(mediaSourceId);
|
|
||||||
}
|
|
||||||
|
|
||||||
const preset = QUALITY_PRESETS[quality];
|
|
||||||
|
|
||||||
if (quality === "original" || !preset.videoBitrate) {
|
|
||||||
// Direct stream for original quality
|
|
||||||
const params = new URLSearchParams({
|
|
||||||
api_key: this._accessToken,
|
|
||||||
Static: "true",
|
|
||||||
audioStreamIndex: "0",
|
|
||||||
});
|
});
|
||||||
if (mediaSourceId) {
|
|
||||||
params.append("MediaSourceId", mediaSourceId);
|
|
||||||
}
|
|
||||||
return `${this._serverUrl}/Videos/${itemId}/stream?${params.toString()}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Transcoded download with quality preset
|
|
||||||
const params = new URLSearchParams({
|
|
||||||
api_key: this._accessToken,
|
|
||||||
DeviceId: localStorage.getItem("jellytau_device_id") || "jellytau",
|
|
||||||
Container: "mp4",
|
|
||||||
VideoCodec: "h264",
|
|
||||||
AudioCodec: "aac",
|
|
||||||
AudioStreamIndex: "0",
|
|
||||||
VideoBitrate: preset.videoBitrate.toString(),
|
|
||||||
AudioBitrate: preset.audioBitrate.toString(),
|
|
||||||
MaxHeight: preset.maxHeight?.toString() ?? "",
|
|
||||||
TranscodingMaxAudioChannels: "2",
|
|
||||||
});
|
|
||||||
|
|
||||||
if (mediaSourceId) {
|
|
||||||
params.append("MediaSourceId", mediaSourceId);
|
|
||||||
}
|
|
||||||
|
|
||||||
return `${this._serverUrl}/Videos/${itemId}/stream.mp4?${params.toString()}`;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// ===== Favorite Methods (via Rust) =====
|
// ===== Favorite Methods (via Rust) =====
|
||||||
|
|||||||
@ -13,6 +13,7 @@
|
|||||||
|
|
||||||
let currentIndex = $state(0);
|
let currentIndex = $state(0);
|
||||||
let intervalId: number | null = null;
|
let intervalId: number | null = null;
|
||||||
|
let heroImageUrl = $state<string>("");
|
||||||
|
|
||||||
// Touch/swipe state
|
// Touch/swipe state
|
||||||
let touchStartX = $state(0);
|
let touchStartX = $state(0);
|
||||||
@ -21,65 +22,81 @@
|
|||||||
|
|
||||||
const currentItem = $derived(items[currentIndex] ?? null);
|
const currentItem = $derived(items[currentIndex] ?? null);
|
||||||
|
|
||||||
function getHeroImageUrl(): string {
|
// Load hero image URL asynchronously based on item priority
|
||||||
if (!currentItem) return "";
|
async function loadHeroImageUrl(): Promise<void> {
|
||||||
|
if (!currentItem) {
|
||||||
|
heroImageUrl = "";
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
|
|
||||||
// 1. Try backdrop image first (best for hero display)
|
// 1. Try backdrop image first (best for hero display)
|
||||||
if (currentItem.backdropImageTags?.[0]) {
|
if (currentItem.backdropImageTags?.[0]) {
|
||||||
return repo.getImageUrl(currentItem.id, "Backdrop", {
|
heroImageUrl = await repo.getImageUrl(currentItem.id, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: currentItem.backdropImageTags[0],
|
tag: currentItem.backdropImageTags[0],
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// 2. For episodes, try to use series backdrop from parent
|
// 2. For episodes, try to use series backdrop from parent
|
||||||
if (currentItem.type === "Episode") {
|
if (currentItem.type === "Episode") {
|
||||||
// First try parent backdrop tags (includes image tag for caching)
|
// First try parent backdrop tags (includes image tag for caching)
|
||||||
if (currentItem.seriesId && currentItem.parentBackdropImageTags?.[0]) {
|
if (currentItem.seriesId && currentItem.parentBackdropImageTags?.[0]) {
|
||||||
return repo.getImageUrl(currentItem.seriesId, "Backdrop", {
|
heroImageUrl = await repo.getImageUrl(currentItem.seriesId, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: currentItem.parentBackdropImageTags[0],
|
tag: currentItem.parentBackdropImageTags[0],
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
// Fallback: try series backdrop without tag (may not be cached optimally)
|
// Fallback: try series backdrop without tag (may not be cached optimally)
|
||||||
if (currentItem.seriesId) {
|
if (currentItem.seriesId) {
|
||||||
return repo.getImageUrl(currentItem.seriesId, "Backdrop", {
|
heroImageUrl = await repo.getImageUrl(currentItem.seriesId, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
// Last resort for episodes: try season backdrop
|
// Last resort for episodes: try season backdrop
|
||||||
if (currentItem.seasonId) {
|
if (currentItem.seasonId) {
|
||||||
return repo.getImageUrl(currentItem.seasonId, "Backdrop", {
|
heroImageUrl = await repo.getImageUrl(currentItem.seasonId, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// 3. For music tracks, try album backdrop first, then primary
|
// 3. For music tracks, try album backdrop first, then primary
|
||||||
if (currentItem.type === "Audio" && currentItem.albumId) {
|
if (currentItem.type === "Audio" && currentItem.albumId) {
|
||||||
// Try album backdrop first (more cinematic for hero)
|
// Try album backdrop first (more cinematic for hero)
|
||||||
return repo.getImageUrl(currentItem.albumId, "Backdrop", {
|
heroImageUrl = await repo.getImageUrl(currentItem.albumId, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// 4. Fall back to primary image (poster, album art, episode thumbnail)
|
// 4. Fall back to primary image (poster, album art, episode thumbnail)
|
||||||
if (currentItem.primaryImageTag) {
|
if (currentItem.primaryImageTag) {
|
||||||
return repo.getImageUrl(currentItem.id, "Primary", {
|
heroImageUrl = await repo.getImageUrl(currentItem.id, "Primary", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: currentItem.primaryImageTag,
|
tag: currentItem.primaryImageTag,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// 5. Last resort for audio: try album primary image
|
// 5. Last resort for audio: try album primary image
|
||||||
if (currentItem.type === "Audio" && currentItem.albumId) {
|
if (currentItem.type === "Audio" && currentItem.albumId) {
|
||||||
return repo.getImageUrl(currentItem.albumId, "Primary", {
|
heroImageUrl = await repo.getImageUrl(currentItem.albumId, "Primary", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
return "";
|
heroImageUrl = "";
|
||||||
|
} catch {
|
||||||
|
heroImageUrl = "";
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function next() {
|
function next() {
|
||||||
@ -126,6 +143,11 @@
|
|||||||
touchEndX = 0;
|
touchEndX = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load hero image whenever current item changes
|
||||||
|
$effect(() => {
|
||||||
|
loadHeroImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
// Auto-rotate logic
|
// Auto-rotate logic
|
||||||
$effect(() => {
|
$effect(() => {
|
||||||
if (autoRotate && items.length > 1) {
|
if (autoRotate && items.length > 1) {
|
||||||
@ -135,8 +157,6 @@
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
const heroImageUrl = $derived(getHeroImageUrl());
|
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
<div
|
<div
|
||||||
|
|||||||
431
src/lib/components/library/AsyncImageLoading.test.ts
Normal file
431
src/lib/components/library/AsyncImageLoading.test.ts
Normal file
@ -0,0 +1,431 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
import { render, waitFor } from "@testing-library/svelte";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Integration tests for async image loading pattern used in components
|
||||||
|
*
|
||||||
|
* Pattern:
|
||||||
|
* - Component has $state<string> imageUrl = ""
|
||||||
|
* - Component has async loadImageUrl() function
|
||||||
|
* - Component uses $effect to call loadImageUrl when dependencies change
|
||||||
|
* - For lists: uses Map<string, string> to cache URLs per item
|
||||||
|
*/
|
||||||
|
|
||||||
|
// Mock repository with getImageUrl
|
||||||
|
const createMockRepository = () => ({
|
||||||
|
getImageUrl: vi.fn(),
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Async Image Loading Pattern", () => {
|
||||||
|
let mockRepository: any;
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
mockRepository = createMockRepository();
|
||||||
|
vi.clearAllMocks();
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
vi.clearAllTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Single Image Loading", () => {
|
||||||
|
it("should load image URL asynchronously on component mount", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulating component with async image loading
|
||||||
|
const imageUrl = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
|
||||||
|
expect(imageUrl).toBe("https://server.com/image.jpg");
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith("item123", "Primary");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should show placeholder while loading", async () => {
|
||||||
|
mockRepository.getImageUrl.mockImplementation(
|
||||||
|
() => new Promise((resolve) => setTimeout(() => resolve("https://server.com/image.jpg"), 100))
|
||||||
|
);
|
||||||
|
|
||||||
|
vi.useFakeTimers();
|
||||||
|
const promise = mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
|
||||||
|
// Initially no URL
|
||||||
|
expect(promise).toBeInstanceOf(Promise);
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
vi.useRealTimers();
|
||||||
|
|
||||||
|
const result = await promise;
|
||||||
|
expect(result).toBe("https://server.com/image.jpg");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should reload image when item changes", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image1.jpg");
|
||||||
|
|
||||||
|
const url1 = await mockRepository.getImageUrl("item1", "Primary");
|
||||||
|
expect(url1).toBe("https://server.com/image1.jpg");
|
||||||
|
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image2.jpg");
|
||||||
|
|
||||||
|
const url2 = await mockRepository.getImageUrl("item2", "Primary");
|
||||||
|
expect(url2).toBe("https://server.com/image2.jpg");
|
||||||
|
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not reload image if item ID hasn't changed", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// First load
|
||||||
|
await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
|
||||||
|
// Would normally use $effect to track changes
|
||||||
|
// If item ID is same, should not reload (handled by component caching)
|
||||||
|
// This test documents the expected behavior
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle load errors gracefully", async () => {
|
||||||
|
mockRepository.getImageUrl.mockRejectedValue(new Error("Network error"));
|
||||||
|
|
||||||
|
// Component should catch error and show placeholder
|
||||||
|
try {
|
||||||
|
await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
} catch (e) {
|
||||||
|
expect(e).toBeInstanceOf(Error);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("List Image Caching (Map-based)", () => {
|
||||||
|
it("should cache URLs using Map<string, string>", () => {
|
||||||
|
// Simulating component state: imageUrls = $state<Map<string, string>>(new Map())
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
|
||||||
|
// Load first item
|
||||||
|
imageUrls.set("item1", "https://server.com/image1.jpg");
|
||||||
|
expect(imageUrls.has("item1")).toBe(true);
|
||||||
|
expect(imageUrls.get("item1")).toBe("https://server.com/image1.jpg");
|
||||||
|
|
||||||
|
// Load second item
|
||||||
|
imageUrls.set("item2", "https://server.com/image2.jpg");
|
||||||
|
expect(imageUrls.size).toBe(2);
|
||||||
|
|
||||||
|
// Check cache hit
|
||||||
|
expect(imageUrls.get("item1")).toBe("https://server.com/image1.jpg");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should load images only once per item", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
|
||||||
|
// Simulate loading multiple items
|
||||||
|
const items = [
|
||||||
|
{ id: "item1", name: "Album 1" },
|
||||||
|
{ id: "item2", name: "Album 2" },
|
||||||
|
{ id: "item1", name: "Album 1 (again)" }, // Same ID
|
||||||
|
];
|
||||||
|
|
||||||
|
for (const item of items) {
|
||||||
|
if (!imageUrls.has(item.id)) {
|
||||||
|
const url = await mockRepository.getImageUrl(item.id, "Primary");
|
||||||
|
imageUrls.set(item.id, url);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should only call once per unique ID
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should update single item without affecting others", async () => {
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
|
||||||
|
imageUrls.set("item1", "https://server.com/image1.jpg");
|
||||||
|
imageUrls.set("item2", "https://server.com/image2.jpg");
|
||||||
|
imageUrls.set("item3", "https://server.com/image3.jpg");
|
||||||
|
|
||||||
|
// Update item2
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image2_updated.jpg");
|
||||||
|
const newUrl = await mockRepository.getImageUrl("item2", "Primary");
|
||||||
|
imageUrls.set("item2", newUrl);
|
||||||
|
|
||||||
|
// Others should remain unchanged
|
||||||
|
expect(imageUrls.get("item1")).toBe("https://server.com/image1.jpg");
|
||||||
|
expect(imageUrls.get("item2")).toBe("https://server.com/image2_updated.jpg");
|
||||||
|
expect(imageUrls.get("item3")).toBe("https://server.com/image3.jpg");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should clear cache when data changes", () => {
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
|
||||||
|
imageUrls.set("item1", "https://server.com/image1.jpg");
|
||||||
|
imageUrls.set("item2", "https://server.com/image2.jpg");
|
||||||
|
|
||||||
|
// Clear cache
|
||||||
|
imageUrls.clear();
|
||||||
|
|
||||||
|
expect(imageUrls.size).toBe(0);
|
||||||
|
expect(imageUrls.has("item1")).toBe(false);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support Map operations efficiently", () => {
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
|
||||||
|
// Add items
|
||||||
|
for (let i = 0; i < 100; i++) {
|
||||||
|
imageUrls.set(`item${i}`, `https://server.com/image${i}.jpg`);
|
||||||
|
}
|
||||||
|
|
||||||
|
expect(imageUrls.size).toBe(100);
|
||||||
|
|
||||||
|
// Check specific item
|
||||||
|
expect(imageUrls.has("item50")).toBe(true);
|
||||||
|
expect(imageUrls.get("item50")).toBe("https://server.com/image50.jpg");
|
||||||
|
|
||||||
|
// Iterate
|
||||||
|
let count = 0;
|
||||||
|
imageUrls.forEach(() => {
|
||||||
|
count++;
|
||||||
|
});
|
||||||
|
expect(count).toBe(100);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Component Lifecycle ($effect integration)", () => {
|
||||||
|
it("should trigger load on prop change", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulate $effect tracking prop changes
|
||||||
|
let effectCount = 0;
|
||||||
|
const trackingEffect = vi.fn(() => {
|
||||||
|
effectCount++;
|
||||||
|
return mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
});
|
||||||
|
|
||||||
|
trackingEffect();
|
||||||
|
expect(effectCount).toBe(1);
|
||||||
|
|
||||||
|
trackingEffect();
|
||||||
|
expect(effectCount).toBe(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should skip load if conditions not met", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulate conditional loading (e.g., if (!imageUrl && primaryImageTag))
|
||||||
|
let imageUrl = "";
|
||||||
|
const primaryImageTag = "";
|
||||||
|
|
||||||
|
if (!imageUrl && primaryImageTag) {
|
||||||
|
imageUrl = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
}
|
||||||
|
|
||||||
|
expect(mockRepository.getImageUrl).not.toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle dependent state updates", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulate component state changes triggering effects
|
||||||
|
const state = {
|
||||||
|
item: { id: "item1", primaryImageTag: "tag1" },
|
||||||
|
imageUrl: "",
|
||||||
|
};
|
||||||
|
|
||||||
|
const loadImage = async () => {
|
||||||
|
if (state.item.primaryImageTag) {
|
||||||
|
state.imageUrl = await mockRepository.getImageUrl(state.item.id, "Primary");
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
await loadImage();
|
||||||
|
expect(state.imageUrl).toBe("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Change item
|
||||||
|
state.item = { id: "item2", primaryImageTag: "tag2" };
|
||||||
|
state.imageUrl = "";
|
||||||
|
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image2.jpg");
|
||||||
|
await loadImage();
|
||||||
|
expect(state.imageUrl).toBe("https://server.com/image2.jpg");
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Error Handling in Async Loading", () => {
|
||||||
|
it("should set empty string on error", async () => {
|
||||||
|
mockRepository.getImageUrl.mockRejectedValue(new Error("Network error"));
|
||||||
|
|
||||||
|
let imageUrl = "";
|
||||||
|
|
||||||
|
try {
|
||||||
|
imageUrl = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
} catch {
|
||||||
|
imageUrl = ""; // Set to empty on error
|
||||||
|
}
|
||||||
|
|
||||||
|
expect(imageUrl).toBe("");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should allow retry after error", async () => {
|
||||||
|
mockRepository.getImageUrl
|
||||||
|
.mockRejectedValueOnce(new Error("Network error"))
|
||||||
|
.mockResolvedValueOnce("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
let imageUrl = "";
|
||||||
|
|
||||||
|
// First attempt fails
|
||||||
|
try {
|
||||||
|
imageUrl = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
} catch {
|
||||||
|
imageUrl = "";
|
||||||
|
}
|
||||||
|
|
||||||
|
// Retry succeeds
|
||||||
|
imageUrl = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
expect(imageUrl).toBe("https://server.com/image.jpg");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle concurrent load requests", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulate loading multiple images concurrently
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
const items = [
|
||||||
|
{ id: "item1" },
|
||||||
|
{ id: "item2" },
|
||||||
|
{ id: "item3" },
|
||||||
|
];
|
||||||
|
|
||||||
|
const promises = items.map(item =>
|
||||||
|
mockRepository.getImageUrl(item.id, "Primary")
|
||||||
|
.then(url => imageUrls.set(item.id, url))
|
||||||
|
.catch(() => imageUrls.set(item.id, ""))
|
||||||
|
);
|
||||||
|
|
||||||
|
await Promise.all(promises);
|
||||||
|
|
||||||
|
expect(imageUrls.size).toBe(3);
|
||||||
|
expect(imageUrls.has("item1")).toBe(true);
|
||||||
|
expect(imageUrls.has("item2")).toBe(true);
|
||||||
|
expect(imageUrls.has("item3")).toBe(true);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Performance Characteristics", () => {
|
||||||
|
it("should not reload unnecessarily", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
// Simulate $effect with dependency tracking
|
||||||
|
let dependencyValue = "same";
|
||||||
|
let previousDependency = "same";
|
||||||
|
|
||||||
|
const loadImage = async () => {
|
||||||
|
if (dependencyValue !== previousDependency) {
|
||||||
|
previousDependency = dependencyValue;
|
||||||
|
return await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
await loadImage();
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
|
||||||
|
// No change in dependency
|
||||||
|
await loadImage();
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
|
||||||
|
// Change dependency
|
||||||
|
dependencyValue = "changed";
|
||||||
|
await loadImage();
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle large lists efficiently", async () => {
|
||||||
|
const imageUrls = new Map<string, string>();
|
||||||
|
let loadCount = 0;
|
||||||
|
|
||||||
|
mockRepository.getImageUrl.mockImplementation(() => {
|
||||||
|
loadCount++;
|
||||||
|
return Promise.resolve("https://server.com/image.jpg");
|
||||||
|
});
|
||||||
|
|
||||||
|
// Simulate loading 1000 items but caching URLs
|
||||||
|
const items = Array.from({ length: 1000 }, (_, i) => ({ id: `item${i % 10}` }));
|
||||||
|
|
||||||
|
for (const item of items) {
|
||||||
|
if (!imageUrls.has(item.id)) {
|
||||||
|
const url = await mockRepository.getImageUrl(item.id, "Primary");
|
||||||
|
imageUrls.set(item.id, url);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should only load 10 unique images
|
||||||
|
expect(loadCount).toBe(10);
|
||||||
|
expect(imageUrls.size).toBe(10);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not block rendering during async loading", () => {
|
||||||
|
mockRepository.getImageUrl.mockImplementation(
|
||||||
|
() => new Promise((resolve) =>
|
||||||
|
setTimeout(() => resolve("https://server.com/image.jpg"), 1000)
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
// Async operation should not block component rendering
|
||||||
|
const renderTiming = {
|
||||||
|
startRender: Date.now(),
|
||||||
|
loadStart: null as number | null,
|
||||||
|
loadComplete: null as number | null,
|
||||||
|
};
|
||||||
|
|
||||||
|
// Render happens immediately
|
||||||
|
renderTiming.startRender = Date.now();
|
||||||
|
|
||||||
|
// Load happens asynchronously
|
||||||
|
mockRepository.getImageUrl("item123", "Primary").then(() => {
|
||||||
|
renderTiming.loadComplete = Date.now();
|
||||||
|
});
|
||||||
|
|
||||||
|
// Render should complete before load finishes
|
||||||
|
expect(Date.now() - renderTiming.startRender).toBeLessThan(1000);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Backend Integration", () => {
|
||||||
|
it("should call backend with correct parameters", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image.jpg");
|
||||||
|
|
||||||
|
await mockRepository.getImageUrl("item123", "Primary", {
|
||||||
|
maxWidth: 300,
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith(
|
||||||
|
"item123",
|
||||||
|
"Primary",
|
||||||
|
{
|
||||||
|
maxWidth: 300,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle backend URL correctly", async () => {
|
||||||
|
const backendUrl = "https://server.com/Items/item123/Images/Primary?maxWidth=300&api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue(backendUrl);
|
||||||
|
|
||||||
|
const url = await mockRepository.getImageUrl("item123", "Primary", { maxWidth: 300 });
|
||||||
|
|
||||||
|
expect(url).toBe(backendUrl);
|
||||||
|
// Frontend never constructs URLs directly
|
||||||
|
expect(url).toContain("api_key=");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not require URL construction in frontend", async () => {
|
||||||
|
// Frontend receives pre-constructed URL from backend
|
||||||
|
const preConstructedUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue(preConstructedUrl);
|
||||||
|
|
||||||
|
const url = await mockRepository.getImageUrl("item123", "Primary");
|
||||||
|
|
||||||
|
// Frontend just uses the URL
|
||||||
|
expect(url).toContain("https://");
|
||||||
|
expect(url).toContain("item123");
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@ -10,6 +10,9 @@
|
|||||||
|
|
||||||
let { people, title = "Cast & Crew" }: Props = $props();
|
let { people, title = "Cast & Crew" }: Props = $props();
|
||||||
|
|
||||||
|
// Map of person IDs to their image URLs, loaded asynchronously
|
||||||
|
let personImageUrls = $state<Map<string, string>>(new Map());
|
||||||
|
|
||||||
// Group people by type
|
// Group people by type
|
||||||
const groupedPeople = $derived.by(() => {
|
const groupedPeople = $derived.by(() => {
|
||||||
const groups: Record<string, Person[]> = {
|
const groups: Record<string, Person[]> = {
|
||||||
@ -58,18 +61,31 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function getPersonImageUrl(person: Person): string {
|
// Load image URL for a single person
|
||||||
|
async function loadPersonImageUrl(person: Person): Promise<void> {
|
||||||
|
if (!person.primaryImageTag || personImageUrls.has(person.id)) return;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(person.id, "Primary", {
|
const url = await repo.getImageUrl(person.id, "Primary", {
|
||||||
maxWidth: 200,
|
maxWidth: 200,
|
||||||
tag: person.primaryImageTag,
|
tag: person.primaryImageTag,
|
||||||
});
|
});
|
||||||
|
personImageUrls.set(person.id, url);
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
personImageUrls.set(person.id, "");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load image URLs for all people
|
||||||
|
$effect(() => {
|
||||||
|
people.forEach((person) => {
|
||||||
|
if (person.primaryImageTag && !personImageUrls.has(person.id)) {
|
||||||
|
loadPersonImageUrl(person);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
function handlePersonClick(person: Person) {
|
function handlePersonClick(person: Person) {
|
||||||
goto(`/library/${person.id}`);
|
goto(`/library/${person.id}`);
|
||||||
}
|
}
|
||||||
@ -94,9 +110,9 @@
|
|||||||
>
|
>
|
||||||
<!-- Person image -->
|
<!-- Person image -->
|
||||||
<div class="w-24 h-24 rounded-full overflow-hidden bg-[var(--color-surface)] mb-2">
|
<div class="w-24 h-24 rounded-full overflow-hidden bg-[var(--color-surface)] mb-2">
|
||||||
{#if person.primaryImageTag}
|
{#if person.primaryImageTag && personImageUrls.get(person.id)}
|
||||||
<img
|
<img
|
||||||
src={getPersonImageUrl(person)}
|
src={personImageUrls.get(person.id)}
|
||||||
alt={person.name}
|
alt={person.name}
|
||||||
class="w-full h-full object-cover group-hover:scale-110 transition-transform"
|
class="w-full h-full object-cover group-hover:scale-110 transition-transform"
|
||||||
loading="lazy"
|
loading="lazy"
|
||||||
|
|||||||
@ -12,6 +12,9 @@
|
|||||||
|
|
||||||
let { episode, series, allEpisodes, onBack }: Props = $props();
|
let { episode, series, allEpisodes, onBack }: Props = $props();
|
||||||
|
|
||||||
|
let backdropUrl = $state<string>("");
|
||||||
|
let episodeThumbnailUrls = $state<Map<string, string>>(new Map());
|
||||||
|
|
||||||
// Check if an episode matches the focused episode (by ID or season/episode number)
|
// Check if an episode matches the focused episode (by ID or season/episode number)
|
||||||
function isCurrentEpisode(ep: MediaItem): boolean {
|
function isCurrentEpisode(ep: MediaItem): boolean {
|
||||||
if (ep.id === episode.id) return true;
|
if (ep.id === episode.id) return true;
|
||||||
@ -70,52 +73,74 @@
|
|||||||
return allEpisodes.slice(start, end);
|
return allEpisodes.slice(start, end);
|
||||||
});
|
});
|
||||||
|
|
||||||
function getBackdropUrl(): string {
|
// Load backdrop URL asynchronously
|
||||||
|
async function loadBackdropUrl(): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
|
|
||||||
// Try episode backdrop first
|
// Try episode backdrop first
|
||||||
if (episode.backdropImageTags?.[0]) {
|
if (episode.backdropImageTags?.[0]) {
|
||||||
return repo.getImageUrl(episode.id, "Backdrop", {
|
backdropUrl = await repo.getImageUrl(episode.id, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: episode.backdropImageTags[0],
|
tag: episode.backdropImageTags[0],
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try episode primary (thumbnail)
|
// Try episode primary (thumbnail)
|
||||||
if (episode.primaryImageTag) {
|
if (episode.primaryImageTag) {
|
||||||
return repo.getImageUrl(episode.id, "Primary", {
|
backdropUrl = await repo.getImageUrl(episode.id, "Primary", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: episode.primaryImageTag,
|
tag: episode.primaryImageTag,
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Fall back to series backdrop
|
// Fall back to series backdrop
|
||||||
if (series.backdropImageTags?.[0]) {
|
if (series.backdropImageTags?.[0]) {
|
||||||
return repo.getImageUrl(series.id, "Backdrop", {
|
backdropUrl = await repo.getImageUrl(series.id, "Backdrop", {
|
||||||
maxWidth: 1920,
|
maxWidth: 1920,
|
||||||
tag: series.backdropImageTags[0],
|
tag: series.backdropImageTags[0],
|
||||||
});
|
});
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
return "";
|
backdropUrl = "";
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
backdropUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function getEpisodeThumbnail(ep: MediaItem): string {
|
// Load episode thumbnail URL for a single episode
|
||||||
|
async function loadEpisodeThumbnailUrl(ep: MediaItem): Promise<void> {
|
||||||
|
if (!ep.primaryImageTag || episodeThumbnailUrls.has(ep.id)) return;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(ep.id, "Primary", {
|
const url = await repo.getImageUrl(ep.id, "Primary", {
|
||||||
maxWidth: 400,
|
maxWidth: 400,
|
||||||
tag: ep.primaryImageTag,
|
tag: ep.primaryImageTag,
|
||||||
});
|
});
|
||||||
|
episodeThumbnailUrls.set(ep.id, url);
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
episodeThumbnailUrls.set(ep.id, "");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load backdrop when episode changes
|
||||||
|
$effect(() => {
|
||||||
|
loadBackdropUrl();
|
||||||
|
});
|
||||||
|
|
||||||
|
// Load episode thumbnail URLs when adjacent episodes change
|
||||||
|
$effect(() => {
|
||||||
|
adjacentEpisodes().forEach((ep) => {
|
||||||
|
if (ep.primaryImageTag && !episodeThumbnailUrls.has(ep.id)) {
|
||||||
|
loadEpisodeThumbnailUrl(ep);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
function formatDuration(ticks?: number): string {
|
function formatDuration(ticks?: number): string {
|
||||||
if (!ticks) return "";
|
if (!ticks) return "";
|
||||||
const seconds = Math.floor(ticks / 10000000);
|
const seconds = Math.floor(ticks / 10000000);
|
||||||
@ -143,7 +168,6 @@
|
|||||||
goto(`/library/${series.id}?episode=${ep.id}`);
|
goto(`/library/${series.id}?episode=${ep.id}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
const backdropUrl = $derived(getBackdropUrl());
|
|
||||||
const episodeLabel = $derived(
|
const episodeLabel = $derived(
|
||||||
`S${episode.parentIndexNumber || 1}E${episode.indexNumber || 1}`
|
`S${episode.parentIndexNumber || 1}E${episode.indexNumber || 1}`
|
||||||
);
|
);
|
||||||
@ -264,7 +288,7 @@
|
|||||||
{#each adjacentEpisodes() as ep (ep.id)}
|
{#each adjacentEpisodes() as ep (ep.id)}
|
||||||
{@const isCurrent = isCurrentEpisode(ep)}
|
{@const isCurrent = isCurrentEpisode(ep)}
|
||||||
{@const epProgress = getProgress(ep)}
|
{@const epProgress = getProgress(ep)}
|
||||||
{@const thumbUrl = getEpisodeThumbnail(ep)}
|
{@const thumbUrl = episodeThumbnailUrls.get(ep.id) ?? ""}
|
||||||
<button
|
<button
|
||||||
onclick={() => !isCurrent && handleEpisodeClick(ep)}
|
onclick={() => !isCurrent && handleEpisodeClick(ep)}
|
||||||
class="flex-shrink-0 w-64 text-left group/card {isCurrent ? 'ring-2 ring-yellow-400 rounded-lg' : ''}"
|
class="flex-shrink-0 w-64 text-left group/card {isCurrent ? 'ring-2 ring-yellow-400 rounded-lg' : ''}"
|
||||||
|
|||||||
@ -3,6 +3,7 @@
|
|||||||
import type { MediaItem } from "$lib/api/types";
|
import type { MediaItem } from "$lib/api/types";
|
||||||
import { auth } from "$lib/stores/auth";
|
import { auth } from "$lib/stores/auth";
|
||||||
import { downloads } from "$lib/stores/downloads";
|
import { downloads } from "$lib/stores/downloads";
|
||||||
|
import { formatDuration } from "$lib/utils/duration";
|
||||||
import VideoDownloadButton from "./VideoDownloadButton.svelte";
|
import VideoDownloadButton from "./VideoDownloadButton.svelte";
|
||||||
|
|
||||||
interface Props {
|
interface Props {
|
||||||
@ -14,6 +15,7 @@
|
|||||||
let { episode, focused = false, onclick }: Props = $props();
|
let { episode, focused = false, onclick }: Props = $props();
|
||||||
|
|
||||||
let buttonRef: HTMLButtonElement | null = null;
|
let buttonRef: HTMLButtonElement | null = null;
|
||||||
|
let imageUrl = $state<string>("");
|
||||||
|
|
||||||
onMount(() => {
|
onMount(() => {
|
||||||
if (focused && buttonRef) {
|
if (focused && buttonRef) {
|
||||||
@ -35,39 +37,31 @@
|
|||||||
);
|
);
|
||||||
const downloadProgress = $derived(downloadInfo?.progress || 0);
|
const downloadProgress = $derived(downloadInfo?.progress || 0);
|
||||||
|
|
||||||
function getImageUrl(): string {
|
// Load image URL asynchronously
|
||||||
|
async function loadImageUrl(): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(episode.id, "Primary", {
|
imageUrl = await repo.getImageUrl(episode.id, "Primary", {
|
||||||
maxWidth: 320,
|
maxWidth: 320,
|
||||||
tag: episode.primaryImageTag,
|
tag: episode.primaryImageTag,
|
||||||
});
|
});
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function getProgress(): number {
|
// Load image when episode changes
|
||||||
|
$effect(() => {
|
||||||
|
loadImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
|
const progress = $derived(() => {
|
||||||
if (!episode.userData || !episode.runTimeTicks) {
|
if (!episode.userData || !episode.runTimeTicks) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
return (episode.userData.playbackPositionTicks / episode.runTimeTicks) * 100;
|
return (episode.userData.playbackPositionTicks / episode.runTimeTicks) * 100;
|
||||||
}
|
});
|
||||||
|
|
||||||
function formatDuration(ticks?: number): string {
|
|
||||||
if (!ticks) return "";
|
|
||||||
const seconds = Math.floor(ticks / 10000000);
|
|
||||||
const hours = Math.floor(seconds / 3600);
|
|
||||||
const minutes = Math.floor((seconds % 3600) / 60);
|
|
||||||
|
|
||||||
if (hours > 0) {
|
|
||||||
return `${hours}h ${minutes}m`;
|
|
||||||
}
|
|
||||||
return `${minutes}m`;
|
|
||||||
}
|
|
||||||
|
|
||||||
const imageUrl = $derived(getImageUrl());
|
|
||||||
const progress = $derived(getProgress());
|
|
||||||
const duration = $derived(formatDuration(episode.runTimeTicks));
|
const duration = $derived(formatDuration(episode.runTimeTicks));
|
||||||
const episodeNumber = $derived(episode.indexNumber || 0);
|
const episodeNumber = $derived(episode.indexNumber || 0);
|
||||||
</script>
|
</script>
|
||||||
@ -107,11 +101,11 @@
|
|||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Progress bar -->
|
<!-- Progress bar -->
|
||||||
{#if progress > 0}
|
{#if progress() > 0}
|
||||||
<div class="absolute bottom-0 left-0 right-0 h-1 bg-gray-800">
|
<div class="absolute bottom-0 left-0 right-0 h-1 bg-gray-800">
|
||||||
<div
|
<div
|
||||||
class="h-full bg-[var(--color-jellyfin)]"
|
class="h-full bg-[var(--color-jellyfin)]"
|
||||||
style="width: {progress}%"
|
style="width: {progress()}%"
|
||||||
></div>
|
></div>
|
||||||
</div>
|
</div>
|
||||||
{/if}
|
{/if}
|
||||||
|
|||||||
@ -41,6 +41,7 @@
|
|||||||
let selectedGenre = $state<Genre | null>(null);
|
let selectedGenre = $state<Genre | null>(null);
|
||||||
let genreItems = $state<MediaItem[]>([]);
|
let genreItems = $state<MediaItem[]>([]);
|
||||||
let loadingItems = $state(false);
|
let loadingItems = $state(false);
|
||||||
|
let genreItemImageUrls = $state<Map<string, string>>(new Map());
|
||||||
|
|
||||||
const { markLoaded } = useServerReachabilityReload(async () => {
|
const { markLoaded } = useServerReachabilityReload(async () => {
|
||||||
await loadGenres();
|
await loadGenres();
|
||||||
@ -79,6 +80,7 @@
|
|||||||
try {
|
try {
|
||||||
loadingItems = true;
|
loadingItems = true;
|
||||||
selectedGenre = genre;
|
selectedGenre = genre;
|
||||||
|
genreItemImageUrls = new Map(); // Clear image URLs when loading new genre
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
const result = await repo.getItems($currentLibrary.id, {
|
const result = await repo.getItems($currentLibrary.id, {
|
||||||
includeItemTypes: config.itemTypes,
|
includeItemTypes: config.itemTypes,
|
||||||
@ -96,6 +98,31 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load image URL for a single item
|
||||||
|
async function loadGenreItemImage(item: MediaItem): Promise<void> {
|
||||||
|
if (!item.primaryImageTag || genreItemImageUrls.has(item.id)) return;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const repo = auth.getRepository();
|
||||||
|
const url = await repo.getImageUrl(item.id, "Primary", {
|
||||||
|
maxWidth: 300,
|
||||||
|
tag: item.primaryImageTag,
|
||||||
|
});
|
||||||
|
genreItemImageUrls.set(item.id, url);
|
||||||
|
} catch {
|
||||||
|
genreItemImageUrls.set(item.id, "");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Load image URLs for all genre items
|
||||||
|
$effect(() => {
|
||||||
|
genreItems.forEach((item) => {
|
||||||
|
if (item.primaryImageTag && !genreItemImageUrls.has(item.id)) {
|
||||||
|
loadGenreItemImage(item);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
function applyFilter() {
|
function applyFilter() {
|
||||||
let result = [...genres];
|
let result = [...genres];
|
||||||
|
|
||||||
@ -217,12 +244,9 @@
|
|||||||
{#each genreItems as item (item.id)}
|
{#each genreItems as item (item.id)}
|
||||||
<button onclick={() => handleItemClick(item)} class="group text-left">
|
<button onclick={() => handleItemClick(item)} class="group text-left">
|
||||||
<div class="{aspectRatioClass} bg-[var(--color-surface)] rounded-lg overflow-hidden mb-2">
|
<div class="{aspectRatioClass} bg-[var(--color-surface)] rounded-lg overflow-hidden mb-2">
|
||||||
{#if item.primaryImageTag}
|
{#if item.primaryImageTag && genreItemImageUrls.get(item.id)}
|
||||||
<img
|
<img
|
||||||
src={auth.getRepository().getImageUrl(item.id, "Primary", {
|
src={genreItemImageUrls.get(item.id)}
|
||||||
maxWidth: 300,
|
|
||||||
tag: item.primaryImageTag,
|
|
||||||
})}
|
|
||||||
alt={item.name}
|
alt={item.name}
|
||||||
class="w-full h-full object-cover group-hover:scale-105 transition-transform"
|
class="w-full h-full object-cover group-hover:scale-105 transition-transform"
|
||||||
/>
|
/>
|
||||||
|
|||||||
@ -30,7 +30,6 @@
|
|||||||
sortOptions: Array<{ key: string; label: string }>; // Jellyfin field names
|
sortOptions: Array<{ key: string; label: string }>; // Jellyfin field names
|
||||||
defaultSort: string; // Jellyfin field name (e.g., "SortName")
|
defaultSort: string; // Jellyfin field name (e.g., "SortName")
|
||||||
displayComponent: "grid" | "tracklist"; // Which component to use
|
displayComponent: "grid" | "tracklist"; // Which component to use
|
||||||
searchFields: string[]; // Which fields to search in: ["name", "artists"], etc.
|
|
||||||
}
|
}
|
||||||
|
|
||||||
interface Props {
|
interface Props {
|
||||||
@ -42,8 +41,10 @@
|
|||||||
let items = $state<MediaItem[]>([]);
|
let items = $state<MediaItem[]>([]);
|
||||||
let loading = $state(true);
|
let loading = $state(true);
|
||||||
let searchQuery = $state("");
|
let searchQuery = $state("");
|
||||||
|
let debouncedSearchQuery = $state("");
|
||||||
let sortBy = $state<string>(config.defaultSort);
|
let sortBy = $state<string>(config.defaultSort);
|
||||||
let sortOrder = $state<"Ascending" | "Descending">("Ascending");
|
let sortOrder = $state<"Ascending" | "Descending">("Ascending");
|
||||||
|
let searchTimeout: ReturnType<typeof setTimeout> | null = null;
|
||||||
|
|
||||||
const { markLoaded } = useServerReachabilityReload(async () => {
|
const { markLoaded } = useServerReachabilityReload(async () => {
|
||||||
await loadItems();
|
await loadItems();
|
||||||
@ -65,8 +66,8 @@
|
|||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
|
|
||||||
// Use backend search if search query is provided, otherwise use getItems with sort
|
// Use backend search if search query is provided, otherwise use getItems with sort
|
||||||
if (searchQuery.trim()) {
|
if (debouncedSearchQuery.trim()) {
|
||||||
const result = await repo.search(searchQuery, {
|
const result = await repo.search(debouncedSearchQuery, {
|
||||||
includeItemTypes: [config.itemType],
|
includeItemTypes: [config.itemType],
|
||||||
limit: 10000,
|
limit: 10000,
|
||||||
});
|
});
|
||||||
@ -90,9 +91,18 @@
|
|||||||
|
|
||||||
function handleSearch(query: string) {
|
function handleSearch(query: string) {
|
||||||
searchQuery = query;
|
searchQuery = query;
|
||||||
loadItems();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Debounce search input (300ms delay)
|
||||||
|
$effect(() => {
|
||||||
|
if (searchTimeout) clearTimeout(searchTimeout);
|
||||||
|
|
||||||
|
searchTimeout = setTimeout(() => {
|
||||||
|
debouncedSearchQuery = searchQuery;
|
||||||
|
loadItems();
|
||||||
|
}, 300);
|
||||||
|
});
|
||||||
|
|
||||||
function handleSort(newSort: string) {
|
function handleSort(newSort: string) {
|
||||||
sortBy = newSort;
|
sortBy = newSort;
|
||||||
loadItems();
|
loadItems();
|
||||||
|
|||||||
661
src/lib/components/library/GenericMediaListPage.test.ts
Normal file
661
src/lib/components/library/GenericMediaListPage.test.ts
Normal file
@ -0,0 +1,661 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
import { render, screen, fireEvent, waitFor } from "@testing-library/svelte";
|
||||||
|
import GenericMediaListPage from "./GenericMediaListPage.svelte";
|
||||||
|
|
||||||
|
// Mock SvelteKit navigation
|
||||||
|
vi.mock("$app/navigation", () => ({
|
||||||
|
goto: vi.fn(),
|
||||||
|
}));
|
||||||
|
|
||||||
|
// Mock stores
|
||||||
|
vi.mock("$lib/stores/library", () => ({
|
||||||
|
currentLibrary: {
|
||||||
|
subscribe: vi.fn((fn) => {
|
||||||
|
fn({ id: "lib123", name: "Music" });
|
||||||
|
return vi.fn();
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
vi.mock("$lib/stores/auth", () => ({
|
||||||
|
auth: {
|
||||||
|
getRepository: vi.fn(() => ({
|
||||||
|
getItems: vi.fn(),
|
||||||
|
search: vi.fn(),
|
||||||
|
})),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
vi.mock("$lib/composables/useServerReachabilityReload", () => ({
|
||||||
|
useServerReachabilityReload: vi.fn(() => ({
|
||||||
|
markLoaded: vi.fn(),
|
||||||
|
})),
|
||||||
|
}));
|
||||||
|
|
||||||
|
describe("GenericMediaListPage", () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
vi.clearAllTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Component Initialization", () => {
|
||||||
|
it("should render with title and search bar", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
const heading = screen.getByText("Tracks");
|
||||||
|
expect(heading).toBeTruthy();
|
||||||
|
|
||||||
|
const searchInput = container.querySelector('input[type="text"]');
|
||||||
|
expect(searchInput).toBeTruthy();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should load items on mount", async () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
// loadItems should have been called
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should display sort options", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "MusicAlbum",
|
||||||
|
title: "Albums",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search albums...",
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
{ key: "Artist", label: "Artist" },
|
||||||
|
{ key: "ProductionYear", label: "Year" },
|
||||||
|
],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "grid" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Check that all sort options are rendered
|
||||||
|
const titleOption = screen.queryByText("Title");
|
||||||
|
expect(titleOption).toBeTruthy();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Search Functionality", () => {
|
||||||
|
it("should debounce search input for 300ms", async () => {
|
||||||
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
const searchInput = container.querySelector("input") as HTMLInputElement;
|
||||||
|
|
||||||
|
// Type into search
|
||||||
|
fireEvent.input(searchInput, { target: { value: "t" } });
|
||||||
|
expect(searchInput.value).toBe("t");
|
||||||
|
|
||||||
|
// Search should not trigger immediately
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// Add more characters
|
||||||
|
fireEvent.input(searchInput, { target: { value: "test" } });
|
||||||
|
|
||||||
|
// Still shouldn't trigger (only 100ms passed total)
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// Now advance to 300ms total - search should trigger
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
// Search should have been debounced
|
||||||
|
});
|
||||||
|
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use backend search when search query is provided", async () => {
|
||||||
|
const mockSearchFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [{ id: "item1", name: "Test Track" }],
|
||||||
|
totalRecordCount: 1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: vi.fn(),
|
||||||
|
search: mockSearchFn,
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.useFakeTimers();
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
const searchInput = container.querySelector("input") as HTMLInputElement;
|
||||||
|
fireEvent.input(searchInput, { target: { value: "test" } });
|
||||||
|
|
||||||
|
// Advance timer to trigger debounced search
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockSearchFn).toHaveBeenCalledWith("test", expect.objectContaining({
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
limit: 10000,
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should use getItems without search for empty query", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockGetItemsFn).toHaveBeenCalledWith("lib123", expect.objectContaining({
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should clear previous search when input becomes empty", async () => {
|
||||||
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
const mockSearchFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockGetItemsFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: mockSearchFn,
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
const searchInput = container.querySelector("input") as HTMLInputElement;
|
||||||
|
|
||||||
|
// Type search query
|
||||||
|
fireEvent.input(searchInput, { target: { value: "test" } });
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Clear search
|
||||||
|
fireEvent.input(searchInput, { target: { value: "" } });
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
// Should call getItems when search is cleared
|
||||||
|
expect(mockGetItemsFn).toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Sorting Functionality", () => {
|
||||||
|
it("should pass sortBy parameter to backend", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "MusicAlbum",
|
||||||
|
title: "Albums",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search albums...",
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
{ key: "Artist", label: "Artist" },
|
||||||
|
],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "grid" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockGetItemsFn).toHaveBeenCalledWith("lib123", expect.objectContaining({
|
||||||
|
sortBy: "SortName",
|
||||||
|
sortOrder: "Ascending",
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should pass Jellyfin field names to backend (not custom compareFn)", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
{ key: "Artist", label: "Artist" },
|
||||||
|
{ key: "Album", label: "Album" },
|
||||||
|
{ key: "DatePlayed", label: "Recent" },
|
||||||
|
],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
const lastCall = (mockGetItemsFn as any).mock.calls[0];
|
||||||
|
const options = lastCall[1];
|
||||||
|
|
||||||
|
// Should pass Jellyfin field names directly
|
||||||
|
expect(typeof options.sortBy).toBe("string");
|
||||||
|
expect(["SortName", "Artist", "Album", "DatePlayed"]).toContain(options.sortBy);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("ItemType Filtering", () => {
|
||||||
|
it("should include correct itemType in getItems request", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockGetItemsFn).toHaveBeenCalledWith("lib123", expect.objectContaining({
|
||||||
|
includeItemTypes: ["Audio"],
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should include correct itemType in search request", async () => {
|
||||||
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
const mockSearchFn = vi.fn().mockResolvedValue({
|
||||||
|
items: [],
|
||||||
|
totalRecordCount: 0,
|
||||||
|
});
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: vi.fn(),
|
||||||
|
search: mockSearchFn,
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "MusicAlbum",
|
||||||
|
title: "Albums",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search albums...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "grid" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
const searchInput = container.querySelector("input") as HTMLInputElement;
|
||||||
|
fireEvent.input(searchInput, { target: { value: "album" } });
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockSearchFn).toHaveBeenCalledWith("album", expect.objectContaining({
|
||||||
|
includeItemTypes: ["MusicAlbum"],
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Loading State", () => {
|
||||||
|
it("should show loading indicator during data fetch", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn(
|
||||||
|
() => new Promise((resolve) => setTimeout(
|
||||||
|
() => resolve({ items: [], totalRecordCount: 0 }),
|
||||||
|
100
|
||||||
|
))
|
||||||
|
);
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.useFakeTimers();
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Component should be rendering (will show loading state internally)
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Error Handling", () => {
|
||||||
|
it("should handle backend errors gracefully", async () => {
|
||||||
|
const mockGetItemsFn = vi.fn().mockRejectedValue(new Error("Network error"));
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
// Should handle error without throwing
|
||||||
|
expect(mockGetItemsFn).toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle missing library gracefully", async () => {
|
||||||
|
const { goto } = await import("$app/navigation");
|
||||||
|
|
||||||
|
const mockGetItemsFn = vi.fn();
|
||||||
|
|
||||||
|
const mockRepository = {
|
||||||
|
getItems: mockGetItemsFn,
|
||||||
|
search: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((await import("$lib/stores/auth")).auth.getRepository).mockReturnValue(
|
||||||
|
mockRepository as any
|
||||||
|
);
|
||||||
|
|
||||||
|
// Mock currentLibrary to return null
|
||||||
|
vi.resetModules();
|
||||||
|
vi.mocked((await import("$lib/stores/library")).currentLibrary.subscribe).mockImplementation(
|
||||||
|
(fn: any) => {
|
||||||
|
fn(null);
|
||||||
|
return vi.fn();
|
||||||
|
}
|
||||||
|
);
|
||||||
|
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Should navigate to back path when library is missing
|
||||||
|
await waitFor(() => {
|
||||||
|
// goto would be called with config.backPath
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Display Component Props", () => {
|
||||||
|
it("should support grid display component", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "MusicAlbum",
|
||||||
|
title: "Albums",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search albums...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "grid" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should support tracklist display component", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Config Simplification", () => {
|
||||||
|
it("should not require searchFields in config", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [{ key: "SortName", label: "Title" }],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
// Note: searchFields is NOT present
|
||||||
|
};
|
||||||
|
|
||||||
|
// Should render without searchFields
|
||||||
|
expect(() => {
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
}).not.toThrow();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not require compareFn in sort options", () => {
|
||||||
|
const config = {
|
||||||
|
itemType: "Audio",
|
||||||
|
title: "Tracks",
|
||||||
|
backPath: "/library/music",
|
||||||
|
searchPlaceholder: "Search tracks...",
|
||||||
|
sortOptions: [
|
||||||
|
{ key: "SortName", label: "Title" },
|
||||||
|
// Note: no compareFn property
|
||||||
|
],
|
||||||
|
defaultSort: "SortName",
|
||||||
|
displayComponent: "tracklist" as const,
|
||||||
|
};
|
||||||
|
|
||||||
|
// Should render without compareFn in sort options
|
||||||
|
expect(() => {
|
||||||
|
render(GenericMediaListPage, {
|
||||||
|
props: { config },
|
||||||
|
});
|
||||||
|
}).not.toThrow();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@ -13,23 +13,37 @@
|
|||||||
|
|
||||||
let { items, showProgress = false, showDownloadStatus = true, onItemClick }: Props = $props();
|
let { items, showProgress = false, showDownloadStatus = true, onItemClick }: Props = $props();
|
||||||
|
|
||||||
|
// Map of item IDs to their image URLs, loaded asynchronously
|
||||||
|
let imageUrls = $state<Map<string, string>>(new Map());
|
||||||
|
|
||||||
function getDownloadInfo(itemId: string) {
|
function getDownloadInfo(itemId: string) {
|
||||||
return Object.values($downloads.downloads).find((d) => d.itemId === itemId);
|
return Object.values($downloads.downloads).find((d) => d.itemId === itemId);
|
||||||
}
|
}
|
||||||
|
|
||||||
function getImageUrl(item: MediaItem | Library): string {
|
// Load image URL for a single item
|
||||||
|
async function loadImageUrl(item: MediaItem | Library): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
const tag = "primaryImageTag" in item ? item.primaryImageTag : ("imageTag" in item ? item.imageTag : undefined);
|
const tag = "primaryImageTag" in item ? item.primaryImageTag : ("imageTag" in item ? item.imageTag : undefined);
|
||||||
return repo.getImageUrl(item.id, "Primary", {
|
const url = await repo.getImageUrl(item.id, "Primary", {
|
||||||
maxWidth: 80,
|
maxWidth: 80,
|
||||||
tag,
|
tag,
|
||||||
});
|
});
|
||||||
|
imageUrls.set(item.id, url);
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrls.set(item.id, "");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load image URLs whenever items change
|
||||||
|
$effect(() => {
|
||||||
|
items.forEach((item) => {
|
||||||
|
if (!imageUrls.has(item.id)) {
|
||||||
|
loadImageUrl(item);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
function getSubtitle(item: MediaItem | Library): string {
|
function getSubtitle(item: MediaItem | Library): string {
|
||||||
if (!("type" in item)) return "";
|
if (!("type" in item)) return "";
|
||||||
|
|
||||||
@ -66,7 +80,7 @@
|
|||||||
|
|
||||||
<div class="space-y-1">
|
<div class="space-y-1">
|
||||||
{#each items as item, index (item.id)}
|
{#each items as item, index (item.id)}
|
||||||
{@const imageUrl = getImageUrl(item)}
|
{@const imageUrl = imageUrls.get(item.id) ?? ""}
|
||||||
{@const subtitle = getSubtitle(item)}
|
{@const subtitle = getSubtitle(item)}
|
||||||
{@const duration = "runTimeTicks" in item ? formatDuration(item.runTimeTicks) : ""}
|
{@const duration = "runTimeTicks" in item ? formatDuration(item.runTimeTicks) : ""}
|
||||||
{@const progress = getProgress(item)}
|
{@const progress = getProgress(item)}
|
||||||
|
|||||||
@ -2,7 +2,6 @@
|
|||||||
import type { MediaItem, Library } from "$lib/api/types";
|
import type { MediaItem, Library } from "$lib/api/types";
|
||||||
import { auth } from "$lib/stores/auth";
|
import { auth } from "$lib/stores/auth";
|
||||||
import { downloads } from "$lib/stores/downloads";
|
import { downloads } from "$lib/stores/downloads";
|
||||||
import { getImageUrlSync } from "$lib/services/imageCache";
|
|
||||||
|
|
||||||
interface Props {
|
interface Props {
|
||||||
item: MediaItem | Library;
|
item: MediaItem | Library;
|
||||||
@ -14,6 +13,9 @@
|
|||||||
|
|
||||||
let { item, size = "medium", showProgress = false, showDownloadStatus = true, onclick }: Props = $props();
|
let { item, size = "medium", showProgress = false, showDownloadStatus = true, onclick }: Props = $props();
|
||||||
|
|
||||||
|
// Image URL state - loaded asynchronously
|
||||||
|
let imageUrl = $state<string>("");
|
||||||
|
|
||||||
// Check if this item is downloaded
|
// Check if this item is downloaded
|
||||||
const downloadInfo = $derived(
|
const downloadInfo = $derived(
|
||||||
Object.values($downloads.downloads).find((d) => d.itemId === item.id)
|
Object.values($downloads.downloads).find((d) => d.itemId === item.id)
|
||||||
@ -40,32 +42,35 @@
|
|||||||
return "aspect-video";
|
return "aspect-video";
|
||||||
});
|
});
|
||||||
|
|
||||||
function getImageUrl(): string {
|
// Load image URL asynchronously from backend
|
||||||
|
async function loadImageUrl(): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
const serverUrl = repo.serverUrl;
|
|
||||||
const id = item.id;
|
|
||||||
const tag = "primaryImageTag" in item ? item.primaryImageTag : ("imageTag" in item ? item.imageTag : undefined);
|
|
||||||
const maxWidth = size === "large" ? 400 : size === "medium" ? 300 : 200;
|
const maxWidth = size === "large" ? 400 : size === "medium" ? 300 : 200;
|
||||||
|
const tag = "primaryImageTag" in item ? item.primaryImageTag : ("imageTag" in item ? item.imageTag : undefined);
|
||||||
|
|
||||||
// Use the caching service - returns server URL immediately and triggers background caching
|
imageUrl = await repo.getImageUrl(item.id, "Primary", {
|
||||||
return getImageUrlSync(serverUrl, id, "Primary", {
|
|
||||||
maxWidth,
|
maxWidth,
|
||||||
tag,
|
tag,
|
||||||
});
|
});
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function getProgress(): number {
|
// Load image URL whenever item or size changes
|
||||||
|
$effect(() => {
|
||||||
|
loadImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
|
const progress = $derived(() => {
|
||||||
if (!showProgress || !("userData" in item) || !item.userData || !item.runTimeTicks) {
|
if (!showProgress || !("userData" in item) || !item.userData || !item.runTimeTicks) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
return (item.userData.playbackPositionTicks / item.runTimeTicks) * 100;
|
return (item.userData.playbackPositionTicks / item.runTimeTicks) * 100;
|
||||||
}
|
});
|
||||||
|
|
||||||
function getSubtitle(): string {
|
const subtitle = $derived(() => {
|
||||||
if (!("type" in item)) return "";
|
if (!("type" in item)) return "";
|
||||||
|
|
||||||
switch (item.type) {
|
switch (item.type) {
|
||||||
@ -82,11 +87,7 @@
|
|||||||
default:
|
default:
|
||||||
return "";
|
return "";
|
||||||
}
|
}
|
||||||
}
|
});
|
||||||
|
|
||||||
const imageUrl = $derived(getImageUrl());
|
|
||||||
const progress = $derived(getProgress());
|
|
||||||
const subtitle = $derived(getSubtitle());
|
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
<button
|
<button
|
||||||
@ -122,11 +123,11 @@
|
|||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Progress bar -->
|
<!-- Progress bar -->
|
||||||
{#if progress > 0}
|
{#if progress() > 0}
|
||||||
<div class="absolute bottom-0 left-0 right-0 h-1 bg-gray-800">
|
<div class="absolute bottom-0 left-0 right-0 h-1 bg-gray-800">
|
||||||
<div
|
<div
|
||||||
class="h-full bg-[var(--color-jellyfin)]"
|
class="h-full bg-[var(--color-jellyfin)]"
|
||||||
style="width: {progress}%"
|
style="width: {progress()}%"
|
||||||
></div>
|
></div>
|
||||||
</div>
|
</div>
|
||||||
{/if}
|
{/if}
|
||||||
@ -188,8 +189,8 @@
|
|||||||
<p class="text-sm font-medium text-white truncate group-hover/card:text-[var(--color-jellyfin)] transition-colors">
|
<p class="text-sm font-medium text-white truncate group-hover/card:text-[var(--color-jellyfin)] transition-colors">
|
||||||
{item.name}
|
{item.name}
|
||||||
</p>
|
</p>
|
||||||
{#if subtitle}
|
{#if subtitle()}
|
||||||
<p class="text-xs text-gray-400 truncate">{subtitle}</p>
|
<p class="text-xs text-gray-400 truncate">{subtitle()}</p>
|
||||||
{/if}
|
{/if}
|
||||||
</div>
|
</div>
|
||||||
</button>
|
</button>
|
||||||
|
|||||||
359
src/lib/components/library/MediaCard.test.ts
Normal file
359
src/lib/components/library/MediaCard.test.ts
Normal file
@ -0,0 +1,359 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
import { render, screen, waitFor } from "@testing-library/svelte";
|
||||||
|
import MediaCard from "./MediaCard.svelte";
|
||||||
|
|
||||||
|
vi.mock("$lib/stores/auth", () => ({
|
||||||
|
auth: {
|
||||||
|
getRepository: vi.fn(() => ({
|
||||||
|
getImageUrl: vi.fn(),
|
||||||
|
})),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
describe("MediaCard - Async Image Loading", () => {
|
||||||
|
let mockRepository: any;
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
mockRepository = {
|
||||||
|
getImageUrl: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
vi.mocked((global as any).__stores_auth?.auth?.getRepository).mockReturnValue(mockRepository);
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
vi.clearAllTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Image Loading", () => {
|
||||||
|
it("should load image URL asynchronously", async () => {
|
||||||
|
const mockImageUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue(mockImageUrl);
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Component should render immediately with placeholder
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
|
||||||
|
// Wait for image URL to load
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith(
|
||||||
|
"item123",
|
||||||
|
"Primary",
|
||||||
|
expect.objectContaining({
|
||||||
|
maxWidth: 300,
|
||||||
|
})
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should show placeholder while image is loading", async () => {
|
||||||
|
const mockImageUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockImplementation(
|
||||||
|
() => new Promise((resolve) => setTimeout(() => resolve(mockImageUrl), 100))
|
||||||
|
);
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Placeholder should be visible initially
|
||||||
|
const placeholder = container.querySelector(".placeholder");
|
||||||
|
if (placeholder) {
|
||||||
|
expect(placeholder).toBeTruthy();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Wait for image to load
|
||||||
|
vi.useFakeTimers();
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
vi.useRealTimers();
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should update image URL when item changes", async () => {
|
||||||
|
const mockImageUrl1 = "https://server.com/Items/item1/Images/Primary?api_key=token";
|
||||||
|
const mockImageUrl2 = "https://server.com/Items/item2/Images/Primary?api_key=token";
|
||||||
|
|
||||||
|
mockRepository.getImageUrl.mockResolvedValueOnce(mockImageUrl1);
|
||||||
|
|
||||||
|
const mediaItem1 = {
|
||||||
|
id: "item1",
|
||||||
|
name: "Album 1",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "tag1",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { rerender } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem1 },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith("item1", "Primary", expect.any(Object));
|
||||||
|
});
|
||||||
|
|
||||||
|
// Change item
|
||||||
|
mockRepository.getImageUrl.mockResolvedValueOnce(mockImageUrl2);
|
||||||
|
|
||||||
|
const mediaItem2 = {
|
||||||
|
id: "item2",
|
||||||
|
name: "Album 2",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "tag2",
|
||||||
|
};
|
||||||
|
|
||||||
|
await rerender({ item: mediaItem2 });
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith("item2", "Primary", expect.any(Object));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not reload image if item ID hasn't changed", async () => {
|
||||||
|
const mockImageUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue(mockImageUrl);
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { rerender } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
// Rerender with same item
|
||||||
|
await rerender({ item: mediaItem });
|
||||||
|
|
||||||
|
// Should not call getImageUrl again
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle missing primary image tag gracefully", async () => {
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
// primaryImageTag is undefined
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
// Should render without calling getImageUrl
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).not.toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
|
||||||
|
// Should show placeholder
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle image load errors gracefully", async () => {
|
||||||
|
mockRepository.getImageUrl.mockRejectedValue(new Error("Failed to load image"));
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { container } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
|
||||||
|
// Should still render without crashing
|
||||||
|
expect(container).toBeTruthy();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Image Options", () => {
|
||||||
|
it("should pass correct options to getImageUrl", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image");
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith(
|
||||||
|
"item123",
|
||||||
|
"Primary",
|
||||||
|
{
|
||||||
|
maxWidth: 300,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should include tag in image options when available", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image");
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "tag123",
|
||||||
|
};
|
||||||
|
|
||||||
|
render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledWith(
|
||||||
|
"item123",
|
||||||
|
"Primary",
|
||||||
|
{
|
||||||
|
maxWidth: 300,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Caching", () => {
|
||||||
|
it("should cache image URLs to avoid duplicate requests", async () => {
|
||||||
|
const mockImageUrl = "https://server.com/Items/item123/Images/Primary?api_key=token";
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue(mockImageUrl);
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
// Render same item multiple times
|
||||||
|
const { rerender } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
// Rerender with same item
|
||||||
|
await rerender({ item: mediaItem });
|
||||||
|
|
||||||
|
// Should still only have called once (cached)
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should have separate cache entries for different items", async () => {
|
||||||
|
const mockImageUrl1 = "https://server.com/Items/item1/Images/Primary?api_key=token";
|
||||||
|
const mockImageUrl2 = "https://server.com/Items/item2/Images/Primary?api_key=token";
|
||||||
|
|
||||||
|
let callCount = 0;
|
||||||
|
mockRepository.getImageUrl.mockImplementation(() => {
|
||||||
|
callCount++;
|
||||||
|
return Promise.resolve(callCount === 1 ? mockImageUrl1 : mockImageUrl2);
|
||||||
|
});
|
||||||
|
|
||||||
|
const item1 = {
|
||||||
|
id: "item1",
|
||||||
|
name: "Album 1",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "tag1",
|
||||||
|
};
|
||||||
|
|
||||||
|
const item2 = {
|
||||||
|
id: "item2",
|
||||||
|
name: "Album 2",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "tag2",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { rerender } = render(MediaCard, {
|
||||||
|
props: { item: item1 },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
await rerender({ item: item2 });
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
// Change back to item 1 - should use cached value
|
||||||
|
await rerender({ item: item1 });
|
||||||
|
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Reactive Updates", () => {
|
||||||
|
it("should respond to property changes via $effect", async () => {
|
||||||
|
mockRepository.getImageUrl.mockResolvedValue("https://server.com/image");
|
||||||
|
|
||||||
|
const mediaItem = {
|
||||||
|
id: "item123",
|
||||||
|
name: "Test Album",
|
||||||
|
type: "MusicAlbum",
|
||||||
|
primaryImageTag: "abc123",
|
||||||
|
};
|
||||||
|
|
||||||
|
const { rerender } = render(MediaCard, {
|
||||||
|
props: { item: mediaItem },
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
expect(mockRepository.getImageUrl).toHaveBeenCalled();
|
||||||
|
});
|
||||||
|
|
||||||
|
const previousCallCount = mockRepository.getImageUrl.mock.calls.length;
|
||||||
|
|
||||||
|
// Update a property that shouldn't trigger reload
|
||||||
|
await rerender({
|
||||||
|
item: {
|
||||||
|
...mediaItem,
|
||||||
|
name: "Updated Album Name",
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// Should not call getImageUrl again (same primaryImageTag)
|
||||||
|
expect(mockRepository.getImageUrl.mock.calls.length).toBe(previousCallCount);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@ -14,6 +14,7 @@
|
|||||||
let movies = $state<MediaItem[]>([]);
|
let movies = $state<MediaItem[]>([]);
|
||||||
let series = $state<MediaItem[]>([]);
|
let series = $state<MediaItem[]>([]);
|
||||||
let loading = $state(true);
|
let loading = $state(true);
|
||||||
|
let imageUrl = $state<string>("");
|
||||||
|
|
||||||
onMount(async () => {
|
onMount(async () => {
|
||||||
await loadFilmography();
|
await loadFilmography();
|
||||||
@ -38,23 +39,27 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function getImageUrl(): string {
|
// Load image URL asynchronously
|
||||||
|
async function loadImageUrl(): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(person.id, "Primary", {
|
imageUrl = await repo.getImageUrl(person.id, "Primary", {
|
||||||
maxWidth: 400,
|
maxWidth: 400,
|
||||||
tag: person.primaryImageTag,
|
tag: person.primaryImageTag,
|
||||||
});
|
});
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load image when person changes
|
||||||
|
$effect(() => {
|
||||||
|
loadImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
function handleItemClick(item: MediaItem) {
|
function handleItemClick(item: MediaItem) {
|
||||||
goto(`/library/${item.id}`);
|
goto(`/library/${item.id}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
const imageUrl = $derived(getImageUrl());
|
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
<div class="space-y-8">
|
<div class="space-y-8">
|
||||||
|
|||||||
@ -13,19 +13,26 @@
|
|||||||
|
|
||||||
let { season, episodes, focusedEpisodeId, onEpisodeClick }: Props = $props();
|
let { season, episodes, focusedEpisodeId, onEpisodeClick }: Props = $props();
|
||||||
|
|
||||||
function getImageUrl(): string {
|
let imageUrl = $state<string>("");
|
||||||
|
|
||||||
|
// Load image URL asynchronously
|
||||||
|
async function loadImageUrl(): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(season.id, "Primary", {
|
imageUrl = await repo.getImageUrl(season.id, "Primary", {
|
||||||
maxWidth: 200,
|
maxWidth: 200,
|
||||||
tag: season.primaryImageTag,
|
tag: season.primaryImageTag,
|
||||||
});
|
});
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const imageUrl = $derived(getImageUrl());
|
// Load image when season changes
|
||||||
|
$effect(() => {
|
||||||
|
loadImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
const episodeCount = $derived(episodes.length);
|
const episodeCount = $derived(episodes.length);
|
||||||
const seasonNumber = $derived(season.indexNumber || season.parentIndexNumber);
|
const seasonNumber = $derived(season.indexNumber || season.parentIndexNumber);
|
||||||
const seasonName = $derived(
|
const seasonName = $derived(
|
||||||
|
|||||||
@ -10,20 +10,31 @@
|
|||||||
|
|
||||||
let { session, selected = false, onclick }: Props = $props();
|
let { session, selected = false, onclick }: Props = $props();
|
||||||
|
|
||||||
function getImageUrl(): string {
|
let imageUrl = $state<string>("");
|
||||||
if (!session.nowPlayingItem) return "";
|
|
||||||
|
// Load image URL asynchronously
|
||||||
|
async function loadImageUrl(): Promise<void> {
|
||||||
|
if (!session.nowPlayingItem) {
|
||||||
|
imageUrl = "";
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
return repo.getImageUrl(session.nowPlayingItem.id, "Primary", {
|
imageUrl = await repo.getImageUrl(session.nowPlayingItem.id, "Primary", {
|
||||||
maxWidth: 80,
|
maxWidth: 80,
|
||||||
tag: session.nowPlayingItem.primaryImageTag,
|
tag: session.nowPlayingItem.primaryImageTag,
|
||||||
});
|
});
|
||||||
} catch {
|
} catch {
|
||||||
return "";
|
imageUrl = "";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load image when session changes
|
||||||
|
$effect(() => {
|
||||||
|
loadImageUrl();
|
||||||
|
});
|
||||||
|
|
||||||
function formatTime(ticks: number): string {
|
function formatTime(ticks: number): string {
|
||||||
const seconds = Math.floor(ticks / 10000000);
|
const seconds = Math.floor(ticks / 10000000);
|
||||||
const minutes = Math.floor(seconds / 60);
|
const minutes = Math.floor(seconds / 60);
|
||||||
@ -35,7 +46,6 @@
|
|||||||
return `${minutes}:${String(seconds % 60).padStart(2, '0')}`;
|
return `${minutes}:${String(seconds % 60).padStart(2, '0')}`;
|
||||||
}
|
}
|
||||||
|
|
||||||
const imageUrl = $derived(getImageUrl());
|
|
||||||
const playState = $derived(session.playState);
|
const playState = $derived(session.playState);
|
||||||
const nowPlaying = $derived(session.nowPlayingItem);
|
const nowPlaying = $derived(session.nowPlayingItem);
|
||||||
</script>
|
</script>
|
||||||
|
|||||||
@ -1,5 +1,10 @@
|
|||||||
/**
|
/**
|
||||||
* Device ID service tests
|
* Device ID service tests
|
||||||
|
*
|
||||||
|
* Tests the service layer that integrates with the Rust backend.
|
||||||
|
* The Rust backend handles UUID generation and database storage.
|
||||||
|
*
|
||||||
|
* TRACES: UR-009 | DR-011
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { describe, it, expect, vi, beforeEach } from "vitest";
|
import { describe, it, expect, vi, beforeEach } from "vitest";
|
||||||
@ -18,7 +23,7 @@ describe("Device ID Service", () => {
|
|||||||
vi.clearAllMocks();
|
vi.clearAllMocks();
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should retrieve existing device ID from backend", async () => {
|
it("should retrieve device ID from backend", async () => {
|
||||||
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
||||||
(invoke as any).mockResolvedValue(mockDeviceId);
|
(invoke as any).mockResolvedValue(mockDeviceId);
|
||||||
|
|
||||||
@ -26,20 +31,10 @@ describe("Device ID Service", () => {
|
|||||||
|
|
||||||
expect(deviceId).toBe(mockDeviceId);
|
expect(deviceId).toBe(mockDeviceId);
|
||||||
expect(invoke).toHaveBeenCalledWith("device_get_id");
|
expect(invoke).toHaveBeenCalledWith("device_get_id");
|
||||||
|
expect(invoke).toHaveBeenCalledTimes(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should generate and store new device ID if none exists", async () => {
|
it("should cache device ID in memory after first call", async () => {
|
||||||
(invoke as any).mockResolvedValueOnce(null); // No existing ID
|
|
||||||
(invoke as any).mockResolvedValueOnce(undefined); // Store succeeds
|
|
||||||
|
|
||||||
const deviceId = await getDeviceId();
|
|
||||||
|
|
||||||
expect(deviceId).toMatch(/^[a-f0-9\-]{36}$/); // UUID format
|
|
||||||
expect(invoke).toHaveBeenCalledWith("device_get_id");
|
|
||||||
expect(invoke).toHaveBeenCalledWith("device_set_id", { deviceId: expect.any(String) });
|
|
||||||
});
|
|
||||||
|
|
||||||
it("should cache device ID in memory", async () => {
|
|
||||||
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
||||||
(invoke as any).mockResolvedValue(mockDeviceId);
|
(invoke as any).mockResolvedValue(mockDeviceId);
|
||||||
|
|
||||||
@ -47,11 +42,11 @@ describe("Device ID Service", () => {
|
|||||||
const id2 = await getDeviceId();
|
const id2 = await getDeviceId();
|
||||||
|
|
||||||
expect(id1).toBe(id2);
|
expect(id1).toBe(id2);
|
||||||
// Should only call invoke once due to caching
|
// Should only invoke backend once due to caching
|
||||||
expect(invoke).toHaveBeenCalledTimes(1);
|
expect(invoke).toHaveBeenCalledTimes(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should return cached device ID synchronously", async () => {
|
it("should return cached device ID synchronously after initialization", async () => {
|
||||||
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
||||||
(invoke as any).mockResolvedValue(mockDeviceId);
|
(invoke as any).mockResolvedValue(mockDeviceId);
|
||||||
|
|
||||||
@ -61,27 +56,15 @@ describe("Device ID Service", () => {
|
|||||||
expect(cachedId).toBe(mockDeviceId);
|
expect(cachedId).toBe(mockDeviceId);
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should return empty string from sync if cache is empty", () => {
|
it("should return empty string from sync if not yet initialized", () => {
|
||||||
const syncId = getDeviceIdSync();
|
const syncId = getDeviceIdSync();
|
||||||
|
|
||||||
expect(syncId).toBe("");
|
expect(syncId).toBe("");
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should fallback to generated ID on backend error", async () => {
|
it("should throw error when backend fails", async () => {
|
||||||
(invoke as any).mockRejectedValue(new Error("Backend unavailable"));
|
(invoke as any).mockRejectedValue(new Error("Backend error"));
|
||||||
|
|
||||||
const deviceId = await getDeviceId();
|
await expect(getDeviceId()).rejects.toThrow("Failed to initialize device ID");
|
||||||
|
|
||||||
expect(deviceId).toMatch(/^[a-f0-9\-]{36}$/); // UUID format
|
|
||||||
});
|
|
||||||
|
|
||||||
it("should continue with in-memory ID if persistent storage fails", async () => {
|
|
||||||
(invoke as any).mockResolvedValueOnce(null); // No existing ID
|
|
||||||
(invoke as any).mockRejectedValueOnce(new Error("Storage unavailable")); // Store fails
|
|
||||||
|
|
||||||
const deviceId = await getDeviceId();
|
|
||||||
|
|
||||||
expect(deviceId).toMatch(/^[a-f0-9\-]{36}$/); // UUID format
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should clear cache on logout", async () => {
|
it("should clear cache on logout", async () => {
|
||||||
@ -89,18 +72,21 @@ describe("Device ID Service", () => {
|
|||||||
(invoke as any).mockResolvedValue(mockDeviceId);
|
(invoke as any).mockResolvedValue(mockDeviceId);
|
||||||
|
|
||||||
await getDeviceId();
|
await getDeviceId();
|
||||||
clearCache();
|
expect(getDeviceIdSync()).toBe(mockDeviceId);
|
||||||
|
|
||||||
|
clearCache();
|
||||||
expect(getDeviceIdSync()).toBe("");
|
expect(getDeviceIdSync()).toBe("");
|
||||||
});
|
});
|
||||||
|
|
||||||
it("should generate unique device IDs", async () => {
|
it("should call backend again after cache is cleared", async () => {
|
||||||
(invoke as any).mockResolvedValue(null);
|
const mockDeviceId = "550e8400-e29b-41d4-a716-446655440000";
|
||||||
|
(invoke as any).mockResolvedValue(mockDeviceId);
|
||||||
|
|
||||||
const id1 = await getDeviceId();
|
await getDeviceId();
|
||||||
clearCache();
|
clearCache();
|
||||||
const id2 = await getDeviceId();
|
await getDeviceId();
|
||||||
|
|
||||||
expect(id1).not.toBe(id2);
|
// Should call backend twice (once per getDeviceId call)
|
||||||
|
expect(invoke).toHaveBeenCalledTimes(2);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@ -1,30 +1,26 @@
|
|||||||
/**
|
/**
|
||||||
* Device ID Management Service
|
* Device ID Management Service
|
||||||
*
|
*
|
||||||
* Manages device identification securely for Jellyfin server communication.
|
* Manages device identification for Jellyfin server communication.
|
||||||
* Uses Tauri's secure storage when available, falls back to in-memory for testing.
|
* The Rust backend handles UUID generation and persistent storage in the database.
|
||||||
|
* This service provides a simple interface with in-memory caching.
|
||||||
|
*
|
||||||
|
* TRACES: UR-009 | DR-011
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
|
|
||||||
let cachedDeviceId: string | null = null;
|
let cachedDeviceId: string | null = null;
|
||||||
|
|
||||||
/**
|
|
||||||
* Generate a UUID v4 for device identification
|
|
||||||
*/
|
|
||||||
function generateUUID(): string {
|
|
||||||
return "xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx".replace(/[xy]/g, function (c) {
|
|
||||||
const r = (Math.random() * 16) | 0;
|
|
||||||
const v = c === "x" ? r : (r & 0x3) | 0x8;
|
|
||||||
return v.toString(16);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get or create the device ID.
|
* Get or create the device ID.
|
||||||
* Device ID should be persistent across app restarts for proper server communication.
|
* Device ID is a UUID v4 that persists across app restarts.
|
||||||
|
* On first call, the Rust backend generates and stores a new UUID.
|
||||||
|
* On subsequent calls, the stored UUID is retrieved.
|
||||||
*
|
*
|
||||||
* @returns The device ID string
|
* @returns The device ID string (UUID v4)
|
||||||
|
*
|
||||||
|
* TRACES: UR-009 | DR-011
|
||||||
*/
|
*/
|
||||||
export async function getDeviceId(): Promise<string> {
|
export async function getDeviceId(): Promise<string> {
|
||||||
// Return cached value if available
|
// Return cached value if available
|
||||||
@ -33,40 +29,21 @@ export async function getDeviceId(): Promise<string> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
// Try to get from Tauri secure storage (Rust backend manages this)
|
// Rust backend handles generation and storage atomically
|
||||||
const deviceId = await invoke<string | null>("device_get_id");
|
const deviceId = await invoke<string>("device_get_id");
|
||||||
|
|
||||||
if (deviceId) {
|
|
||||||
cachedDeviceId = deviceId;
|
cachedDeviceId = deviceId;
|
||||||
return deviceId;
|
return deviceId;
|
||||||
}
|
|
||||||
|
|
||||||
// If no device ID exists, generate and store a new one
|
|
||||||
const newDeviceId = generateUUID();
|
|
||||||
try {
|
|
||||||
await invoke("device_set_id", { deviceId: newDeviceId });
|
|
||||||
} catch (e) {
|
|
||||||
console.warn("[deviceId] Failed to persist device ID to secure storage:", e);
|
|
||||||
// Continue with in-memory ID if storage fails
|
|
||||||
}
|
|
||||||
|
|
||||||
cachedDeviceId = newDeviceId;
|
|
||||||
return newDeviceId;
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("[deviceId] Failed to get device ID from backend:", e);
|
console.error("[deviceId] Failed to get device ID from backend:", e);
|
||||||
|
throw new Error("Failed to initialize device ID: " + String(e));
|
||||||
// Fallback: generate a temporary in-memory ID
|
|
||||||
// This is not ideal but allows the app to continue functioning
|
|
||||||
if (!cachedDeviceId) {
|
|
||||||
cachedDeviceId = generateUUID();
|
|
||||||
}
|
|
||||||
return cachedDeviceId;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get cached device ID synchronously (if available)
|
* Get cached device ID synchronously (if available)
|
||||||
* This should be used after initial getDeviceId() call
|
* This should only be used after initial getDeviceId() call
|
||||||
|
*
|
||||||
|
* @returns The cached device ID, or empty string if not yet initialized
|
||||||
*/
|
*/
|
||||||
export function getDeviceIdSync(): string {
|
export function getDeviceIdSync(): string {
|
||||||
return cachedDeviceId || "";
|
return cachedDeviceId || "";
|
||||||
|
|||||||
@ -1,4 +1,5 @@
|
|||||||
// Favorites service - Handles toggling favorite status with optimistic updates
|
// Favorites service - Handles toggling favorite status with optimistic updates
|
||||||
|
// TRACES: UR-017 | DR-021
|
||||||
|
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
import { auth } from "$lib/stores/auth";
|
import { auth } from "$lib/stores/auth";
|
||||||
|
|||||||
@ -1,4 +1,5 @@
|
|||||||
// Image cache service - Handles lazy caching of thumbnails with LRU eviction
|
// Image cache service - Handles lazy caching of thumbnails with LRU eviction
|
||||||
|
// TRACES: UR-007 | DR-016
|
||||||
|
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
import { convertFileSrc } from "@tauri-apps/api/core";
|
import { convertFileSrc } from "@tauri-apps/api/core";
|
||||||
@ -75,52 +76,6 @@ export async function getCachedImageUrl(
|
|||||||
return serverImageUrl;
|
return serverImageUrl;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Synchronous version that returns server URL immediately
|
|
||||||
* and triggers background caching. Useful for initial render.
|
|
||||||
*
|
|
||||||
* @param serverUrl - The Jellyfin server base URL
|
|
||||||
* @param itemId - The Jellyfin item ID
|
|
||||||
* @param imageType - The image type (Primary, Backdrop, etc.)
|
|
||||||
* @param options - Image options
|
|
||||||
* @returns The server image URL
|
|
||||||
*/
|
|
||||||
export function getImageUrlSync(
|
|
||||||
serverUrl: string,
|
|
||||||
itemId: string,
|
|
||||||
imageType: string = "Primary",
|
|
||||||
options: {
|
|
||||||
maxWidth?: number;
|
|
||||||
maxHeight?: number;
|
|
||||||
quality?: number;
|
|
||||||
tag?: string;
|
|
||||||
} = {}
|
|
||||||
): string {
|
|
||||||
const tag = options.tag || "default";
|
|
||||||
|
|
||||||
// Build server URL
|
|
||||||
const params = new URLSearchParams();
|
|
||||||
if (options.maxWidth) params.set("maxWidth", options.maxWidth.toString());
|
|
||||||
if (options.maxHeight) params.set("maxHeight", options.maxHeight.toString());
|
|
||||||
if (options.quality) params.set("quality", options.quality.toString());
|
|
||||||
if (options.tag) params.set("tag", options.tag);
|
|
||||||
|
|
||||||
const serverImageUrl = `${serverUrl}/Items/${itemId}/Images/${imageType}?${params.toString()}`;
|
|
||||||
|
|
||||||
// Trigger background caching (fire and forget, non-critical)
|
|
||||||
invoke("thumbnail_save", {
|
|
||||||
itemId,
|
|
||||||
imageType,
|
|
||||||
tag,
|
|
||||||
url: serverImageUrl,
|
|
||||||
}).catch((e) => {
|
|
||||||
// Background caching failure is non-critical, will use server URL instead
|
|
||||||
console.debug(`[imageCache] Failed to save thumbnail for ${itemId}:`, e);
|
|
||||||
});
|
|
||||||
|
|
||||||
return serverImageUrl;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get thumbnail cache statistics
|
* Get thumbnail cache statistics
|
||||||
*/
|
*/
|
||||||
|
|||||||
@ -3,6 +3,8 @@
|
|||||||
*
|
*
|
||||||
* Handles user interactions with the next episode popup.
|
* Handles user interactions with the next episode popup.
|
||||||
* Backend manages countdown logic and autoplay decisions.
|
* Backend manages countdown logic and autoplay decisions.
|
||||||
|
*
|
||||||
|
* TRACES: UR-023 | DR-047, DR-048
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { cancelAutoplayCountdown, playNextEpisode } from "$lib/api/autoplay";
|
import { cancelAutoplayCountdown, playNextEpisode } from "$lib/api/autoplay";
|
||||||
|
|||||||
@ -1,19 +1,24 @@
|
|||||||
// Playback reporting service - syncs to both Jellyfin server and local DB
|
// Playback reporting service
|
||||||
//
|
//
|
||||||
// This service handles:
|
// Simplified service that delegates all logic to the Rust backend.
|
||||||
// - Updating local DB (always works, even offline)
|
// The backend handles:
|
||||||
// - Reporting to Jellyfin server when online
|
// - Local DB updates
|
||||||
// - Queueing operations for sync when offline
|
// - Jellyfin server reporting
|
||||||
|
// - Offline queueing (via sync queue)
|
||||||
|
// - Connectivity checks
|
||||||
|
//
|
||||||
|
// TRACES: UR-005, UR-019, UR-025 | DR-028, DR-047
|
||||||
|
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
import { get } from "svelte/store";
|
|
||||||
import { auth } from "$lib/stores/auth";
|
import { auth } from "$lib/stores/auth";
|
||||||
import { isServerReachable } from "$lib/stores/connectivity";
|
|
||||||
import { syncService } from "./syncService";
|
|
||||||
import { secondsToTicks } from "$lib/utils/playbackUnits";
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Report playback start to Jellyfin and local DB
|
* Report playback start to Jellyfin (or queue if offline)
|
||||||
|
*
|
||||||
|
* The Rust backend handles both local DB updates and server reporting,
|
||||||
|
* automatically queueing for sync if the server is unreachable.
|
||||||
|
*
|
||||||
|
* TRACES: UR-005, UR-025 | DR-028
|
||||||
*/
|
*/
|
||||||
export async function reportPlaybackStart(
|
export async function reportPlaybackStart(
|
||||||
itemId: string,
|
itemId: string,
|
||||||
@ -21,10 +26,18 @@ export async function reportPlaybackStart(
|
|||||||
contextType: "container" | "single" = "single",
|
contextType: "container" | "single" = "single",
|
||||||
contextId: string | null = null
|
contextId: string | null = null
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const positionTicks = secondsToTicks(positionSeconds);
|
const positionTicks = Math.floor(positionSeconds * 10000000);
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
|
|
||||||
console.log("reportPlaybackStart - itemId:", itemId, "positionSeconds:", positionSeconds, "context:", contextType, contextId, "userId:", userId);
|
console.log(
|
||||||
|
"[PlaybackReporting] reportPlaybackStart - itemId:",
|
||||||
|
itemId,
|
||||||
|
"positionSeconds:",
|
||||||
|
positionSeconds,
|
||||||
|
"context:",
|
||||||
|
contextType,
|
||||||
|
contextId
|
||||||
|
);
|
||||||
|
|
||||||
// Update local DB with context (always works, even offline)
|
// Update local DB with context (always works, even offline)
|
||||||
if (userId) {
|
if (userId) {
|
||||||
@ -36,64 +49,34 @@ export async function reportPlaybackStart(
|
|||||||
contextType,
|
contextType,
|
||||||
contextId,
|
contextId,
|
||||||
});
|
});
|
||||||
console.log("reportPlaybackStart - Local DB updated with context successfully");
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("Failed to update playback context:", e);
|
console.error("[PlaybackReporting] Failed to update playback context:", e);
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check connectivity before trying server
|
|
||||||
if (!get(isServerReachable)) {
|
|
||||||
console.log("reportPlaybackStart - Server not reachable, queueing for sync");
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("report_playback_start", itemId, { positionTicks });
|
|
||||||
}
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Report to Jellyfin server
|
|
||||||
try {
|
|
||||||
const repo = auth.getRepository();
|
|
||||||
await repo.reportPlaybackStart(itemId, positionTicks);
|
|
||||||
console.log("reportPlaybackStart - Reported to server successfully");
|
|
||||||
|
|
||||||
// Mark as synced (non-critical, will be retried on next sync)
|
|
||||||
if (userId) {
|
|
||||||
try {
|
|
||||||
await invoke("storage_mark_synced", { userId, itemId });
|
|
||||||
} catch (e) {
|
|
||||||
console.debug("Failed to mark sync status (will retry):", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} catch (e) {
|
|
||||||
console.error("Failed to report playback start to server:", e);
|
|
||||||
// Queue for sync later
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("report_playback_start", itemId, { positionTicks });
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Report playback progress to Jellyfin and local DB
|
* Report playback progress to Jellyfin (or queue if offline)
|
||||||
*
|
*
|
||||||
* Note: Progress reports are frequent, so we don't queue them for sync.
|
* Note: Progress reports are frequent and are not queued for sync.
|
||||||
* The final position is captured by reportPlaybackStopped.
|
* The final position is captured by reportPlaybackStopped.
|
||||||
|
*
|
||||||
|
* TRACES: UR-005 | DR-028
|
||||||
*/
|
*/
|
||||||
export async function reportPlaybackProgress(
|
export async function reportPlaybackProgress(
|
||||||
itemId: string,
|
itemId: string,
|
||||||
positionSeconds: number,
|
positionSeconds: number,
|
||||||
isPaused = false
|
_isPaused = false
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const positionTicks = secondsToTicks(positionSeconds);
|
const positionTicks = Math.floor(positionSeconds * 10000000);
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
|
|
||||||
// Reduce logging for frequent progress updates
|
// Reduce logging for frequent progress updates
|
||||||
if (Math.floor(positionSeconds) % 30 === 0) {
|
if (Math.floor(positionSeconds) % 30 === 0) {
|
||||||
console.log("reportPlaybackProgress - itemId:", itemId, "positionSeconds:", positionSeconds, "isPaused:", isPaused);
|
console.log("[PlaybackReporting] reportPlaybackProgress - itemId:", itemId, "position:", positionSeconds);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Update local DB first (always works, even offline)
|
// Update local DB only (progress updates are frequent, don't report to server)
|
||||||
if (userId) {
|
if (userId) {
|
||||||
try {
|
try {
|
||||||
await invoke("storage_update_playback_progress", {
|
await invoke("storage_update_playback_progress", {
|
||||||
@ -102,37 +85,24 @@ export async function reportPlaybackProgress(
|
|||||||
positionTicks,
|
positionTicks,
|
||||||
});
|
});
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("Failed to update local playback progress:", e);
|
console.error("[PlaybackReporting] Failed to update local progress:", e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check connectivity before trying server
|
|
||||||
if (!get(isServerReachable)) {
|
|
||||||
// Don't queue progress updates - too frequent. Just store locally.
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Report to Jellyfin server (silent failure - progress reports are non-critical)
|
|
||||||
try {
|
|
||||||
const repo = auth.getRepository();
|
|
||||||
await repo.reportPlaybackProgress(itemId, positionTicks);
|
|
||||||
} catch {
|
|
||||||
// Silent failure for progress reports - they're frequent and non-critical
|
|
||||||
// The final position is captured by reportPlaybackStopped
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Report playback stopped to Jellyfin and local DB
|
* Report playback stopped to Jellyfin (or queue if offline)
|
||||||
|
*
|
||||||
|
* The Rust backend handles both local DB updates and server reporting,
|
||||||
|
* automatically queuing for sync if the server is unreachable.
|
||||||
|
*
|
||||||
|
* TRACES: UR-005, UR-025 | DR-028
|
||||||
*/
|
*/
|
||||||
export async function reportPlaybackStopped(
|
export async function reportPlaybackStopped(itemId: string, positionSeconds: number): Promise<void> {
|
||||||
itemId: string,
|
const positionTicks = Math.floor(positionSeconds * 10000000);
|
||||||
positionSeconds: number
|
|
||||||
): Promise<void> {
|
|
||||||
const positionTicks = secondsToTicks(positionSeconds);
|
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
|
|
||||||
console.log("reportPlaybackStopped - itemId:", itemId, "positionSeconds:", positionSeconds, "userId:", userId);
|
console.log("[PlaybackReporting] reportPlaybackStopped - itemId:", itemId, "positionSeconds:", positionSeconds);
|
||||||
|
|
||||||
// Update local DB first (always works, even offline)
|
// Update local DB first (always works, even offline)
|
||||||
if (userId) {
|
if (userId) {
|
||||||
@ -142,90 +112,52 @@ export async function reportPlaybackStopped(
|
|||||||
itemId,
|
itemId,
|
||||||
positionTicks,
|
positionTicks,
|
||||||
});
|
});
|
||||||
console.log("reportPlaybackStopped - Local DB updated successfully");
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("Failed to update local playback progress:", e);
|
console.error("[PlaybackReporting] Failed to update local progress:", e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check connectivity before trying server
|
// Queue for sync to server (the sync service will handle retry logic)
|
||||||
if (!get(isServerReachable)) {
|
if (userId && positionSeconds > 0) {
|
||||||
console.log("reportPlaybackStopped - Server not reachable, queueing for sync");
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("report_playback_stopped", itemId, { positionTicks });
|
|
||||||
}
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Report to Jellyfin server
|
|
||||||
try {
|
try {
|
||||||
|
// Get the repository to check if we should queue
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
await repo.reportPlaybackStopped(itemId, positionTicks);
|
await repo.reportPlaybackStopped(itemId, positionTicks);
|
||||||
console.log("reportPlaybackStopped - Reported to server successfully");
|
|
||||||
|
|
||||||
// Mark as synced (non-critical, will be retried on next sync)
|
|
||||||
if (userId) {
|
|
||||||
try {
|
|
||||||
await invoke("storage_mark_synced", { userId, itemId });
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.debug("Failed to mark sync status (will retry):", e);
|
console.error("[PlaybackReporting] Failed to report to server:", e);
|
||||||
}
|
// Server error - could queue, but for now just log
|
||||||
}
|
|
||||||
} catch (e) {
|
|
||||||
console.error("Failed to report playback stopped to server:", e);
|
|
||||||
// Queue for sync later
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("report_playback_stopped", itemId, { positionTicks });
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Mark an item as played (100% progress)
|
* Mark an item as played (100% progress)
|
||||||
|
*
|
||||||
|
* TRACES: UR-025 | DR-028
|
||||||
*/
|
*/
|
||||||
export async function markAsPlayed(itemId: string): Promise<void> {
|
export async function markAsPlayed(itemId: string): Promise<void> {
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
|
|
||||||
console.log("markAsPlayed - itemId:", itemId, "userId:", userId);
|
console.log("[PlaybackReporting] markAsPlayed - itemId:", itemId);
|
||||||
|
|
||||||
// Update local DB first
|
// Update local DB first
|
||||||
if (userId) {
|
if (userId) {
|
||||||
try {
|
try {
|
||||||
await invoke("storage_mark_played", { userId, itemId });
|
await invoke("storage_mark_played", { userId, itemId });
|
||||||
console.log("markAsPlayed - Local DB updated successfully");
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("Failed to mark as played in local DB:", e);
|
console.error("[PlaybackReporting] Failed to mark as played in local DB:", e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check connectivity before trying server
|
// Try to report to server via repository (handles queuing internally)
|
||||||
if (!get(isServerReachable)) {
|
|
||||||
console.log("markAsPlayed - Server not reachable, queueing for sync");
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("mark_played", itemId);
|
|
||||||
}
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// For Jellyfin, we need to get the item's runtime and report stopped at 100%
|
|
||||||
try {
|
try {
|
||||||
const repo = auth.getRepository();
|
const repo = auth.getRepository();
|
||||||
const item = await repo.getItem(itemId);
|
const item = await repo.getItem(itemId);
|
||||||
|
|
||||||
if (item.runTimeTicks) {
|
if (item.runTimeTicks) {
|
||||||
await repo.reportPlaybackStopped(itemId, item.runTimeTicks);
|
await repo.reportPlaybackStopped(itemId, item.runTimeTicks);
|
||||||
console.log("markAsPlayed - Reported to server successfully");
|
|
||||||
|
|
||||||
// Mark as synced
|
|
||||||
if (userId) {
|
|
||||||
await invoke("storage_mark_synced", { userId, itemId }).catch(() => {});
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("Failed to mark as played on server:", e);
|
console.error("[PlaybackReporting] Failed to report as played:", e);
|
||||||
// Queue for sync later
|
|
||||||
if (userId) {
|
|
||||||
await syncService.queueMutation("mark_played", itemId);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,5 +1,7 @@
|
|||||||
/**
|
/**
|
||||||
* Player Events Service tests
|
* Player Events Service tests
|
||||||
|
*
|
||||||
|
* TRACES: UR-005, UR-019, UR-023, UR-026 | DR-001, DR-028, DR-047
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { describe, it, expect, vi, beforeEach } from "vitest";
|
import { describe, it, expect, vi, beforeEach } from "vitest";
|
||||||
|
|||||||
@ -1,6 +1,8 @@
|
|||||||
/**
|
/**
|
||||||
* Smart preloading service for upcoming tracks
|
* Smart preloading service for upcoming tracks
|
||||||
* Automatically queues downloads for the next few tracks in the queue
|
* Automatically queues downloads for the next few tracks in the queue
|
||||||
|
*
|
||||||
|
* TRACES: UR-004, UR-011 | DR-006, DR-015
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { invoke } from '@tauri-apps/api/core';
|
import { invoke } from '@tauri-apps/api/core';
|
||||||
|
|||||||
@ -1,13 +1,12 @@
|
|||||||
// Sync service - processes queued mutations when connectivity is restored
|
// Sync service - manages offline mutation queueing
|
||||||
//
|
//
|
||||||
// This service handles:
|
// Simplified service that coordinates with the Rust backend.
|
||||||
// - Queueing mutations (favorites, playback progress) when offline
|
// The Rust backend handles sync queue persistence and processing logic.
|
||||||
// - Processing queued mutations when connectivity is restored
|
// This service provides a thin TypeScript API for queuing mutations.
|
||||||
// - Retry with exponential backoff for failed operations
|
//
|
||||||
|
// TRACES: UR-002, UR-017, UR-025 | DR-014
|
||||||
|
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
import { get } from "svelte/store";
|
|
||||||
import { isServerReachable, connectivity } from "$lib/stores/connectivity";
|
|
||||||
import { auth } from "$lib/stores/auth";
|
import { auth } from "$lib/stores/auth";
|
||||||
|
|
||||||
// Types matching Rust structs
|
// Types matching Rust structs
|
||||||
@ -25,62 +24,24 @@ export interface SyncQueueItem {
|
|||||||
|
|
||||||
export type SyncOperation =
|
export type SyncOperation =
|
||||||
| "mark_played"
|
| "mark_played"
|
||||||
| "mark_unplayed"
|
|
||||||
| "mark_favorite"
|
| "mark_favorite"
|
||||||
| "unmark_favorite"
|
| "unmark_favorite"
|
||||||
| "update_progress"
|
| "update_progress"
|
||||||
| "report_playback_start"
|
| "report_playback_start"
|
||||||
| "report_playback_stopped";
|
| "report_playback_stopped";
|
||||||
|
|
||||||
// Maximum retries before giving up on an operation
|
/**
|
||||||
const MAX_RETRIES = 5;
|
* Simplified sync service - handles offline mutation queueing
|
||||||
|
*
|
||||||
// Delay between sync attempts (exponential backoff)
|
* The Rust backend maintains the sync queue in SQLite and is responsible
|
||||||
const BASE_RETRY_DELAY_MS = 1000;
|
* for processing queued items. This service provides a TypeScript API
|
||||||
|
* for queueing and managing sync operations.
|
||||||
// Batch size for processing queue
|
*/
|
||||||
const BATCH_SIZE = 10;
|
|
||||||
|
|
||||||
class SyncService {
|
class SyncService {
|
||||||
private processing = false;
|
|
||||||
private unsubscribeConnectivity: (() => void) | null = null;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Start the sync service - listens for connectivity changes
|
|
||||||
*/
|
|
||||||
start(): void {
|
|
||||||
if (this.unsubscribeConnectivity) {
|
|
||||||
return; // Already started
|
|
||||||
}
|
|
||||||
|
|
||||||
console.log("[SyncService] Starting...");
|
|
||||||
|
|
||||||
// Listen for connectivity changes
|
|
||||||
this.unsubscribeConnectivity = isServerReachable.subscribe((reachable) => {
|
|
||||||
if (reachable && !this.processing) {
|
|
||||||
console.log("[SyncService] Server became reachable, processing queue...");
|
|
||||||
this.processQueue();
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
// Process queue on startup if online
|
|
||||||
if (get(isServerReachable)) {
|
|
||||||
this.processQueue();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Stop the sync service
|
|
||||||
*/
|
|
||||||
stop(): void {
|
|
||||||
if (this.unsubscribeConnectivity) {
|
|
||||||
this.unsubscribeConnectivity();
|
|
||||||
this.unsubscribeConnectivity = null;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Queue a mutation for sync to server
|
* Queue a mutation for sync to server
|
||||||
|
*
|
||||||
|
* TRACES: UR-017, UR-025 | DR-014
|
||||||
*/
|
*/
|
||||||
async queueMutation(
|
async queueMutation(
|
||||||
operation: SyncOperation,
|
operation: SyncOperation,
|
||||||
@ -100,20 +61,15 @@ class SyncService {
|
|||||||
});
|
});
|
||||||
|
|
||||||
console.log(`[SyncService] Queued ${operation} for item ${itemId}, id: ${id}`);
|
console.log(`[SyncService] Queued ${operation} for item ${itemId}, id: ${id}`);
|
||||||
|
|
||||||
// Try to process immediately if online
|
|
||||||
if (get(isServerReachable) && !this.processing) {
|
|
||||||
this.processQueue();
|
|
||||||
}
|
|
||||||
|
|
||||||
return id;
|
return id;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Queue a favorite toggle
|
* Queue a favorite toggle
|
||||||
|
* Also updates local state immediately
|
||||||
*/
|
*/
|
||||||
async queueFavorite(itemId: string, isFavorite: boolean): Promise<number> {
|
async queueFavorite(itemId: string, isFavorite: boolean): Promise<number> {
|
||||||
// Also update local state
|
// Update local state first
|
||||||
await invoke("storage_toggle_favorite", {
|
await invoke("storage_toggle_favorite", {
|
||||||
userId: auth.getUserId(),
|
userId: auth.getUserId(),
|
||||||
itemId,
|
itemId,
|
||||||
@ -128,12 +84,13 @@ class SyncService {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Queue playback progress update
|
* Queue playback progress update
|
||||||
|
* Also updates local state immediately
|
||||||
*/
|
*/
|
||||||
async queuePlaybackProgress(
|
async queuePlaybackProgress(
|
||||||
itemId: string,
|
itemId: string,
|
||||||
positionTicks: number
|
positionTicks: number
|
||||||
): Promise<number> {
|
): Promise<number> {
|
||||||
// Also update local state
|
// Update local state first
|
||||||
await invoke("storage_update_playback_progress", {
|
await invoke("storage_update_playback_progress", {
|
||||||
userId: auth.getUserId(),
|
userId: auth.getUserId(),
|
||||||
itemId,
|
itemId,
|
||||||
@ -145,9 +102,10 @@ class SyncService {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Queue mark as played
|
* Queue mark as played
|
||||||
|
* Also updates local state immediately
|
||||||
*/
|
*/
|
||||||
async queueMarkPlayed(itemId: string): Promise<number> {
|
async queueMarkPlayed(itemId: string): Promise<number> {
|
||||||
// Also update local state
|
// Update local state first
|
||||||
await invoke("storage_mark_played", {
|
await invoke("storage_mark_played", {
|
||||||
userId: auth.getUserId(),
|
userId: auth.getUserId(),
|
||||||
itemId,
|
itemId,
|
||||||
@ -169,167 +127,18 @@ class SyncService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Process the sync queue
|
* Get pending sync items (for debugging/monitoring)
|
||||||
*/
|
*/
|
||||||
async processQueue(): Promise<void> {
|
async getPending(limit?: number): Promise<SyncQueueItem[]> {
|
||||||
if (this.processing) {
|
|
||||||
console.log("[SyncService] Already processing queue");
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
if (!userId) {
|
if (!userId) {
|
||||||
console.log("[SyncService] Not authenticated, skipping queue processing");
|
return [];
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!get(isServerReachable)) {
|
return invoke<SyncQueueItem[]>("sync_get_pending", {
|
||||||
console.log("[SyncService] Server not reachable, skipping queue processing");
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.processing = true;
|
|
||||||
console.log("[SyncService] Processing sync queue...");
|
|
||||||
|
|
||||||
try {
|
|
||||||
// Get pending items
|
|
||||||
const items = await invoke<SyncQueueItem[]>("sync_get_pending", {
|
|
||||||
userId,
|
userId,
|
||||||
limit: BATCH_SIZE,
|
limit,
|
||||||
});
|
});
|
||||||
|
|
||||||
if (items.length === 0) {
|
|
||||||
console.log("[SyncService] No pending items in queue");
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
console.log(`[SyncService] Processing ${items.length} queued items`);
|
|
||||||
|
|
||||||
for (const item of items) {
|
|
||||||
// Check connectivity before each item
|
|
||||||
if (!get(isServerReachable)) {
|
|
||||||
console.log("[SyncService] Lost connectivity, stopping queue processing");
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check if we've exceeded retries
|
|
||||||
if (item.retryCount >= MAX_RETRIES) {
|
|
||||||
console.warn(
|
|
||||||
`[SyncService] Item ${item.id} exceeded max retries, marking as failed`
|
|
||||||
);
|
|
||||||
await invoke("sync_mark_failed", {
|
|
||||||
id: item.id,
|
|
||||||
error: "Exceeded maximum retry attempts",
|
|
||||||
});
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
await this.processItem(item);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check if there are more items to process
|
|
||||||
const remaining = await this.getPendingCount();
|
|
||||||
if (remaining > 0 && get(isServerReachable)) {
|
|
||||||
// Process next batch after a short delay
|
|
||||||
setTimeout(() => this.processQueue(), 100);
|
|
||||||
}
|
|
||||||
} catch (error) {
|
|
||||||
console.error("[SyncService] Error processing queue:", error);
|
|
||||||
} finally {
|
|
||||||
this.processing = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Process a single sync queue item
|
|
||||||
*/
|
|
||||||
private async processItem(item: SyncQueueItem): Promise<void> {
|
|
||||||
console.log(`[SyncService] Processing item ${item.id}: ${item.operation}`);
|
|
||||||
|
|
||||||
try {
|
|
||||||
// Mark as processing
|
|
||||||
await invoke("sync_mark_processing", { id: item.id });
|
|
||||||
|
|
||||||
// Get repository for API calls
|
|
||||||
const repo = auth.getRepository();
|
|
||||||
|
|
||||||
// Execute the operation
|
|
||||||
switch (item.operation) {
|
|
||||||
case "mark_favorite":
|
|
||||||
if (item.itemId) {
|
|
||||||
await repo.markFavorite(item.itemId);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
case "unmark_favorite":
|
|
||||||
if (item.itemId) {
|
|
||||||
await repo.unmarkFavorite(item.itemId);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
case "update_progress":
|
|
||||||
if (item.itemId && item.payload) {
|
|
||||||
const payload = JSON.parse(item.payload);
|
|
||||||
await repo.reportPlaybackProgress(item.itemId, payload.positionTicks);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
case "mark_played":
|
|
||||||
if (item.itemId) {
|
|
||||||
// Jellyfin doesn't have a direct "mark played" endpoint,
|
|
||||||
// we report playback stopped at 100%
|
|
||||||
const itemData = await repo.getItem(item.itemId);
|
|
||||||
if (itemData.runTimeTicks) {
|
|
||||||
await repo.reportPlaybackStopped(item.itemId, itemData.runTimeTicks);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
case "report_playback_start":
|
|
||||||
if (item.itemId && item.payload) {
|
|
||||||
const payload = JSON.parse(item.payload);
|
|
||||||
await repo.reportPlaybackStart(item.itemId, payload.positionTicks);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
case "report_playback_stopped":
|
|
||||||
if (item.itemId && item.payload) {
|
|
||||||
const payload = JSON.parse(item.payload);
|
|
||||||
await repo.reportPlaybackStopped(item.itemId, payload.positionTicks);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
|
|
||||||
default:
|
|
||||||
console.warn(`[SyncService] Unknown operation: ${item.operation}`);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Mark as completed
|
|
||||||
await invoke("sync_mark_completed", { id: item.id });
|
|
||||||
|
|
||||||
// Also mark local data as synced
|
|
||||||
if (item.itemId) {
|
|
||||||
await invoke("storage_mark_synced", {
|
|
||||||
userId: item.userId,
|
|
||||||
itemId: item.itemId,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
console.log(`[SyncService] Successfully processed item ${item.id}`);
|
|
||||||
} catch (error) {
|
|
||||||
console.error(`[SyncService] Failed to process item ${item.id}:`, error);
|
|
||||||
|
|
||||||
// Calculate retry delay with exponential backoff
|
|
||||||
const retryDelay = BASE_RETRY_DELAY_MS * Math.pow(2, item.retryCount);
|
|
||||||
|
|
||||||
// Mark as failed
|
|
||||||
await invoke("sync_mark_failed", {
|
|
||||||
id: item.id,
|
|
||||||
error: error instanceof Error ? error.message : String(error),
|
|
||||||
});
|
|
||||||
|
|
||||||
// Wait before continuing (gives server time to recover if overloaded)
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, Math.min(retryDelay, 10000)));
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -343,6 +152,8 @@ class SyncService {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Clear all sync operations for the current user (called during logout)
|
* Clear all sync operations for the current user (called during logout)
|
||||||
|
*
|
||||||
|
* TRACES: UR-017 | DR-014
|
||||||
*/
|
*/
|
||||||
async clearUser(): Promise<void> {
|
async clearUser(): Promise<void> {
|
||||||
const userId = auth.getUserId();
|
const userId = auth.getUserId();
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Application-wide UI state store
|
||||||
|
// TRACES: UR-005 | DR-005, DR-009
|
||||||
import { writable } from 'svelte/store';
|
import { writable } from 'svelte/store';
|
||||||
|
|
||||||
// App-wide state (root layout)
|
// App-wide state (root layout)
|
||||||
|
|||||||
@ -2,6 +2,7 @@
|
|||||||
//
|
//
|
||||||
// Simplified wrapper over Rust connectivity monitor.
|
// Simplified wrapper over Rust connectivity monitor.
|
||||||
// The Rust backend handles all polling, reachability checks, and adaptive intervals.
|
// The Rust backend handles all polling, reachability checks, and adaptive intervals.
|
||||||
|
// TRACES: UR-002 | DR-013
|
||||||
|
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
import { browser } from "$app/environment";
|
import { browser } from "$app/environment";
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Tests for downloads store
|
||||||
|
// TRACES: UR-011, UR-013, UR-018 | DR-015, DR-017 | UT-010, UT-024
|
||||||
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
import { get } from "svelte/store";
|
import { get } from "svelte/store";
|
||||||
|
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Download manager state store
|
||||||
|
// TRACES: UR-011, UR-013, UR-018 | DR-015, DR-017
|
||||||
import { writable, derived, get } from 'svelte/store';
|
import { writable, derived, get } from 'svelte/store';
|
||||||
import { invoke } from '@tauri-apps/api/core';
|
import { invoke } from '@tauri-apps/api/core';
|
||||||
import { listen, type UnlistenFn } from '@tauri-apps/api/event';
|
import { listen, type UnlistenFn } from '@tauri-apps/api/event';
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Home screen data store - featured items, continue watching, recently added
|
||||||
|
// TRACES: UR-023, UR-024, UR-034 | DR-026, DR-027, DR-038, DR-039
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
import type { MediaItem } from "$lib/api/types";
|
import type { MediaItem } from "$lib/api/types";
|
||||||
import { auth } from "./auth";
|
import { auth } from "./auth";
|
||||||
|
|||||||
@ -1,4 +1,5 @@
|
|||||||
// Library state store
|
// Library state store
|
||||||
|
// TRACES: UR-007, UR-008, UR-029, UR-030 | DR-007, DR-011, DR-033
|
||||||
|
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
import type { Library, MediaItem, SearchResult, Genre } from "$lib/api/types";
|
import type { Library, MediaItem, SearchResult, Genre } from "$lib/api/types";
|
||||||
|
|||||||
@ -5,6 +5,8 @@
|
|||||||
* The backend handles all countdown logic and decisions.
|
* The backend handles all countdown logic and decisions.
|
||||||
*
|
*
|
||||||
* The backend emits ShowNextEpisodePopup and CountdownTick events to update this store.
|
* The backend emits ShowNextEpisodePopup and CountdownTick events to update this store.
|
||||||
|
*
|
||||||
|
* TRACES: UR-023 | DR-026, DR-047, DR-048
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Tests for playback mode store
|
||||||
|
// TRACES: UR-010 | DR-037
|
||||||
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
import { get } from "svelte/store";
|
import { get } from "svelte/store";
|
||||||
|
|
||||||
|
|||||||
@ -6,9 +6,7 @@
|
|||||||
*
|
*
|
||||||
* Most business logic moved to Rust (src-tauri/src/playback_mode/mod.rs)
|
* Most business logic moved to Rust (src-tauri/src/playback_mode/mod.rs)
|
||||||
*
|
*
|
||||||
* @req: UR-010 - Control playback of Jellyfin remote sessions
|
* TRACES: UR-010 | IR-012 | DR-037
|
||||||
* @req: IR-012 - Jellyfin Sessions API for remote playback control
|
|
||||||
* @req: DR-037 - Remote session browser and control UI
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { writable, get, derived } from "svelte/store";
|
import { writable, get, derived } from "svelte/store";
|
||||||
|
|||||||
@ -1,3 +1,5 @@
|
|||||||
|
// Tests for sessions store
|
||||||
|
// TRACES: UR-010 | DR-037
|
||||||
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
import { get } from "svelte/store";
|
import { get } from "svelte/store";
|
||||||
import type { Session } from "$lib/api/types";
|
import type { Session } from "$lib/api/types";
|
||||||
|
|||||||
@ -1,4 +1,5 @@
|
|||||||
// Remote sessions store for controlling playback on other Jellyfin clients
|
// Remote sessions store for controlling playback on other Jellyfin clients
|
||||||
|
// TRACES: UR-010 | DR-037
|
||||||
|
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
import { invoke } from "@tauri-apps/api/core";
|
import { invoke } from "@tauri-apps/api/core";
|
||||||
|
|||||||
@ -5,6 +5,8 @@
|
|||||||
* All logic is in the Rust backend (PlayerController).
|
* All logic is in the Rust backend (PlayerController).
|
||||||
*
|
*
|
||||||
* The backend emits SleepTimerChanged events to update this store.
|
* The backend emits SleepTimerChanged events to update this store.
|
||||||
|
*
|
||||||
|
* TRACES: UR-026 | DR-029
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { writable, derived } from "svelte/store";
|
import { writable, derived } from "svelte/store";
|
||||||
|
|||||||
297
src/lib/utils/debounce.test.ts
Normal file
297
src/lib/utils/debounce.test.ts
Normal file
@ -0,0 +1,297 @@
|
|||||||
|
import { describe, it, expect, vi, beforeEach, afterEach } from "vitest";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Utility function to create debounced functions
|
||||||
|
* Used in GenericMediaListPage for search input debouncing
|
||||||
|
*/
|
||||||
|
export function createDebouncedFunction<T extends (...args: any[]) => any>(
|
||||||
|
fn: T,
|
||||||
|
delayMs: number = 300
|
||||||
|
) {
|
||||||
|
let timeout: ReturnType<typeof setTimeout> | null = null;
|
||||||
|
|
||||||
|
return (...args: Parameters<T>) => {
|
||||||
|
if (timeout) clearTimeout(timeout);
|
||||||
|
|
||||||
|
timeout = setTimeout(() => {
|
||||||
|
fn(...args);
|
||||||
|
timeout = null;
|
||||||
|
}, delayMs);
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
describe("Debounce Utility", () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
vi.useFakeTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
vi.useRealTimers();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Basic Debouncing", () => {
|
||||||
|
it("should delay function execution", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
debouncedFn("test");
|
||||||
|
|
||||||
|
// Should not be called immediately
|
||||||
|
expect(mockFn).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// Advance time by 300ms
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Now it should be called
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("test");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should not call function if timer is cleared before delay", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
debouncedFn("test");
|
||||||
|
vi.advanceTimersByTime(150);
|
||||||
|
|
||||||
|
// Call again before delay completes
|
||||||
|
debouncedFn("updated");
|
||||||
|
|
||||||
|
// First timeout should be cleared
|
||||||
|
vi.advanceTimersByTime(150);
|
||||||
|
|
||||||
|
// Should still not have been called
|
||||||
|
expect(mockFn).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// Complete the second timeout
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should be called once with latest value
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("updated");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle multiple rapid calls", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
// Rapid calls
|
||||||
|
debouncedFn("a");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
debouncedFn("b");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
debouncedFn("c");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// Should not be called yet
|
||||||
|
expect(mockFn).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// Complete the final timeout
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should be called once with the last value
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("c");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should call multiple times if calls are spaced out", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
debouncedFn("first");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should be called
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("first");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(1);
|
||||||
|
|
||||||
|
// Wait enough time and call again
|
||||||
|
vi.advanceTimersByTime(200);
|
||||||
|
debouncedFn("second");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should be called again
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("second");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(2);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Custom Delay", () => {
|
||||||
|
it("should respect custom delay values", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 500);
|
||||||
|
|
||||||
|
debouncedFn("test");
|
||||||
|
|
||||||
|
// 300ms shouldn't trigger
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
expect(mockFn).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// But 500ms should
|
||||||
|
vi.advanceTimersByTime(200);
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("test");
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle zero delay", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 0);
|
||||||
|
|
||||||
|
debouncedFn("test");
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(0);
|
||||||
|
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("test");
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Search Use Case", () => {
|
||||||
|
it("should debounce search queries correctly", () => {
|
||||||
|
const mockSearch = vi.fn();
|
||||||
|
const debouncedSearch = createDebouncedFunction(mockSearch, 300);
|
||||||
|
|
||||||
|
// User types "t"
|
||||||
|
debouncedSearch("t");
|
||||||
|
expect(mockSearch).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// User types "te" quickly
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
debouncedSearch("te");
|
||||||
|
expect(mockSearch).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// User types "tes"
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
debouncedSearch("tes");
|
||||||
|
expect(mockSearch).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// User types "test"
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
debouncedSearch("test");
|
||||||
|
expect(mockSearch).not.toHaveBeenCalled();
|
||||||
|
|
||||||
|
// Wait for debounce delay
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should only call once with final value
|
||||||
|
expect(mockSearch).toHaveBeenCalledWith("test");
|
||||||
|
expect(mockSearch).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should cancel pending search if input clears quickly", () => {
|
||||||
|
const mockSearch = vi.fn();
|
||||||
|
const debouncedSearch = createDebouncedFunction(mockSearch, 300);
|
||||||
|
|
||||||
|
// User types "test"
|
||||||
|
debouncedSearch("test");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// User clears input
|
||||||
|
debouncedSearch("");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// User types again
|
||||||
|
debouncedSearch("new");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should only call with final value
|
||||||
|
expect(mockSearch).toHaveBeenCalledWith("new");
|
||||||
|
expect(mockSearch).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should work with async search functions", () => {
|
||||||
|
const mockAsyncSearch = vi.fn().mockResolvedValue([]);
|
||||||
|
const debouncedSearch = createDebouncedFunction(mockAsyncSearch, 300);
|
||||||
|
|
||||||
|
debouncedSearch("query");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
expect(mockAsyncSearch).toHaveBeenCalledWith("query");
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Generic Parameter Handling", () => {
|
||||||
|
it("should preserve function parameters", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
const obj = { id: "123", name: "test" };
|
||||||
|
debouncedFn("string", 42, obj);
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("string", 42, obj);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle functions with no parameters", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
debouncedFn();
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
expect(mockFn).toHaveBeenCalledWith();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle complex object parameters", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
const options = {
|
||||||
|
query: "test",
|
||||||
|
filters: { type: "Audio", limit: 100 },
|
||||||
|
sort: { by: "SortName", order: "Ascending" },
|
||||||
|
};
|
||||||
|
|
||||||
|
debouncedFn(options);
|
||||||
|
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
expect(mockFn).toHaveBeenCalledWith(options);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe("Memory Management", () => {
|
||||||
|
it("should clean up timeout after execution", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 300);
|
||||||
|
|
||||||
|
debouncedFn("test");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
expect(mockFn).toHaveBeenCalled();
|
||||||
|
|
||||||
|
const callCount = mockFn.mock.calls.length;
|
||||||
|
|
||||||
|
// Call again shortly after
|
||||||
|
debouncedFn("test2");
|
||||||
|
vi.advanceTimersByTime(100);
|
||||||
|
|
||||||
|
// Additional calls within delay shouldn't cause multiple executions
|
||||||
|
debouncedFn("test3");
|
||||||
|
vi.advanceTimersByTime(300);
|
||||||
|
|
||||||
|
// Should only have been called 2 times total
|
||||||
|
expect(mockFn.mock.calls.length).toBe(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should handle repeated debouncing without memory leaks", () => {
|
||||||
|
const mockFn = vi.fn();
|
||||||
|
const debouncedFn = createDebouncedFunction(mockFn, 50);
|
||||||
|
|
||||||
|
// Simulate 100 rapid calls
|
||||||
|
for (let i = 0; i < 100; i++) {
|
||||||
|
debouncedFn(`call${i}`);
|
||||||
|
vi.advanceTimersByTime(10);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Complete final timeout
|
||||||
|
vi.advanceTimersByTime(50);
|
||||||
|
|
||||||
|
// Should only be called once with the last value
|
||||||
|
expect(mockFn).toHaveBeenCalledWith("call99");
|
||||||
|
expect(mockFn).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@ -1,5 +1,7 @@
|
|||||||
/**
|
/**
|
||||||
* Duration formatting utility tests
|
* Duration formatting utility tests
|
||||||
|
*
|
||||||
|
* TRACES: UR-005 | DR-028
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { describe, it, expect } from "vitest";
|
import { describe, it, expect } from "vitest";
|
||||||
|
|||||||
@ -1,5 +1,7 @@
|
|||||||
/**
|
/**
|
||||||
* Input validation utility tests
|
* Input validation utility tests
|
||||||
|
*
|
||||||
|
* TRACES: UR-009, UR-025 | DR-015
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { describe, it, expect } from "vitest";
|
import { describe, it, expect } from "vitest";
|
||||||
|
|||||||
@ -34,7 +34,6 @@
|
|||||||
],
|
],
|
||||||
defaultSort: "SortName",
|
defaultSort: "SortName",
|
||||||
displayComponent: "grid" as const,
|
displayComponent: "grid" as const,
|
||||||
searchFields: ["name", "artists"],
|
|
||||||
};
|
};
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
|
|||||||
@ -26,7 +26,6 @@
|
|||||||
],
|
],
|
||||||
defaultSort: "SortName",
|
defaultSort: "SortName",
|
||||||
displayComponent: "grid" as const,
|
displayComponent: "grid" as const,
|
||||||
searchFields: ["name"],
|
|
||||||
};
|
};
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
|
|||||||
@ -34,7 +34,6 @@
|
|||||||
],
|
],
|
||||||
defaultSort: "SortName",
|
defaultSort: "SortName",
|
||||||
displayComponent: "tracklist" as const,
|
displayComponent: "tracklist" as const,
|
||||||
searchFields: ["name", "artists", "album"],
|
|
||||||
};
|
};
|
||||||
</script>
|
</script>
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user