28 Commits

Author SHA1 Message Date
6a55de3299 Perf: Optimiser builds backend avec image de base (90-95% plus rapide)
All checks were successful
Build Base Docker Image / Build Base Image (push) Successful in 16m30s
Build and Push Docker Images / Build Backend Image (push) Successful in 15m34s
Build and Push Docker Images / Build Frontend Image (push) Successful in 5m51s
Architecture en 2 images:
- Image base (audio-classifier-base): deps système + Python (~15min, 1x/semaine)
- Image app (audio-classifier-backend): code uniquement (~30s-2min, chaque commit)

Fichiers ajoutés:
- backend/Dockerfile.base: Image de base avec toutes les dépendances
- .gitea/workflows/docker-base.yml: CI pour build de l'image de base
- backend/DOCKER_BUILD.md: Documentation complète

Fichiers modifiés:
- backend/Dockerfile: Utilise l'image de base (FROM audio-classifier-base)
- .gitea/workflows/docker.yml: Passe BASE_IMAGE en build-arg

Gains de performance:
- Build normal: 15-25min → 30s-2min (90-95% plus rapide)
- Trigger auto du build base: quand requirements.txt change
- Trigger manuel: via interface Gitea Actions

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 22:04:13 +01:00
f3f321511d Feature: Sélection multiple d'instruments dans les filtres
Frontend:
- FilterPanel: Remplacer select par checkboxes pour instruments
- Zone scrollable (max-height 12rem) pour la liste
- Affichage des instruments sélectionnés dans résumé filtres actifs

Backend:
- API tracks: Nouveau paramètre instruments (List[str])
- Backward compatible avec ancien paramètre instrument
- CRUD: Filtrage AND (track doit avoir TOUS les instruments)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 20:10:39 +01:00
34fcbe1223 Fix: Résoudre tous les conflits d'authentification
All checks were successful
Build and Push Docker Images / Build Frontend Image (push) Successful in 5m31s
Build and Push Docker Images / Build Backend Image (push) Successful in 8m5s
- Nettoyer logs de debug dans auth.py
- Routes /api/audio/* : auth interne au lieu de middleware global
- /stream et /download : token obligatoire en query param (compat <audio>/<a>)
- /waveform : auth standard via header
- Polling scan/status : ajouter Authorization header
- Player : token JWT sur toutes les requêtes (waveform, stream, download)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 19:26:01 +01:00
f05958ed36 J'ai :
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 9m36s
Build and Push Docker Images / Build Frontend Image (push) Successful in 7m23s
Nettoyé les logs de debug dans backend/src/core/auth.py - supprimé tous les logger.info/warning de la fonction authenticate_user
Ajouté les tokens JWT à toutes les requêtes du player :
frontend/components/AudioPlayer.tsx : Ajouté Authorization header à loadWaveform()
frontend/components/AudioPlayer.tsx : Créé getAuthenticatedStreamUrl() qui ajoute le token en query param pour les <audio> et <a> tags
backend/src/api/routes/audio.py : Ajouté support du token en query param pour /stream et /download (compatibilité avec les tags HTML qui ne supportent pas les headers)
Le player devrait maintenant fonctionner entièrement avec l'authentification.
2025-12-26 17:46:39 +01:00
aa252487b8 Fix: Ajouter Authorization header aux requêtes fetch du scan
Problème: Les requêtes fetch() vers /api/library/scan utilisaient
pas l'interceptor axios, donc le token JWT n'était pas envoyé.
Résultat: 403 Forbidden

Solution: Ajouter manuellement le header Authorization avec le token
depuis localStorage pour les requêtes fetch du rescan.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 17:42:54 +01:00
ed7034f55b Fix: Passer les variables d'auth au container backend
Problème: Les variables ADMIN_EMAIL, ADMIN_PASSWORD, JWT_SECRET_KEY
étaient dans le .env mais n'étaient PAS passées au container Docker.
Le backend utilisait donc les valeurs par défaut.

Solution: Ajouter les 4 variables d'auth dans docker-compose.yml
- ADMIN_EMAIL
- ADMIN_PASSWORD
- JWT_SECRET_KEY
- JWT_EXPIRATION_HOURS

Maintenant le container charge les variables depuis le .env du serveur.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 17:39:44 +01:00
0fbfb6f8ed Debug: Ajouter logs détaillés pour authentification
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 9m48s
Build and Push Docker Images / Build Frontend Image (push) Successful in 3m0s
Problème: Login échoue avec 401, besoin de debug
Ajout logs INFO pour:
- Email fourni vs attendu
- Comparaison email
- Longueur des mots de passe
- Résultat authentification

À retirer en production une fois le problème résolu.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 17:20:58 +01:00
16b3fdabed Ajout documentation dépendances + script de vérification
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Successful in 14m27s
Build and Push Docker Images / Build Frontend Image (push) Failing after 11m42s
- DEPENDENCIES.md: Documentation complète de toutes les dépendances
  * Backend Python (requirements.txt)
  * Dépendances système (apt packages)
  * Frontend Node.js (package.json)
  * Modèles Essentia (28 MB)
  * Variables d'environnement requises

- check_dependencies.py: Script pour vérifier l'installation
  * Teste tous les imports Python
  * Affiche statut / pour chaque package
  * Utile pour debug d'installation

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 13:03:17 +01:00
eeee538fcd Fix: Ajouter email-validator pour Pydantic EmailStr
Erreur: ImportError: email-validator is not installed
Cause: EmailStr de Pydantic nécessite email-validator
Fix: Ajout de email-validator==2.1.0 dans requirements.txt

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 13:00:11 +01:00
c366ca5ce0 Include Essentia models in repo + optimize CI/CD
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 13m24s
Build and Push Docker Images / Build Frontend Image (push) Successful in 4m56s
Problème: Les modèles Essentia (28 MB) étaient téléchargés à chaque build CI/CD
- Ralentit les builds (~30 secondes de download)
- Consomme bande passante
- Point de défaillance si serveur Essentia down

Solution:
- Commit les 6 modèles dans backend/models/
- Supprime steps "Download Essentia models" du workflow Gitea
- Retire backend/models/*.pb et *.json du .gitignore

Modèles inclus (~28 MB total):
- discogs-effnet-bs64-1.pb (18 MB) - embedding model
- genre_discogs400-discogs-effnet-1.pb (2 MB) - genre classifier
- genre_discogs400-discogs-effnet-1.json (15 KB) - genre metadata
- mtg_jamendo_moodtheme-discogs-effnet-1.pb (2.6 MB) - mood
- mtg_jamendo_instrument-discogs-effnet-1.pb (2.6 MB) - instruments
- mtg_jamendo_genre-discogs-effnet-1.pb (2.7 MB) - genre alt

Bénéfices:
 Builds CI/CD plus rapides (~30s gagnées)
 Pas de dépendance externe au serveur Essentia
 Versioning des modèles avec le code
 Repo offline-friendly

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 10:11:03 +01:00
774cb799a2 Ajout authentification JWT complète (app 100% protégée)
Backend:
- Nouveau module auth.py avec JWT et password handling
- Endpoint /api/auth/login (public)
- Endpoint /api/auth/me (protégé)
- TOUS les endpoints API protégés par require_auth
- Variables env: ADMIN_EMAIL, ADMIN_PASSWORD, JWT_SECRET_KEY
- Dependencies: python-jose, passlib

Frontend:
- Page de login (/login)
- AuthGuard component pour redirection automatique
- Axios interceptor: ajoute JWT token à chaque requête
- Gestion erreur 401: redirect automatique vers /login
- Bouton logout dans header
- Token stocké dans localStorage

Configuration:
- .env.example mis à jour avec variables auth
- Credentials admin configurables via env

Sécurité:
- Aucun endpoint public (sauf /api/auth/login et /health)
- JWT expiration configurable (24h par défaut)
- Password en clair dans env (à améliorer avec hash en prod)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 10:05:36 +01:00
6ae861ff54 Actualiser .gitea/workflows/docker.yml
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 12m28s
Build and Push Docker Images / Build Frontend Image (push) Successful in 58s
Remise en état build backend
2025-12-26 00:15:54 +01:00
b74c6b0b40 Fix scan infini: exclure dossiers transcoded et waveforms
All checks were successful
Build and Push Docker Images / Build Frontend Image (push) Successful in 57s
Problème: Le scanner scannait TOUS les dossiers, y compris les dossiers
générés (transcoded/ et waveforms/), créant:
1. Boucle infinie: scan original → crée transcoded → re-scan transcoded
2. Segfaults: tentative de transcoder des fichiers déjà transcodés
3. Doublons en base de données

Solution:
- library.py: Exclut transcoded, waveforms, .transcoded, .waveforms
- scanner.py: Même exclusion dans le CLI

Technique: Modifie dirs[:] dans os.walk() pour skip ces dossiers.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-26 00:08:13 +01:00
4d8fa57ab2 Fix tous les appels API pour utiliser getApiUrl() au lieu de process.env
All checks were successful
Build and Push Docker Images / Build Frontend Image (push) Successful in 3m31s
Problème: Le commit précédent n'avait corrigé que api.ts, mais AudioPlayer
et page.tsx utilisaient encore directement process.env.NEXT_PUBLIC_API_URL,
ce qui ignorait la config runtime.

Fichiers corrigés:
1. lib/api.ts:
   - Export getApiUrl() pour usage externe

2. app/page.tsx:
   - Import getApiUrl
   - /api/library/scan: process.env → getApiUrl()
   - /api/library/scan/status: process.env → getApiUrl()

3. components/AudioPlayer.tsx:
   - Import getApiUrl
   - /api/audio/waveform: process.env → getApiUrl()
   - /api/audio/stream: process.env → getApiUrl()
   - /api/audio/download: process.env → getApiUrl()

Maintenant TOUS les appels API utilisent la config runtime
(window.__RUNTIME_CONFIG__) côté client.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-24 10:54:38 +01:00
36652ea2cc Fix API URL configuration pour production
All checks were successful
Build and Push Docker Images / Build Frontend Image (push) Successful in 3m59s
Problème: Le frontend utilisait localhost:8001 au lieu de l'URL de production
car NEXT_PUBLIC_API_URL était évalué au build time et non au runtime.

Changements:
1. Frontend (api.ts):
   - Remplace apiClient statique par getApiClient() dynamique
   - Chaque appel crée une instance axios avec l'URL runtime
   - getStreamUrl/getDownloadUrl utilisent getApiUrl() au lieu de API_BASE_URL
   - Supprime l'export default apiClient (non utilisé)

2. Docker Compose:
   - Configure NEXT_PUBLIC_API_URL=https://api.audioclassifier.benoitsz.com
   - Simplifie la config (retire le fallback)

Le runtime config (window.__RUNTIME_CONFIG__) fonctionne maintenant correctement
car il est évalué à chaque appel API côté client.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-12-24 10:12:38 +01:00
3b236d6af4 Merge pull request 'Fix localhost en prod' (#4) from prochains-changements into main
All checks were successful
Build and Push Docker Images / Build Frontend Image (push) Successful in 5m45s
Reviewed-on: #4
2025-12-23 15:34:34 +01:00
04603cd5e9 Fix localhost en prod
Résumé des modifications
J'ai implémenté une configuration runtime pour le frontend qui permet de changer l'URL de l'API sans rebuilder l'image Docker. Voici ce qui a été fait :
📝 Fichiers modifiés :
frontend/Dockerfile - Ajout du script de génération de config au démarrage
frontend/lib/api.ts - Lecture de la config depuis window.__RUNTIME_CONFIG__
frontend/app/layout.tsx - Chargement du script config.js
docker-compose.yml - Utilisation de la variable d'environnement
.env.example - Documentation de la variable
DEPLOYMENT.md - Documentation de la configuration runtime
📄 Fichiers créés :
frontend/generate-config.sh - Script de génération de config
frontend/public/config.js - Fichier de config (placeholder)
frontend/README.md - Documentation du frontend
🚀 Pour résoudre votre problème en production :
Sur votre serveur, modifiez le fichier .env :

# URL publique de l'API (utilisée par le navigateur)
NEXT_PUBLIC_API_URL=https://audioclassifier.benoitsz.com:8001

# CORS doit accepter les requêtes du frontend
CORS_ORIGINS=https://audioclassifier.benoitsz.com,https://audioclassifier.benoitsz.com:3000
Ensuite :

# Pull les dernières modifications
git pull

# Rebuild l'image frontend (une seule fois)
# Soit via Gitea Actions en poussant sur main
# Soit manuellement :
# docker build -t git.benoitsz.com/benoit/audio-classifier-frontend:dev -f frontend/Dockerfile .
# docker push git.benoitsz.com/benoit/audio-classifier-frontend:dev

# Redémarrer les containers
docker-compose down
docker-compose up -d
 Avantages :
 Aucun rebuild nécessaire après le premier déploiement
 Configuration flexible via variables d'environnement
 Compatible avec tous les environnements (dev, staging, prod)
 Testé et fonctionnel localement
Le frontend générera automatiquement le bon fichier de configuration au démarrage du container avec l'URL de votre serveur !
2025-12-23 15:33:52 +01:00
64ba7f9006 Merge pull request 'Fix CORS' (#3) from prochains-changements into main
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 12m50s
Build and Push Docker Images / Build Frontend Image (push) Successful in 7m10s
Reviewed-on: #3
2025-12-23 14:34:05 +01:00
cc2f1d0051 Fix CORS 2025-12-23 14:33:25 +01:00
169a759b57 Fix Build backend Gitea
All checks were successful
Build and Push Docker Images / Build Backend Image (push) Successful in 27m9s
Build and Push Docker Images / Build Frontend Image (push) Successful in 56s
2025-12-23 13:27:50 +01:00
88db8cc9c8 Fix build backend depuis Gitea 2025-12-23 13:27:33 +01:00
3e225b158f Fix build et actions
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Failing after 37s
Build and Push Docker Images / Build Frontend Image (push) Has been cancelled
2025-12-23 13:23:07 +01:00
8ec8b1aa42 Merge branch 'main' of https://git.benoitsz.com/benoit/Audio-Classifier
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Failing after 38s
Build and Push Docker Images / Build Frontend Image (push) Failing after 2m6s
2025-12-23 13:10:37 +01:00
e3d85f4775 Merge branch 'Backend'
Merge Backend
2025-12-23 13:08:43 +01:00
2a0d022e37 Fix Actions avec qwen
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Failing after 37s
Build and Push Docker Images / Build Frontend Image (push) Failing after 46s
2025-12-23 12:10:51 +01:00
5fb56a636f Fix Gitea Actions
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Failing after 38s
Build and Push Docker Images / Build Frontend Image (push) Failing after 46s
2025-12-23 12:03:55 +01:00
721f7b51f7 Ajouter .gitea/workflows/docker.yml
Some checks failed
Build and Push Docker Images / Build Backend Image (push) Failing after 38s
Build and Push Docker Images / Build Frontend Image (push) Failing after 3m15s
2025-12-23 11:24:24 +01:00
54086236c6 Merge pull request 'Backend' (#1) from Backend into main
Reviewed-on: #1
2025-12-23 10:58:10 +01:00
46 changed files with 2107 additions and 157 deletions

View File

@@ -10,7 +10,8 @@
"Bash(curl:*)",
"Bash(docker logs:*)",
"Bash(docker exec:*)",
"Bash(ls:*)"
"Bash(ls:*)",
"Bash(docker build:*)"
]
}
}

View File

@@ -5,7 +5,9 @@ POSTGRES_PASSWORD=audio_password
POSTGRES_DB=audio_classifier
# Backend API
CORS_ORIGINS=http://localhost:3000,http://127.0.0.1:3000
# Use "*" to allow all origins (recommended for development/local deployment)
# Or specify comma-separated URLs for production: http://yourdomain.com,https://yourdomain.com
CORS_ORIGINS=*
API_HOST=0.0.0.0
API_PORT=8000
@@ -15,5 +17,14 @@ ANALYSIS_NUM_WORKERS=4
ESSENTIA_MODELS_PATH=/app/models
AUDIO_LIBRARY_PATH=/path/to/your/audio/library
# Authentication
ADMIN_EMAIL=admin@example.com
ADMIN_PASSWORD=changeme
JWT_SECRET_KEY=your-super-secret-jwt-key-change-this-in-production
JWT_EXPIRATION_HOURS=24
# Frontend
NEXT_PUBLIC_API_URL=http://localhost:8000
# API URL accessed by the browser (use port 8001 since backend is mapped to 8001)
# For production on a remote server, set this to your server's public URL
# Example: NEXT_PUBLIC_API_URL=http://yourserver.com:8001
NEXT_PUBLIC_API_URL=http://localhost:8001

View File

@@ -0,0 +1,61 @@
name: Build Base Docker Image
# Build base image only when requirements.txt changes or manually triggered
on:
push:
branches:
- main
paths:
- 'backend/requirements.txt'
- 'backend/Dockerfile.base'
workflow_dispatch: # Allow manual trigger
env:
REGISTRY: git.benoitsz.com
IMAGE_BASE: audio-classifier-base
jobs:
build-base:
name: Build Base Image
runs-on: ubuntu-latest
steps:
- name: Checkout code
uses: actions/checkout@v4
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v3
- name: Login to Gitea Container Registry
uses: docker/login-action@v3
with:
registry: ${{ env.REGISTRY }}
username: ${{ gitea.actor }}
password: ${{ secrets.REGISTRY_TOKEN }}
- name: Extract metadata
id: meta
uses: docker/metadata-action@v5
with:
images: ${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BASE }}
tags: |
type=raw,value=latest
type=sha,prefix=sha-,format=short
- name: Build and push base image
uses: docker/build-push-action@v5
with:
context: ./backend
file: ./backend/Dockerfile.base
push: true
tags: ${{ steps.meta.outputs.tags }}
labels: ${{ steps.meta.outputs.labels }}
cache-from: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BASE }}:buildcache
cache-to: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BASE }}:buildcache,mode=max
platforms: linux/amd64
- name: Image built successfully
run: |
echo "✅ Base image built and pushed successfully"
echo "📦 Image: ${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BASE }}:latest"
echo "⏱️ This image will be used by the main backend builds to speed up CI/CD"

124
.gitea/workflows/docker.yml Normal file
View File

@@ -0,0 +1,124 @@
name: Build and Push Docker Images
on:
push:
branches:
- main
tags:
- 'v*.*.*'
env:
REGISTRY: git.benoitsz.com
IMAGE_BACKEND: audio-classifier-backend
IMAGE_FRONTEND: audio-classifier-frontend
jobs:
build-backend:
name: Build Backend Image
runs-on: ubuntu-latest
steps:
- name: Checkout code
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v3
- name: Login to Gitea Container Registry
uses: docker/login-action@v3
with:
registry: ${{ env.REGISTRY }}
username: ${{ gitea.actor }}
password: ${{ secrets.REGISTRY_TOKEN }}
- name: Determine version
id: version
run: |
if [[ "${{ gitea.ref }}" == refs/tags/v* ]]; then
echo "VERSION=${GITEA_REF#refs/tags/}" >> $GITHUB_OUTPUT
else
echo "VERSION=dev-$(git rev-parse --short HEAD)" >> $GITHUB_OUTPUT
fi
- name: Extract metadata
id: meta
uses: docker/metadata-action@v5
with:
images: ${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BACKEND }}
tags: |
type=semver,pattern={{version}}
type=semver,pattern={{major}}.{{minor}}
type=raw,value=latest,enable=${{ startsWith(gitea.ref, 'refs/tags/v') }}
type=raw,value=dev,enable=${{ gitea.ref == 'refs/heads/main' }}
type=sha,prefix=dev-,format=short,enable=${{ gitea.ref == 'refs/heads/main' }}
- name: Build and push backend
uses: docker/build-push-action@v5
with:
context: .
file: ./backend/Dockerfile
push: true
build-args: |
VERSION=${{ steps.version.outputs.VERSION }}
BASE_IMAGE=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/audio-classifier-base:latest
tags: ${{ steps.meta.outputs.tags }}
labels: ${{ steps.meta.outputs.labels }}
cache-from: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BACKEND }}:buildcache
cache-to: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_BACKEND }}:buildcache,mode=max
platforms: linux/amd64
build-frontend:
name: Build Frontend Image
runs-on: ubuntu-latest
steps:
- name: Checkout code
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v3
- name: Login to Gitea Container Registry
uses: docker/login-action@v3
with:
registry: ${{ env.REGISTRY }}
username: ${{ gitea.actor }}
password: ${{ secrets.REGISTRY_TOKEN }}
- name: Determine version
id: version
run: |
if [[ "${{ gitea.ref }}" == refs/tags/v* ]]; then
echo "VERSION=${GITEA_REF#refs/tags/}" >> $GITHUB_OUTPUT
else
echo "VERSION=dev-$(git rev-parse --short HEAD)" >> $GITHUB_OUTPUT
fi
- name: Extract metadata
id: meta
uses: docker/metadata-action@v5
with:
images: ${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_FRONTEND }}
tags: |
type=semver,pattern={{version}}
type=semver,pattern={{major}}.{{minor}}
type=raw,value=latest,enable=${{ startsWith(gitea.ref, 'refs/tags/v') }}
type=raw,value=dev,enable=${{ gitea.ref == 'refs/heads/main' }}
type=sha,prefix=dev-,format=short,enable=${{ gitea.ref == 'refs/heads/main' }}
- name: Build and push frontend
uses: docker/build-push-action@v5
with:
context: .
file: ./frontend/Dockerfile
push: true
build-args: |
VERSION=${{ steps.version.outputs.VERSION }}
tags: ${{ steps.meta.outputs.tags }}
labels: ${{ steps.meta.outputs.labels }}
cache-from: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_FRONTEND }}:buildcache
cache-to: type=registry,ref=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/${{ env.IMAGE_FRONTEND }}:buildcache,mode=max

4
.gitignore vendored
View File

@@ -75,10 +75,6 @@ yarn-error.log*
# Docker
postgres_data/
# Essentia models (large files, download separately)
backend/models/*.pb
backend/models/*.json
# Audio analysis cache
*.peaks.json
.audio_cache/

140
DEPENDENCIES.md Normal file
View File

@@ -0,0 +1,140 @@
# Dépendances du projet
## Backend Python (requirements.txt)
### Web Framework
- `fastapi==0.109.0` - Framework web moderne
- `uvicorn[standard]==0.27.0` - Serveur ASGI
- `python-multipart==0.0.6` - Support formulaires multipart
### Database
- `sqlalchemy==2.0.25` - ORM
- `psycopg2-binary==2.9.9` - Driver PostgreSQL
- `pgvector==0.2.4` - Extension vecteurs PostgreSQL
- `alembic==1.13.1` - Migrations de base de données
### Audio Processing
- `librosa==0.10.1` - Analyse audio
- `soundfile==0.12.1` - Lecture/écriture fichiers audio
- `audioread==3.0.1` - Décodage formats audio
- `mutagen==1.47.0` - Métadonnées ID3
### Machine Learning
- `essentia-tensorflow` - Classification genre/mood/instruments (installé via Dockerfile)
- `numpy==1.24.3` - Calcul numérique
- `scipy==1.11.4` - Calcul scientifique
### Configuration & Validation
- `pydantic==2.5.3` - Validation de données
- `pydantic-settings==2.1.0` - Configuration via env vars
- `python-dotenv==1.0.0` - Chargement fichier .env
- `email-validator==2.1.0` - Validation emails (requis par Pydantic EmailStr)
### Authentication
- `python-jose[cryptography]==3.3.0` - JWT tokens
- `passlib[bcrypt]==1.7.4` - Hashing passwords
### Utilities
- `aiofiles==23.2.1` - I/O fichiers asynchrones
- `httpx==0.26.0` - Client HTTP asynchrone
## Dépendances Système (Dockerfile)
### Requis pour le backend
```bash
apt-get install -y \
ffmpeg # Transcodage audio (MP3, etc.)
libsndfile1 # Lecture formats audio
gcc g++ gfortran # Compilation packages Python
libopenblas-dev # Algèbre linéaire optimisée
liblapack-dev # Routines algèbre linéaire
libfftw3-dev # Transformées de Fourier rapides
libavcodec-dev # Codecs audio/vidéo
libavformat-dev # Formats conteneurs
libavutil-dev # Utilitaires FFmpeg
libswresample-dev # Resampling audio
libsamplerate0-dev # Conversion taux d'échantillonnage
libtag1-dev # Métadonnées audio
libchromaprint-dev # Audio fingerprinting
```
## Frontend (package.json)
### Framework
- `next@15.5.6` - Framework React
- `react@19.0.0` - Bibliothèque UI
- `react-dom@19.0.0` - Rendu React
### State Management & Data Fetching
- `@tanstack/react-query@5.62.11` - Gestion état serveur
- `axios@1.7.9` - Client HTTP
### UI & Styling
- `tailwindcss@3.4.17` - Framework CSS utility-first
### Types
- `typescript@5.7.2` - Typage statique
- `@types/react@19.0.1`
- `@types/node@22.10.1`
## Modèles Essentia (inclus dans le repo)
Total: ~28 MB
- `discogs-effnet-bs64-1.pb` (18 MB) - Modèle d'embedding
- `genre_discogs400-discogs-effnet-1.pb` (2 MB) - Classification genre
- `genre_discogs400-discogs-effnet-1.json` (15 KB) - Métadonnées genres
- `mtg_jamendo_moodtheme-discogs-effnet-1.pb` (2.6 MB) - Classification mood
- `mtg_jamendo_instrument-discogs-effnet-1.pb` (2.6 MB) - Classification instruments
- `mtg_jamendo_genre-discogs-effnet-1.pb` (2.7 MB) - Classification genre (alternatif)
## Vérification des dépendances
### Backend
```bash
cd backend
python check_dependencies.py
```
### Build Docker
```bash
# Backend
docker build -t audio-classifier-backend -f backend/Dockerfile .
# Frontend
docker build -t audio-classifier-frontend -f frontend/Dockerfile .
```
## Notes de compatibilité
- **Python**: 3.9 (requis pour essentia-tensorflow)
- **Architecture**: amd64 (meilleure compatibilité Essentia)
- **Node.js**: 20+ (pour Next.js 15)
- **PostgreSQL**: 16+ avec extension pgvector
## Installation locale
### Backend
```bash
cd backend
python -m venv venv
source venv/bin/activate # Windows: venv\Scripts\activate
pip install -r requirements.txt
pip install essentia-tensorflow
```
### Frontend
```bash
cd frontend
npm install
```
## Variables d'environnement requises
Voir `.env.example` pour la liste complète des variables nécessaires.
### Critiques
- `DATABASE_URL` - Connexion PostgreSQL
- `ADMIN_EMAIL` - Email admin
- `ADMIN_PASSWORD` - Mot de passe admin
- `JWT_SECRET_KEY` - Secret pour JWT (générer avec `openssl rand -hex 32`)

View File

@@ -14,7 +14,7 @@ Le système est **100% autonome** - aucune action manuelle requise ! Les modèle
1. **Cloner le projet** :
```bash
git clone <votre-repo>
git clone https://git.benoitsz.com/benoit/Audio-Classifier.git
cd Audio-Classifier
```
@@ -36,6 +36,8 @@ docker-compose up -d
C'est tout ! 🎉
**Note** : Les images Docker sont automatiquement téléchargées depuis git.benoitsz.com. Aucun build nécessaire !
### Premier Scan
1. Ouvrir http://localhost:3000
@@ -202,13 +204,20 @@ cd Audio-Classifier
# Chemin vers musique
AUDIO_LIBRARY_PATH=/mnt/musique
# Domaine public
CORS_ORIGINS=http://votre-domaine.com,https://votre-domaine.com
# URL publique de l'API (IMPORTANT pour le frontend)
# Cette URL est utilisée par le navigateur pour accéder à l'API
# Remplacer par votre domaine ou IP publique + port 8001
NEXT_PUBLIC_API_URL=https://votre-serveur.com:8001
# Domaine public pour CORS (doit inclure l'URL du frontend)
CORS_ORIGINS=https://votre-domaine.com,https://votre-domaine.com:3000
# Credentials BDD (sécurisés !)
POSTGRES_PASSWORD=motdepasse_fort_aleatoire
```
**Important :** Le frontend utilise maintenant une configuration **runtime**, ce qui signifie que vous pouvez changer `NEXT_PUBLIC_API_URL` dans le fichier `.env` et redémarrer les containers sans avoir à rebuilder les images.
4. **Démarrer** :
```bash
docker-compose up -d

View File

@@ -41,8 +41,8 @@ Outil de classification audio automatique capable d'indexer et analyser des bibl
```bash
# 1. Cloner le projet
git clone <repo>
cd audio-classifier
git clone https://git.benoitsz.com/benoit/Audio-Classifier.git
cd Audio-Classifier
# 2. Configurer le chemin audio (optionnel)
echo "AUDIO_LIBRARY_PATH=/chemin/vers/votre/musique" > .env
@@ -53,6 +53,8 @@ docker-compose up -d
**C'est tout !** 🎉
Les images Docker sont automatiquement téléchargées depuis le registry Gitea.
- Frontend : http://localhost:3000
- API : http://localhost:8001
- API Docs : http://localhost:8001/docs
@@ -66,13 +68,26 @@ docker-compose up -d
### ✨ Particularités
- **Aucun téléchargement manuel** : Les modèles Essentia (28 MB) sont inclus dans l'image Docker
- **Images pré-construites** : Téléchargées automatiquement depuis git.benoitsz.com
- **Modèles inclus** : Les modèles Essentia (28 MB) sont intégrés dans l'image
- **Aucune configuration** : Tout fonctionne out-of-the-box
- **Transcodage automatique** : MP3 128kbps créés pour streaming rapide
- **Waveforms pré-calculées** : Chargement instantané
📖 **Documentation complète** : Voir [DEPLOYMENT.md](DEPLOYMENT.md)
### 🛠 Build local (développement)
Si vous voulez builder les images localement, les modèles Essentia doivent être présents dans `backend/models/` (28 MB).
```bash
# Build avec docker-compose
docker-compose -f docker-compose.build.yml build
docker-compose -f docker-compose.build.yml up -d
```
**Note** : Les modèles Essentia (`.pb`, 28 MB) ne sont pas versionnés dans Git. Le workflow CI/CD les télécharge automatiquement depuis essentia.upf.edu pendant le build.
## 📖 Utilisation
### Scanner un dossier

136
backend/DOCKER_BUILD.md Normal file
View File

@@ -0,0 +1,136 @@
# Docker Build Optimization
Cette configuration utilise une approche en 2 images pour accélérer les builds backend de **15-25 minutes** à **30 secondes - 2 minutes**.
## Architecture
### Image 1 : Base (`audio-classifier-base`)
Contient toutes les dépendances système et Python qui changent rarement :
- Python 3.9 + apt packages (ffmpeg, libsndfile, etc.)
- numpy, scipy, essentia-tensorflow
- Toutes les dépendances de `requirements.txt`
**Build** : ~15 minutes (1 fois par semaine ou quand `requirements.txt` change)
### Image 2 : App (`audio-classifier-backend`)
Hérite de l'image de base et ajoute uniquement le code applicatif :
- Code source (`src/`)
- Fichiers de configuration (`alembic.ini`)
- Modèles Essentia (`models/`)
**Build** : ~30 secondes - 2 minutes (à chaque commit)
## Workflows CI/CD
### 1. Build de l'image de base (`.gitea/workflows/docker-base.yml`)
Se déclenche automatiquement quand :
- `backend/requirements.txt` est modifié
- `backend/Dockerfile.base` est modifié
- Déclenchement manuel via l'interface Gitea
```bash
# Image produite :
git.benoitsz.com/benoit/audio-classifier-base:latest
git.benoitsz.com/benoit/audio-classifier-base:sha-<commit>
```
### 2. Build de l'image app (`.gitea/workflows/docker.yml`)
Se déclenche à chaque push sur `main` :
- Utilise l'image de base comme FROM
- Copie uniquement le code source
- Build rapide (~30s-2min)
```bash
# Image produite :
git.benoitsz.com/benoit/audio-classifier-backend:dev
git.benoitsz.com/benoit/audio-classifier-backend:dev-<commit>
```
## Utilisation en local
### Build de l'image de base
```bash
cd backend
docker build -f Dockerfile.base -t audio-classifier-base:local .
```
### Build de l'image app (utilise l'image de base)
```bash
# Depuis la racine du projet
docker build \
--build-arg BASE_IMAGE=audio-classifier-base:local \
-f backend/Dockerfile \
-t audio-classifier-backend:local \
.
```
### Build direct (sans image de base) - pour tests
Si tu veux tester un build complet sans dépendre de l'image de base :
```bash
# Revenir temporairement au Dockerfile original
git show HEAD~1:backend/Dockerfile > backend/Dockerfile.monolithic
docker build -f backend/Dockerfile.monolithic -t audio-classifier-backend:monolithic .
```
## Mise à jour des dépendances
Quand tu modifies `requirements.txt` :
1. **Push les changements sur `main`**
```bash
git add backend/requirements.txt
git commit -m "Update dependencies"
git push
```
2. **Le workflow `docker-base.yml` se déclenche automatiquement**
- Build de la nouvelle image de base (~15 min)
- Push vers `git.benoitsz.com/benoit/audio-classifier-base:latest`
3. **Les prochains builds backend utiliseront la nouvelle base**
- Builds futurs rapides (~30s-2min)
## Déclenchement manuel
Pour rebuild l'image de base manuellement (sans modifier `requirements.txt`) :
1. Va sur Gitea : `https://git.benoitsz.com/benoit/audio-classifier/actions`
2. Sélectionne le workflow "Build Base Docker Image"
3. Clique sur "Run workflow"
## Monitoring
Vérifie les builds dans Gitea Actions :
- **Base image** : `.gitea/workflows/docker-base.yml`
- **App image** : `.gitea/workflows/docker.yml`
Les logs montrent la durée de build pour chaque étape.
## Gains de performance attendus
| Scénario | Avant | Après | Gain |
|----------|-------|-------|------|
| Build normal (code change) | 15-25 min | 30s-2min | **90-95%** |
| Build après update deps | 15-25 min | 15-25 min (base) + 30s-2min (app) | 0% (1ère fois) |
| Builds suivants | 15-25 min | 30s-2min | **90-95%** |
## Troubleshooting
### Erreur "base image not found"
L'image de base n'existe pas encore dans le registry. Solutions :
1. Trigger le workflow `docker-base.yml` manuellement
2. Ou build localement et push :
```bash
docker build -f backend/Dockerfile.base -t git.benoitsz.com/benoit/audio-classifier-base:latest backend/
docker push git.benoitsz.com/benoit/audio-classifier-base:latest
```
### Build app lent malgré l'image de base
Vérifie que le build-arg `BASE_IMAGE` est bien passé :
```yaml
build-args: |
BASE_IMAGE=${{ env.REGISTRY }}/${{ gitea.repository_owner }}/audio-classifier-base:latest
```
### Dépendances Python pas à jour dans l'app
L'image de base doit être rebuildée. Trigger `docker-base.yml`.

View File

@@ -1,55 +1,18 @@
# Use amd64 platform for better Essentia compatibility, works with emulation on ARM
FROM --platform=linux/amd64 python:3.9-slim
# Use pre-built base image with all dependencies
# Base image includes: Python 3.9, system deps, numpy, scipy, essentia-tensorflow, all pip deps
# Only rebuild base when requirements.txt changes
ARG BASE_IMAGE=git.benoitsz.com/benoit/audio-classifier-base:latest
FROM ${BASE_IMAGE}
# Install system dependencies
RUN apt-get update && apt-get install -y \
ffmpeg \
libsndfile1 \
libsndfile1-dev \
gcc \
g++ \
gfortran \
libopenblas-dev \
liblapack-dev \
pkg-config \
curl \
build-essential \
libyaml-dev \
libfftw3-dev \
libavcodec-dev \
libavformat-dev \
libavutil-dev \
libswresample-dev \
libsamplerate0-dev \
libtag1-dev \
libchromaprint-dev \
&& rm -rf /var/lib/apt/lists/*
# Set working directory
# Working directory already set in base image
WORKDIR /app
# Upgrade pip, setuptools, wheel
RUN pip install --no-cache-dir --upgrade pip setuptools wheel
# Copy requirements
COPY requirements.txt .
# Install Python dependencies in stages for better caching
# Using versions compatible with Python 3.9
RUN pip install --no-cache-dir numpy==1.24.3
RUN pip install --no-cache-dir scipy==1.11.4
# Install Essentia-TensorFlow - Python 3.9 AMD64 support
RUN pip install --no-cache-dir essentia-tensorflow
RUN pip install --no-cache-dir -r requirements.txt
# Copy application code
COPY src/ ./src/
COPY alembic.ini .
COPY backend/src/ ./src/
COPY backend/alembic.ini .
# Copy Essentia models into image
COPY models/ ./models/
# Copy Essentia models into image (28 MB total)
COPY backend/models/ ./models/
RUN ls -lh /app/models
# Expose port

59
backend/Dockerfile.base Normal file
View File

@@ -0,0 +1,59 @@
# Base image for Audio Classifier Backend
# This image contains all system dependencies and Python packages
# Build this image only when dependencies change (requirements.txt updates)
# Use amd64 platform for better Essentia compatibility
FROM --platform=linux/amd64 python:3.9-slim
LABEL maintainer="benoit.schw@gmail.com"
LABEL description="Base image with all dependencies for Audio Classifier Backend"
# Install system dependencies
RUN apt-get update && apt-get install -y \
ffmpeg \
libsndfile1 \
libsndfile1-dev \
gcc \
g++ \
gfortran \
libopenblas-dev \
liblapack-dev \
pkg-config \
curl \
build-essential \
libyaml-dev \
libfftw3-dev \
libavcodec-dev \
libavformat-dev \
libavutil-dev \
libswresample-dev \
libsamplerate0-dev \
libtag1-dev \
libchromaprint-dev \
&& rm -rf /var/lib/apt/lists/*
# Set working directory
WORKDIR /app
# Upgrade pip, setuptools, wheel
RUN pip install --no-cache-dir --upgrade pip setuptools wheel
# Copy requirements
COPY requirements.txt .
# Install Python dependencies in stages for better caching
# Using versions compatible with Python 3.9
RUN pip install --no-cache-dir numpy==1.24.3
RUN pip install --no-cache-dir scipy==1.11.4
# Install Essentia-TensorFlow - Python 3.9 AMD64 support
RUN pip install --no-cache-dir essentia-tensorflow
# Install remaining dependencies
RUN pip install --no-cache-dir -r requirements.txt
# Verify installations
RUN python -c "import essentia.standard; import numpy; import scipy; import fastapi; print('All dependencies installed successfully')"
# This image is meant to be used as a base
# The application code will be copied in the derived Dockerfile

View File

@@ -0,0 +1,76 @@
#!/usr/bin/env python3
"""Check all required dependencies are installed."""
import sys
def check_import(module_name, package_name=None):
"""Try to import a module and report status."""
package = package_name or module_name
try:
__import__(module_name)
print(f"{package}")
return True
except ImportError as e:
print(f"{package}: {e}")
return False
def main():
"""Check all dependencies."""
print("🔍 Checking Python dependencies...\n")
dependencies = [
# Web Framework
("fastapi", "fastapi"),
("uvicorn", "uvicorn"),
("multipart", "python-multipart"),
# Database
("sqlalchemy", "sqlalchemy"),
("psycopg2", "psycopg2-binary"),
("pgvector.sqlalchemy", "pgvector"),
("alembic", "alembic"),
# Audio Processing
("librosa", "librosa"),
("soundfile", "soundfile"),
("audioread", "audioread"),
("mutagen", "mutagen"),
# Scientific
("numpy", "numpy"),
("scipy", "scipy"),
# Configuration
("pydantic", "pydantic"),
("pydantic_settings", "pydantic-settings"),
("dotenv", "python-dotenv"),
("email_validator", "email-validator"),
# Authentication
("jose", "python-jose"),
("passlib", "passlib"),
# Utilities
("aiofiles", "aiofiles"),
("httpx", "httpx"),
# Essentia (optional)
("essentia.standard", "essentia-tensorflow"),
]
all_ok = True
for module, package in dependencies:
if not check_import(module, package):
all_ok = False
print("\n" + "="*50)
if all_ok:
print("✅ All dependencies installed!")
return 0
else:
print("❌ Some dependencies are missing")
print("\nInstall missing dependencies with:")
print(" pip install -r requirements.txt")
return 1
if __name__ == "__main__":
sys.exit(main())

52
backend/models/README.md Normal file
View File

@@ -0,0 +1,52 @@
# Essentia Models
Ce dossier contient les modèles pré-entraînés Essentia-TensorFlow pour la classification audio (28 MB total).
## Modèles requis
Les fichiers suivants sont nécessaires pour le fonctionnement de l'application :
1. **discogs-effnet-bs64-1.pb** (18 MB) - Embedding model
2. **genre_discogs400-discogs-effnet-1.pb** (2 MB) - Genre classifier
3. **genre_discogs400-discogs-effnet-1.json** (15 KB) - Genre metadata
4. **mtg_jamendo_moodtheme-discogs-effnet-1.pb** (2.7 MB) - Mood classifier
5. **mtg_jamendo_instrument-discogs-effnet-1.pb** (2.6 MB) - Instrument classifier
6. **mtg_jamendo_genre-discogs-effnet-1.pb** (2.7 MB) - Alternative genre classifier
## Téléchargement automatique
**Pour les utilisateurs** : Les modèles sont déjà inclus dans les images Docker depuis le registry `git.benoitsz.com`. Aucune action nécessaire.
**Pour le CI/CD** : Les modèles sont téléchargés automatiquement depuis essentia.upf.edu pendant le build (voir `.gitea/workflows/docker.yml`).
**Pour le développement local** : Si vous avez besoin de builder localement, vous devez avoir les modèles dans ce dossier. Ils ne sont pas versionnés dans Git car ils pèsent 28 MB.
### Téléchargement manuel (si nécessaire)
```bash
cd backend/models
# Embedding model (18 MB)
curl -L -O https://essentia.upf.edu/models/feature-extractors/discogs-effnet/discogs-effnet-bs64-1.pb
# Genre classifier (2 MB)
curl -L -O https://essentia.upf.edu/models/classification-heads/genre_discogs400/genre_discogs400-discogs-effnet-1.pb
curl -L -O https://essentia.upf.edu/models/classification-heads/genre_discogs400/genre_discogs400-discogs-effnet-1.json
# Mood classifier (2.7 MB)
curl -L -O https://essentia.upf.edu/models/classification-heads/mtg_jamendo_moodtheme/mtg_jamendo_moodtheme-discogs-effnet-1.pb
# Instrument classifier (2.6 MB)
curl -L -O https://essentia.upf.edu/models/classification-heads/mtg_jamendo_instrument/mtg_jamendo_instrument-discogs-effnet-1.pb
# Alternative genre classifier (2.7 MB)
curl -L -O https://essentia.upf.edu/models/classification-heads/mtg_jamendo_genre/mtg_jamendo_genre-discogs-effnet-1.pb
```
## Source
Tous les modèles proviennent du projet Essentia : https://essentia.upf.edu/models/
## Licence
Ces modèles sont fournis par le Music Technology Group de l'Universitat Pompeu Fabra sous licence permissive pour usage académique et commercial.

Binary file not shown.

View File

@@ -0,0 +1,462 @@
{
"name": "Genre Discogs400",
"type": "Music genre classification",
"link": "https://essentia.upf.edu/models/classification-heads/genre_discogs400/genre_discogs400-discogs-effnet-1.pb",
"version": "1",
"description": "Prediction of 400 music styles in the from the Discogs taxonomy",
"author": "Pablo Alonso",
"email": "pablo.alonso@upf.edu",
"release_date": "2023-05-04",
"framework": "tensorflow",
"framework_version": "2.8.0",
"classes": [
"Blues---Boogie Woogie",
"Blues---Chicago Blues",
"Blues---Country Blues",
"Blues---Delta Blues",
"Blues---Electric Blues",
"Blues---Harmonica Blues",
"Blues---Jump Blues",
"Blues---Louisiana Blues",
"Blues---Modern Electric Blues",
"Blues---Piano Blues",
"Blues---Rhythm & Blues",
"Blues---Texas Blues",
"Brass & Military---Brass Band",
"Brass & Military---Marches",
"Brass & Military---Military",
"Children's---Educational",
"Children's---Nursery Rhymes",
"Children's---Story",
"Classical---Baroque",
"Classical---Choral",
"Classical---Classical",
"Classical---Contemporary",
"Classical---Impressionist",
"Classical---Medieval",
"Classical---Modern",
"Classical---Neo-Classical",
"Classical---Neo-Romantic",
"Classical---Opera",
"Classical---Post-Modern",
"Classical---Renaissance",
"Classical---Romantic",
"Electronic---Abstract",
"Electronic---Acid",
"Electronic---Acid House",
"Electronic---Acid Jazz",
"Electronic---Ambient",
"Electronic---Bassline",
"Electronic---Beatdown",
"Electronic---Berlin-School",
"Electronic---Big Beat",
"Electronic---Bleep",
"Electronic---Breakbeat",
"Electronic---Breakcore",
"Electronic---Breaks",
"Electronic---Broken Beat",
"Electronic---Chillwave",
"Electronic---Chiptune",
"Electronic---Dance-pop",
"Electronic---Dark Ambient",
"Electronic---Darkwave",
"Electronic---Deep House",
"Electronic---Deep Techno",
"Electronic---Disco",
"Electronic---Disco Polo",
"Electronic---Donk",
"Electronic---Downtempo",
"Electronic---Drone",
"Electronic---Drum n Bass",
"Electronic---Dub",
"Electronic---Dub Techno",
"Electronic---Dubstep",
"Electronic---Dungeon Synth",
"Electronic---EBM",
"Electronic---Electro",
"Electronic---Electro House",
"Electronic---Electroclash",
"Electronic---Euro House",
"Electronic---Euro-Disco",
"Electronic---Eurobeat",
"Electronic---Eurodance",
"Electronic---Experimental",
"Electronic---Freestyle",
"Electronic---Future Jazz",
"Electronic---Gabber",
"Electronic---Garage House",
"Electronic---Ghetto",
"Electronic---Ghetto House",
"Electronic---Glitch",
"Electronic---Goa Trance",
"Electronic---Grime",
"Electronic---Halftime",
"Electronic---Hands Up",
"Electronic---Happy Hardcore",
"Electronic---Hard House",
"Electronic---Hard Techno",
"Electronic---Hard Trance",
"Electronic---Hardcore",
"Electronic---Hardstyle",
"Electronic---Hi NRG",
"Electronic---Hip Hop",
"Electronic---Hip-House",
"Electronic---House",
"Electronic---IDM",
"Electronic---Illbient",
"Electronic---Industrial",
"Electronic---Italo House",
"Electronic---Italo-Disco",
"Electronic---Italodance",
"Electronic---Jazzdance",
"Electronic---Juke",
"Electronic---Jumpstyle",
"Electronic---Jungle",
"Electronic---Latin",
"Electronic---Leftfield",
"Electronic---Makina",
"Electronic---Minimal",
"Electronic---Minimal Techno",
"Electronic---Modern Classical",
"Electronic---Musique Concr\u00e8te",
"Electronic---Neofolk",
"Electronic---New Age",
"Electronic---New Beat",
"Electronic---New Wave",
"Electronic---Noise",
"Electronic---Nu-Disco",
"Electronic---Power Electronics",
"Electronic---Progressive Breaks",
"Electronic---Progressive House",
"Electronic---Progressive Trance",
"Electronic---Psy-Trance",
"Electronic---Rhythmic Noise",
"Electronic---Schranz",
"Electronic---Sound Collage",
"Electronic---Speed Garage",
"Electronic---Speedcore",
"Electronic---Synth-pop",
"Electronic---Synthwave",
"Electronic---Tech House",
"Electronic---Tech Trance",
"Electronic---Techno",
"Electronic---Trance",
"Electronic---Tribal",
"Electronic---Tribal House",
"Electronic---Trip Hop",
"Electronic---Tropical House",
"Electronic---UK Garage",
"Electronic---Vaporwave",
"Folk, World, & Country---African",
"Folk, World, & Country---Bluegrass",
"Folk, World, & Country---Cajun",
"Folk, World, & Country---Canzone Napoletana",
"Folk, World, & Country---Catalan Music",
"Folk, World, & Country---Celtic",
"Folk, World, & Country---Country",
"Folk, World, & Country---Fado",
"Folk, World, & Country---Flamenco",
"Folk, World, & Country---Folk",
"Folk, World, & Country---Gospel",
"Folk, World, & Country---Highlife",
"Folk, World, & Country---Hillbilly",
"Folk, World, & Country---Hindustani",
"Folk, World, & Country---Honky Tonk",
"Folk, World, & Country---Indian Classical",
"Folk, World, & Country---La\u00efk\u00f3",
"Folk, World, & Country---Nordic",
"Folk, World, & Country---Pacific",
"Folk, World, & Country---Polka",
"Folk, World, & Country---Ra\u00ef",
"Folk, World, & Country---Romani",
"Folk, World, & Country---Soukous",
"Folk, World, & Country---S\u00e9ga",
"Folk, World, & Country---Volksmusik",
"Folk, World, & Country---Zouk",
"Folk, World, & Country---\u00c9ntekhno",
"Funk / Soul---Afrobeat",
"Funk / Soul---Boogie",
"Funk / Soul---Contemporary R&B",
"Funk / Soul---Disco",
"Funk / Soul---Free Funk",
"Funk / Soul---Funk",
"Funk / Soul---Gospel",
"Funk / Soul---Neo Soul",
"Funk / Soul---New Jack Swing",
"Funk / Soul---P.Funk",
"Funk / Soul---Psychedelic",
"Funk / Soul---Rhythm & Blues",
"Funk / Soul---Soul",
"Funk / Soul---Swingbeat",
"Funk / Soul---UK Street Soul",
"Hip Hop---Bass Music",
"Hip Hop---Boom Bap",
"Hip Hop---Bounce",
"Hip Hop---Britcore",
"Hip Hop---Cloud Rap",
"Hip Hop---Conscious",
"Hip Hop---Crunk",
"Hip Hop---Cut-up/DJ",
"Hip Hop---DJ Battle Tool",
"Hip Hop---Electro",
"Hip Hop---G-Funk",
"Hip Hop---Gangsta",
"Hip Hop---Grime",
"Hip Hop---Hardcore Hip-Hop",
"Hip Hop---Horrorcore",
"Hip Hop---Instrumental",
"Hip Hop---Jazzy Hip-Hop",
"Hip Hop---Miami Bass",
"Hip Hop---Pop Rap",
"Hip Hop---Ragga HipHop",
"Hip Hop---RnB/Swing",
"Hip Hop---Screw",
"Hip Hop---Thug Rap",
"Hip Hop---Trap",
"Hip Hop---Trip Hop",
"Hip Hop---Turntablism",
"Jazz---Afro-Cuban Jazz",
"Jazz---Afrobeat",
"Jazz---Avant-garde Jazz",
"Jazz---Big Band",
"Jazz---Bop",
"Jazz---Bossa Nova",
"Jazz---Contemporary Jazz",
"Jazz---Cool Jazz",
"Jazz---Dixieland",
"Jazz---Easy Listening",
"Jazz---Free Improvisation",
"Jazz---Free Jazz",
"Jazz---Fusion",
"Jazz---Gypsy Jazz",
"Jazz---Hard Bop",
"Jazz---Jazz-Funk",
"Jazz---Jazz-Rock",
"Jazz---Latin Jazz",
"Jazz---Modal",
"Jazz---Post Bop",
"Jazz---Ragtime",
"Jazz---Smooth Jazz",
"Jazz---Soul-Jazz",
"Jazz---Space-Age",
"Jazz---Swing",
"Latin---Afro-Cuban",
"Latin---Bai\u00e3o",
"Latin---Batucada",
"Latin---Beguine",
"Latin---Bolero",
"Latin---Boogaloo",
"Latin---Bossanova",
"Latin---Cha-Cha",
"Latin---Charanga",
"Latin---Compas",
"Latin---Cubano",
"Latin---Cumbia",
"Latin---Descarga",
"Latin---Forr\u00f3",
"Latin---Guaguanc\u00f3",
"Latin---Guajira",
"Latin---Guaracha",
"Latin---MPB",
"Latin---Mambo",
"Latin---Mariachi",
"Latin---Merengue",
"Latin---Norte\u00f1o",
"Latin---Nueva Cancion",
"Latin---Pachanga",
"Latin---Porro",
"Latin---Ranchera",
"Latin---Reggaeton",
"Latin---Rumba",
"Latin---Salsa",
"Latin---Samba",
"Latin---Son",
"Latin---Son Montuno",
"Latin---Tango",
"Latin---Tejano",
"Latin---Vallenato",
"Non-Music---Audiobook",
"Non-Music---Comedy",
"Non-Music---Dialogue",
"Non-Music---Education",
"Non-Music---Field Recording",
"Non-Music---Interview",
"Non-Music---Monolog",
"Non-Music---Poetry",
"Non-Music---Political",
"Non-Music---Promotional",
"Non-Music---Radioplay",
"Non-Music---Religious",
"Non-Music---Spoken Word",
"Pop---Ballad",
"Pop---Bollywood",
"Pop---Bubblegum",
"Pop---Chanson",
"Pop---City Pop",
"Pop---Europop",
"Pop---Indie Pop",
"Pop---J-pop",
"Pop---K-pop",
"Pop---Kay\u014dkyoku",
"Pop---Light Music",
"Pop---Music Hall",
"Pop---Novelty",
"Pop---Parody",
"Pop---Schlager",
"Pop---Vocal",
"Reggae---Calypso",
"Reggae---Dancehall",
"Reggae---Dub",
"Reggae---Lovers Rock",
"Reggae---Ragga",
"Reggae---Reggae",
"Reggae---Reggae-Pop",
"Reggae---Rocksteady",
"Reggae---Roots Reggae",
"Reggae---Ska",
"Reggae---Soca",
"Rock---AOR",
"Rock---Acid Rock",
"Rock---Acoustic",
"Rock---Alternative Rock",
"Rock---Arena Rock",
"Rock---Art Rock",
"Rock---Atmospheric Black Metal",
"Rock---Avantgarde",
"Rock---Beat",
"Rock---Black Metal",
"Rock---Blues Rock",
"Rock---Brit Pop",
"Rock---Classic Rock",
"Rock---Coldwave",
"Rock---Country Rock",
"Rock---Crust",
"Rock---Death Metal",
"Rock---Deathcore",
"Rock---Deathrock",
"Rock---Depressive Black Metal",
"Rock---Doo Wop",
"Rock---Doom Metal",
"Rock---Dream Pop",
"Rock---Emo",
"Rock---Ethereal",
"Rock---Experimental",
"Rock---Folk Metal",
"Rock---Folk Rock",
"Rock---Funeral Doom Metal",
"Rock---Funk Metal",
"Rock---Garage Rock",
"Rock---Glam",
"Rock---Goregrind",
"Rock---Goth Rock",
"Rock---Gothic Metal",
"Rock---Grindcore",
"Rock---Grunge",
"Rock---Hard Rock",
"Rock---Hardcore",
"Rock---Heavy Metal",
"Rock---Indie Rock",
"Rock---Industrial",
"Rock---Krautrock",
"Rock---Lo-Fi",
"Rock---Lounge",
"Rock---Math Rock",
"Rock---Melodic Death Metal",
"Rock---Melodic Hardcore",
"Rock---Metalcore",
"Rock---Mod",
"Rock---Neofolk",
"Rock---New Wave",
"Rock---No Wave",
"Rock---Noise",
"Rock---Noisecore",
"Rock---Nu Metal",
"Rock---Oi",
"Rock---Parody",
"Rock---Pop Punk",
"Rock---Pop Rock",
"Rock---Pornogrind",
"Rock---Post Rock",
"Rock---Post-Hardcore",
"Rock---Post-Metal",
"Rock---Post-Punk",
"Rock---Power Metal",
"Rock---Power Pop",
"Rock---Power Violence",
"Rock---Prog Rock",
"Rock---Progressive Metal",
"Rock---Psychedelic Rock",
"Rock---Psychobilly",
"Rock---Pub Rock",
"Rock---Punk",
"Rock---Rock & Roll",
"Rock---Rockabilly",
"Rock---Shoegaze",
"Rock---Ska",
"Rock---Sludge Metal",
"Rock---Soft Rock",
"Rock---Southern Rock",
"Rock---Space Rock",
"Rock---Speed Metal",
"Rock---Stoner Rock",
"Rock---Surf",
"Rock---Symphonic Rock",
"Rock---Technical Death Metal",
"Rock---Thrash",
"Rock---Twist",
"Rock---Viking Metal",
"Rock---Y\u00e9-Y\u00e9",
"Stage & Screen---Musical",
"Stage & Screen---Score",
"Stage & Screen---Soundtrack",
"Stage & Screen---Theme"
],
"model_types": [
"frozen_model",
"SavedModel",
"onnx"
],
"dataset": {
"name": "Discogs-4M (unreleased)",
"citation": "In-house dataset",
"size": "4M full tracks (3.3M used)",
"metrics": {
"ROC-AUC": 0.95417,
"PR-AUC": 0.20629
}
},
"schema": {
"inputs": [
{
"name": "serving_default_model_Placeholder",
"type": "float",
"shape": [
"batch_size",
1280
]
}
],
"outputs": [
{
"name": "PartitionedCall:0",
"type": "float",
"shape": [
"batch_size",
400
],
"op": "Sigmoid",
"output_purpose": "predictions"
}
]
},
"citation": "@inproceedings{alonso2022music,\n title={Music Representation Learning Based on Editorial Metadata from Discogs},\n author={Alonso-Jim{\\'e}nez, Pablo and Serra, Xavier and Bogdanov, Dmitry},\n booktitle={Conference of the International Society for Music Information Retrieval (ISMIR)},\n year={2022}\n}",
"inference": {
"sample_rate": 16000,
"algorithm": "TensorflowPredict2D",
"embedding_model": {
"algorithm": "TensorflowPredictEffnetDiscogs",
"model_name": "discogs-effnet-bs64-1",
"link": "https://essentia.upf.edu/models/music-style-classification/discogs-effnet/discogs-effnet-bs64-1.pb"
}
}
}

Binary file not shown.

Binary file not shown.

View File

@@ -26,6 +26,11 @@ scipy==1.11.4
pydantic==2.5.3
pydantic-settings==2.1.0
python-dotenv==1.0.0
email-validator==2.1.0
# Authentication
python-jose[cryptography]==3.3.0
passlib[bcrypt]==1.7.4
# Utilities
aiofiles==23.2.1

View File

@@ -1,14 +1,15 @@
"""FastAPI main application."""
from fastapi import FastAPI
from fastapi import FastAPI, Depends
from fastapi.middleware.cors import CORSMiddleware
from contextlib import asynccontextmanager
from ..utils.config import settings
from ..utils.logging import setup_logging, get_logger
from ..models.database import engine, Base
from ..core.auth import require_auth
# Import routes
from .routes import tracks, search, audio, analyze, similar, stats, library
from .routes import tracks, search, audio, analyze, similar, stats, library, auth
# Setup logging
setup_logging()
@@ -62,13 +63,18 @@ async def health_check():
# Include routers
app.include_router(tracks.router, prefix="/api/tracks", tags=["tracks"])
app.include_router(search.router, prefix="/api/search", tags=["search"])
# Auth endpoints (public - no auth required)
app.include_router(auth.router, prefix="/api/auth", tags=["auth"])
# Protected endpoints (auth required for ALL routes)
app.include_router(tracks.router, prefix="/api/tracks", tags=["tracks"], dependencies=[Depends(require_auth)])
app.include_router(search.router, prefix="/api/search", tags=["search"], dependencies=[Depends(require_auth)])
# Audio endpoints handle auth internally (support both header and query param)
app.include_router(audio.router, prefix="/api/audio", tags=["audio"])
app.include_router(analyze.router, prefix="/api/analyze", tags=["analyze"])
app.include_router(similar.router, prefix="/api", tags=["similar"])
app.include_router(stats.router, prefix="/api/stats", tags=["stats"])
app.include_router(library.router, prefix="/api/library", tags=["library"])
app.include_router(analyze.router, prefix="/api/analyze", tags=["analyze"], dependencies=[Depends(require_auth)])
app.include_router(similar.router, prefix="/api", tags=["similar"], dependencies=[Depends(require_auth)])
app.include_router(stats.router, prefix="/api/stats", tags=["stats"], dependencies=[Depends(require_auth)])
app.include_router(library.router, prefix="/api/library", tags=["library"], dependencies=[Depends(require_auth)])
@app.get("/", tags=["root"])

View File

@@ -1,13 +1,15 @@
"""Audio streaming and download endpoints."""
from fastapi import APIRouter, Depends, HTTPException, Request
from fastapi import APIRouter, Depends, HTTPException, Request, Query, status
from fastapi.responses import FileResponse
from sqlalchemy.orm import Session
from uuid import UUID
from pathlib import Path
from typing import Optional
from ...models.database import get_db
from ...models import crud
from ...core.waveform_generator import get_waveform_data
from ...core.auth import verify_token, require_auth
from ...utils.logging import get_logger
router = APIRouter()
@@ -18,6 +20,7 @@ logger = get_logger(__name__)
async def stream_audio(
track_id: UUID,
request: Request,
token: Optional[str] = Query(None),
db: Session = Depends(get_db),
):
"""Stream audio file with range request support.
@@ -28,6 +31,7 @@ async def stream_audio(
Args:
track_id: Track UUID
request: HTTP request
token: Optional JWT token for authentication (for <audio> tag compatibility)
db: Database session
Returns:
@@ -36,6 +40,15 @@ async def stream_audio(
Raises:
HTTPException: 404 if track not found or file doesn't exist
"""
# Verify authentication via query parameter for <audio> tag
if not token:
raise HTTPException(
status_code=status.HTTP_401_UNAUTHORIZED,
detail="Authentication required",
headers={"WWW-Authenticate": "Bearer"},
)
verify_token(token)
track = crud.get_track_by_id(db, track_id)
if not track:
@@ -79,12 +92,14 @@ async def stream_audio(
@router.get("/download/{track_id}")
async def download_audio(
track_id: UUID,
token: Optional[str] = Query(None),
db: Session = Depends(get_db),
):
"""Download audio file.
Args:
track_id: Track UUID
token: Optional JWT token for authentication (for <a> tag compatibility)
db: Database session
Returns:
@@ -93,6 +108,15 @@ async def download_audio(
Raises:
HTTPException: 404 if track not found or file doesn't exist
"""
# Verify authentication via query parameter for <a> tag
if not token:
raise HTTPException(
status_code=status.HTTP_401_UNAUTHORIZED,
detail="Authentication required",
headers={"WWW-Authenticate": "Bearer"},
)
verify_token(token)
track = crud.get_track_by_id(db, track_id)
if not track:
@@ -129,6 +153,7 @@ async def get_waveform(
track_id: UUID,
num_peaks: int = 800,
db: Session = Depends(get_db),
current_user: dict = Depends(require_auth),
):
"""Get waveform peak data for visualization.
@@ -138,6 +163,7 @@ async def get_waveform(
track_id: Track UUID
num_peaks: Number of peaks to generate
db: Database session
current_user: Current authenticated user
Returns:
Waveform data with peaks and duration

View File

@@ -0,0 +1,82 @@
"""Authentication endpoints."""
from datetime import timedelta
from fastapi import APIRouter, HTTPException, status, Depends
from pydantic import BaseModel, EmailStr
from ...core.auth import authenticate_user, create_access_token, get_current_user
from ...utils.config import settings
from ...utils.logging import get_logger
router = APIRouter()
logger = get_logger(__name__)
class LoginRequest(BaseModel):
"""Login request model."""
email: EmailStr
password: str
class LoginResponse(BaseModel):
"""Login response model."""
access_token: str
token_type: str = "bearer"
user: dict
class UserResponse(BaseModel):
"""User response model."""
email: str
role: str
@router.post("/login", response_model=LoginResponse)
async def login(request: LoginRequest):
"""Authenticate user and return JWT token.
Args:
request: Login credentials
Returns:
Access token and user info
Raises:
HTTPException: 401 if credentials are invalid
"""
user = authenticate_user(request.email, request.password)
if not user:
logger.warning(f"Failed login attempt for: {request.email}")
raise HTTPException(
status_code=status.HTTP_401_UNAUTHORIZED,
detail="Incorrect email or password",
headers={"WWW-Authenticate": "Bearer"},
)
# Create access token
access_token_expires = timedelta(hours=settings.JWT_EXPIRATION_HOURS)
access_token = create_access_token(
data={"sub": user["email"], "role": user["role"]},
expires_delta=access_token_expires
)
logger.info(f"User logged in: {user['email']}")
return {
"access_token": access_token,
"token_type": "bearer",
"user": user
}
@router.get("/me", response_model=UserResponse)
async def get_me(current_user: dict = Depends(get_current_user)):
"""Get current authenticated user info.
Args:
current_user: Current user from JWT token
Returns:
User information
"""
return current_user

View File

@@ -41,6 +41,9 @@ def find_audio_files(directory: str) -> list[Path]:
return []
for root, dirs, files in os.walk(directory_path):
# Skip transcoded and waveforms directories
dirs[:] = [d for d in dirs if d not in ['transcoded', 'waveforms', '.transcoded', '.waveforms']]
for file in files:
file_path = Path(root) / file
if file_path.suffix.lower() in AUDIO_EXTENSIONS:

View File

@@ -24,6 +24,7 @@ async def get_tracks(
has_vocals: Optional[bool] = None,
key: Optional[str] = None,
instrument: Optional[str] = None,
instruments: Optional[List[str]] = Query(None),
tempo_range: Optional[str] = Query(None, regex="^(slow|medium|fast)$"),
sort_by: str = Query("analyzed_at", regex="^(analyzed_at|tempo_bpm|duration_seconds|filename|energy)$"),
sort_desc: bool = True,
@@ -42,7 +43,8 @@ async def get_tracks(
energy_max: Maximum energy
has_vocals: Filter by vocal presence
key: Filter by musical key
instrument: Filter by instrument
instrument: Filter by instrument (deprecated, use instruments)
instruments: Filter by multiple instruments (must have ALL)
tempo_range: Filter by tempo range (slow: <100, medium: 100-140, fast: >140)
sort_by: Field to sort by
sort_desc: Sort descending
@@ -61,6 +63,9 @@ async def get_tracks(
elif tempo_range == "fast":
bpm_min = 140.0 if bpm_min is None else max(bpm_min, 140.0)
# Use instruments if provided, otherwise fall back to instrument
final_instruments = instruments if instruments else ([instrument] if instrument else None)
tracks, total = crud.get_tracks(
db=db,
skip=skip,
@@ -73,7 +78,7 @@ async def get_tracks(
energy_max=energy_max,
has_vocals=has_vocals,
key=key,
instrument=instrument,
instruments=final_instruments,
sort_by=sort_by,
sort_desc=sort_desc,
)

View File

@@ -46,6 +46,9 @@ def find_audio_files(directory: str) -> List[Path]:
logger.info(f"Scanning directory: {directory}")
for root, dirs, files in os.walk(directory_path):
# Skip transcoded and waveforms directories
dirs[:] = [d for d in dirs if d not in ['transcoded', 'waveforms', '.transcoded', '.waveforms']]
for file in files:
file_path = Path(root) / file
if file_path.suffix.lower() in AUDIO_EXTENSIONS:

151
backend/src/core/auth.py Normal file
View File

@@ -0,0 +1,151 @@
"""Authentication utilities."""
from datetime import datetime, timedelta
from typing import Optional
from jose import JWTError, jwt
from passlib.context import CryptContext
from fastapi import HTTPException, status, Depends
from fastapi.security import HTTPBearer, HTTPAuthorizationCredentials
from ..utils.config import settings
from ..utils.logging import get_logger
logger = get_logger(__name__)
# Password hashing
pwd_context = CryptContext(schemes=["bcrypt"], deprecated="auto")
# HTTP Bearer for JWT
security = HTTPBearer()
def verify_password(plain_password: str, hashed_password: str) -> bool:
"""Verify a password against its hash.
Args:
plain_password: Plain text password
hashed_password: Hashed password
Returns:
True if password matches
"""
return pwd_context.verify(plain_password, hashed_password)
def get_password_hash(password: str) -> str:
"""Hash a password.
Args:
password: Plain text password
Returns:
Hashed password
"""
return pwd_context.hash(password)
def create_access_token(data: dict, expires_delta: Optional[timedelta] = None) -> str:
"""Create JWT access token.
Args:
data: Data to encode in token
expires_delta: Token expiration time
Returns:
JWT token string
"""
to_encode = data.copy()
if expires_delta:
expire = datetime.utcnow() + expires_delta
else:
expire = datetime.utcnow() + timedelta(hours=settings.JWT_EXPIRATION_HOURS)
to_encode.update({"exp": expire})
encoded_jwt = jwt.encode(to_encode, settings.JWT_SECRET_KEY, algorithm="HS256")
return encoded_jwt
def verify_token(token: str) -> dict:
"""Verify and decode JWT token.
Args:
token: JWT token string
Returns:
Decoded token payload
Raises:
HTTPException: If token is invalid
"""
try:
payload = jwt.decode(token, settings.JWT_SECRET_KEY, algorithms=["HS256"])
return payload
except JWTError as e:
logger.error(f"Token verification failed: {e}")
raise HTTPException(
status_code=status.HTTP_401_UNAUTHORIZED,
detail="Could not validate credentials",
headers={"WWW-Authenticate": "Bearer"},
)
def authenticate_user(email: str, password: str) -> Optional[dict]:
"""Authenticate user with email and password.
Args:
email: User email
password: User password
Returns:
User data if authenticated, None otherwise
"""
# Check against admin credentials from environment
if email == settings.ADMIN_EMAIL and password == settings.ADMIN_PASSWORD:
return {
"email": email,
"role": "admin"
}
return None
async def get_current_user(credentials: HTTPAuthorizationCredentials = Depends(security)) -> dict:
"""Get current authenticated user from JWT token.
Args:
credentials: HTTP Bearer credentials
Returns:
User data from token
Raises:
HTTPException: If authentication fails
"""
token = credentials.credentials
payload = verify_token(token)
email: str = payload.get("sub")
if email is None:
raise HTTPException(
status_code=status.HTTP_401_UNAUTHORIZED,
detail="Could not validate credentials",
headers={"WWW-Authenticate": "Bearer"},
)
return {
"email": email,
"role": payload.get("role", "user")
}
async def require_auth(current_user: dict = Depends(get_current_user)) -> dict:
"""Dependency to require authentication.
Args:
current_user: Current user from get_current_user
Returns:
Current user data
"""
return current_user

View File

@@ -104,7 +104,7 @@ def get_tracks(
energy_max: Optional[float] = None,
has_vocals: Optional[bool] = None,
key: Optional[str] = None,
instrument: Optional[str] = None,
instruments: Optional[List[str]] = None,
sort_by: str = "analyzed_at",
sort_desc: bool = True,
) -> Tuple[List[AudioTrack], int]:
@@ -122,7 +122,7 @@ def get_tracks(
energy_max: Maximum energy (0-1)
has_vocals: Filter by vocal presence
key: Filter by musical key
instrument: Filter by instrument
instruments: Filter by instruments (track must have ALL instruments in the list)
sort_by: Field to sort by
sort_desc: Sort descending if True
@@ -168,7 +168,9 @@ def get_tracks(
if key:
query = query.filter(AudioTrack.key == key)
if instrument:
if instruments:
# Track must have ALL specified instruments
for instrument in instruments:
query = query.filter(AudioTrack.instruments.any(instrument))
# Get total count before pagination

View File

@@ -10,7 +10,8 @@ class Settings(BaseSettings):
DATABASE_URL: str = "postgresql://audio_user:audio_password@localhost:5432/audio_classifier"
# API Configuration
CORS_ORIGINS: str = "http://localhost:3000,http://127.0.0.1:3000"
# Comma-separated list of allowed origins, or use "*" to allow all
CORS_ORIGINS: str = "*"
API_HOST: str = "0.0.0.0"
API_PORT: int = 8000
@@ -20,6 +21,12 @@ class Settings(BaseSettings):
ESSENTIA_MODELS_PATH: str = "./models"
AUDIO_LIBRARY_PATH: str = "/audio"
# Authentication
ADMIN_EMAIL: str = "admin@example.com"
ADMIN_PASSWORD: str = "changeme"
JWT_SECRET_KEY: str = "your-secret-key-change-in-production"
JWT_EXPIRATION_HOURS: int = 24
# Application
APP_NAME: str = "Audio Classifier API"
APP_VERSION: str = "1.0.0"
@@ -33,7 +40,13 @@ class Settings(BaseSettings):
@property
def cors_origins_list(self) -> List[str]:
"""Parse CORS origins string to list."""
"""Parse CORS origins string to list.
If CORS_ORIGINS is "*", allow all origins.
Otherwise, parse comma-separated list.
"""
if self.CORS_ORIGINS.strip() == "*":
return ["*"]
return [origin.strip() for origin in self.CORS_ORIGINS.split(",")]

64
docker-compose.build.yml Normal file
View File

@@ -0,0 +1,64 @@
# Docker Compose pour build local (développement)
# Usage: docker-compose -f docker-compose.build.yml build
services:
postgres:
image: pgvector/pgvector:pg16
container_name: audio_classifier_db
environment:
POSTGRES_USER: ${POSTGRES_USER:-audio_user}
POSTGRES_PASSWORD: ${POSTGRES_PASSWORD:-audio_password}
POSTGRES_DB: ${POSTGRES_DB:-audio_classifier}
ports:
- "5433:5432"
volumes:
- postgres_data:/var/lib/postgresql/data
- ./backend/init-db.sql:/docker-entrypoint-initdb.d/init-db.sql
healthcheck:
test: ["CMD-SHELL", "pg_isready -U ${POSTGRES_USER:-audio_user}"]
interval: 10s
timeout: 5s
retries: 5
restart: unless-stopped
backend:
build:
context: .
dockerfile: backend/Dockerfile
container_name: audio_classifier_api
depends_on:
postgres:
condition: service_healthy
environment:
DATABASE_URL: postgresql://${POSTGRES_USER:-audio_user}:${POSTGRES_PASSWORD:-audio_password}@postgres:5432/${POSTGRES_DB:-audio_classifier}
CORS_ORIGINS: ${CORS_ORIGINS:-*}
ANALYSIS_USE_CLAP: ${ANALYSIS_USE_CLAP:-false}
ANALYSIS_NUM_WORKERS: ${ANALYSIS_NUM_WORKERS:-4}
ESSENTIA_MODELS_PATH: /app/models
ports:
- "8001:8000"
volumes:
# Mount your audio library (read-write for transcoding and waveforms)
- ${AUDIO_LIBRARY_PATH:-./audio_samples}:/audio
restart: unless-stopped
frontend:
build:
context: .
dockerfile: frontend/Dockerfile
args:
NEXT_PUBLIC_API_URL: http://localhost:8001
container_name: audio_classifier_ui
environment:
# Use localhost:8001 because the browser (client-side) needs to access the API
# The backend is mapped to port 8001 on the host machine
NEXT_PUBLIC_API_URL: http://localhost:8001
ports:
- "3000:3000"
depends_on:
- backend
restart: unless-stopped
volumes:
postgres_data:
driver: local

View File

@@ -19,17 +19,22 @@ services:
restart: unless-stopped
backend:
build: ./backend
image: git.benoitsz.com/benoit/audio-classifier-backend:dev
container_name: audio_classifier_api
depends_on:
postgres:
condition: service_healthy
environment:
DATABASE_URL: postgresql://${POSTGRES_USER:-audio_user}:${POSTGRES_PASSWORD:-audio_password}@postgres:5432/${POSTGRES_DB:-audio_classifier}
CORS_ORIGINS: ${CORS_ORIGINS:-http://localhost:3000}
CORS_ORIGINS: ${CORS_ORIGINS:-*}
ANALYSIS_USE_CLAP: ${ANALYSIS_USE_CLAP:-false}
ANALYSIS_NUM_WORKERS: ${ANALYSIS_NUM_WORKERS:-4}
ESSENTIA_MODELS_PATH: /app/models
# Authentication
ADMIN_EMAIL: ${ADMIN_EMAIL:-admin@example.com}
ADMIN_PASSWORD: ${ADMIN_PASSWORD:-changeme}
JWT_SECRET_KEY: ${JWT_SECRET_KEY:-your-secret-key-change-in-production}
JWT_EXPIRATION_HOURS: ${JWT_EXPIRATION_HOURS:-24}
ports:
- "8001:8000"
volumes:
@@ -38,15 +43,10 @@ services:
restart: unless-stopped
frontend:
build:
context: ./frontend
args:
NEXT_PUBLIC_API_URL: http://localhost:8001
image: git.benoitsz.com/benoit/audio-classifier-frontend:dev
container_name: audio_classifier_ui
environment:
# Use localhost:8001 because the browser (client-side) needs to access the API
# The backend is mapped to port 8001 on the host machine
NEXT_PUBLIC_API_URL: http://localhost:8001
NEXT_PUBLIC_API_URL: https://api.audioclassifier.benoitsz.com
ports:
- "3000:3000"
depends_on:

View File

@@ -4,23 +4,27 @@ FROM node:20-alpine
WORKDIR /app
# Copy package files
COPY package*.json ./
COPY frontend/package*.json ./
# Install dependencies
RUN npm ci
# Copy application code
COPY . .
COPY frontend/ .
# Build argument for API URL
# Build argument for API URL (used for default build)
ARG NEXT_PUBLIC_API_URL=http://localhost:8001
ENV NEXT_PUBLIC_API_URL=${NEXT_PUBLIC_API_URL}
# Build the application
RUN npm run build
# Copy runtime config generation script
COPY frontend/generate-config.sh /app/generate-config.sh
RUN chmod +x /app/generate-config.sh
# Expose port
EXPOSE 3000
# Start the application
CMD ["npm", "start"]
# Generate runtime config and start the application
CMD ["/bin/sh", "-c", "/app/generate-config.sh && npm start"]

19
frontend/Dockerfile.dev Normal file
View File

@@ -0,0 +1,19 @@
FROM node:20-alpine
# Set working directory
WORKDIR /app
# Copy package files
COPY package*.json ./
# Debug: List files and Node.js version
RUN ls -la && node --version && npm --version
# Install dependencies with more verbose output
RUN npm install --verbose
# Expose port
EXPOSE 3000
# Start the development server
CMD ["npm", "run", "dev"]

93
frontend/README.md Normal file
View File

@@ -0,0 +1,93 @@
# Frontend - Audio Classifier
Frontend Next.js pour Audio Classifier avec configuration runtime.
## Configuration Runtime
Le frontend utilise un système de **configuration runtime** qui permet de changer l'URL de l'API sans rebuilder l'image Docker.
### Comment ça fonctionne
1. Au démarrage du container, le script `generate-config.sh` génère un fichier `/app/public/config.js`
2. Ce fichier contient l'URL de l'API basée sur la variable `NEXT_PUBLIC_API_URL`
3. Le fichier est chargé dans le navigateur via `<Script src="/config.js">`
4. Le code API lit la configuration depuis `window.__RUNTIME_CONFIG__.API_URL`
### Développement Local
```bash
# Installer les dépendances
npm install
# Créer un fichier .env.local
echo "NEXT_PUBLIC_API_URL=http://localhost:8001" > .env.local
# Lancer en mode dev
npm run dev
```
### Production avec Docker
```bash
# Build l'image
docker build -t audio-classifier-frontend -f frontend/Dockerfile .
# Lancer avec une URL personnalisée
docker run -p 3000:3000 \
-e NEXT_PUBLIC_API_URL=https://mon-serveur.com:8001 \
audio-classifier-frontend
```
### Docker Compose
```yaml
frontend:
image: audio-classifier-frontend
environment:
NEXT_PUBLIC_API_URL: ${NEXT_PUBLIC_API_URL:-http://localhost:8001}
ports:
- "3000:3000"
```
## Structure
```
frontend/
├── app/ # Pages Next.js (App Router)
│ ├── layout.tsx # Layout principal (charge config.js)
│ └── page.tsx # Page d'accueil
├── components/ # Composants React
├── lib/ # Utilitaires
│ ├── api.ts # Client API (lit la config runtime)
│ └── types.ts # Types TypeScript
├── public/ # Fichiers statiques
│ └── config.js # Configuration runtime (généré au démarrage)
├── generate-config.sh # Script de génération de config
└── Dockerfile # Image Docker de production
```
## Variables d'Environnement
- `NEXT_PUBLIC_API_URL` : URL de l'API backend (ex: `https://api.example.com:8001`)
## Troubleshooting
### L'API n'est pas accessible
Vérifiez que :
1. La variable `NEXT_PUBLIC_API_URL` est correctement définie
2. Le fichier `/app/public/config.js` existe dans le container
3. Le navigateur peut accéder à l'URL de l'API (pas de CORS, firewall, etc.)
### Voir la configuration active
Ouvrez la console du navigateur et tapez :
```javascript
console.log(window.__RUNTIME_CONFIG__)
```
### Vérifier la config dans le container
```bash
docker exec audio_classifier_ui cat /app/public/config.js
```

View File

@@ -2,6 +2,8 @@ import type { Metadata } from "next"
import { Inter } from "next/font/google"
import "./globals.css"
import { QueryProvider } from "@/components/providers/QueryProvider"
import AuthGuard from "@/components/AuthGuard"
import Script from "next/script"
const inter = Inter({ subsets: ["latin"] })
@@ -17,9 +19,14 @@ export default function RootLayout({
}) {
return (
<html lang="en">
<head>
<Script src="/config.js" strategy="beforeInteractive" />
</head>
<body className={inter.className}>
<QueryProvider>
<AuthGuard>
{children}
</AuthGuard>
</QueryProvider>
</body>
</html>

124
frontend/app/login/page.tsx Normal file
View File

@@ -0,0 +1,124 @@
"use client"
import { useState } from "react"
import { useRouter } from "next/navigation"
import { getApiUrl } from "@/lib/api"
export default function LoginPage() {
const router = useRouter()
const [email, setEmail] = useState("")
const [password, setPassword] = useState("")
const [error, setError] = useState("")
const [isLoading, setIsLoading] = useState(false)
const handleSubmit = async (e: React.FormEvent) => {
e.preventDefault()
setError("")
setIsLoading(true)
try {
const response = await fetch(`${getApiUrl()}/api/auth/login`, {
method: "POST",
headers: {
"Content-Type": "application/json",
},
body: JSON.stringify({ email, password }),
})
if (!response.ok) {
const data = await response.json()
throw new Error(data.detail || "Login failed")
}
const data = await response.json()
// Store token in localStorage
localStorage.setItem("access_token", data.access_token)
localStorage.setItem("user", JSON.stringify(data.user))
// Redirect to home
router.push("/")
} catch (err) {
setError(err instanceof Error ? err.message : "Login failed")
} finally {
setIsLoading(false)
}
}
return (
<div className="min-h-screen flex items-center justify-center bg-gradient-to-br from-gray-900 via-gray-800 to-gray-900">
<div className="max-w-md w-full mx-4">
<div className="bg-white rounded-lg shadow-2xl p-8">
{/* Logo/Title */}
<div className="text-center mb-8">
<h1 className="text-3xl font-bold text-gray-900 mb-2">
Audio Classifier
</h1>
<p className="text-gray-600">Sign in to continue</p>
</div>
{/* Error message */}
{error && (
<div className="mb-4 p-3 bg-red-50 border border-red-200 text-red-700 rounded-md text-sm">
{error}
</div>
)}
{/* Login form */}
<form onSubmit={handleSubmit} className="space-y-6">
<div>
<label
htmlFor="email"
className="block text-sm font-medium text-gray-700 mb-1"
>
Email
</label>
<input
id="email"
type="email"
required
value={email}
onChange={(e) => setEmail(e.target.value)}
className="w-full px-3 py-2 border border-gray-300 rounded-md focus:outline-none focus:ring-2 focus:ring-blue-500 focus:border-transparent"
placeholder="admin@example.com"
disabled={isLoading}
/>
</div>
<div>
<label
htmlFor="password"
className="block text-sm font-medium text-gray-700 mb-1"
>
Password
</label>
<input
id="password"
type="password"
required
value={password}
onChange={(e) => setPassword(e.target.value)}
className="w-full px-3 py-2 border border-gray-300 rounded-md focus:outline-none focus:ring-2 focus:ring-blue-500 focus:border-transparent"
placeholder="••••••••"
disabled={isLoading}
/>
</div>
<button
type="submit"
disabled={isLoading}
className="w-full bg-blue-600 hover:bg-blue-700 text-white font-medium py-2 px-4 rounded-md transition-colors disabled:bg-blue-400 disabled:cursor-not-allowed"
>
{isLoading ? "Signing in..." : "Sign in"}
</button>
</form>
</div>
{/* Footer */}
<p className="text-center text-gray-400 text-sm mt-6">
Audio Classifier v1.0.0
</p>
</div>
</div>
)
}

View File

@@ -2,7 +2,8 @@
import { useState, useMemo } from "react"
import { useQuery } from "@tanstack/react-query"
import { getTracks } from "@/lib/api"
import { getTracks, getApiUrl } from "@/lib/api"
import { logout, getUser } from "@/lib/auth"
import type { FilterParams, Track } from "@/lib/types"
import FilterPanel from "@/components/FilterPanel"
import AudioPlayer from "@/components/AudioPlayer"
@@ -52,6 +53,7 @@ export default function Home() {
const [filters, setFilters] = useState<FilterParams>({})
const [page, setPage] = useState(0)
const [currentTrack, setCurrentTrack] = useState<Track | null>(null)
const [isPlaying, setIsPlaying] = useState(false)
const [searchQuery, setSearchQuery] = useState("")
const [isScanning, setIsScanning] = useState(false)
const [scanStatus, setScanStatus] = useState<string>("")
@@ -89,8 +91,17 @@ export default function Home() {
setIsScanning(true)
setScanStatus("Démarrage du scan...")
const response = await fetch(`${process.env.NEXT_PUBLIC_API_URL}/api/library/scan`, {
const token = localStorage.getItem('access_token')
const headers: HeadersInit = {
'Content-Type': 'application/json',
}
if (token) {
headers['Authorization'] = `Bearer ${token}`
}
const response = await fetch(`${getApiUrl()}/api/library/scan`, {
method: 'POST',
headers,
})
if (!response.ok) {
@@ -102,7 +113,15 @@ export default function Home() {
// Poll scan status
const pollInterval = setInterval(async () => {
try {
const statusResponse = await fetch(`${process.env.NEXT_PUBLIC_API_URL}/api/library/scan/status`)
const token = localStorage.getItem('access_token')
const pollHeaders: HeadersInit = {}
if (token) {
pollHeaders['Authorization'] = `Bearer ${token}`
}
const statusResponse = await fetch(`${getApiUrl()}/api/library/scan/status`, {
headers: pollHeaders,
})
const status = await statusResponse.json()
if (!status.is_scanning) {
@@ -159,6 +178,18 @@ export default function Home() {
{tracksData?.total || 0} piste{(tracksData?.total || 0) > 1 ? 's' : ''}
</div>
{/* Logout button */}
<button
onClick={logout}
className="px-3 py-2 text-sm text-slate-600 hover:text-slate-900 hover:bg-slate-100 rounded-lg transition-colors flex items-center gap-2"
title="Déconnexion"
>
<svg className="w-4 h-4" fill="none" stroke="currentColor" viewBox="0 0 24 24">
<path strokeLinecap="round" strokeLinejoin="round" strokeWidth={2} d="M17 16l4-4m0 0l-4-4m4 4H7m6 4v1a3 3 0 01-3 3H6a3 3 0 01-3-3V7a3 3 0 013-3h4a3 3 0 013 3v1" />
</svg>
Logout
</button>
{/* Rescan button */}
<button
onClick={handleRescan}
@@ -233,10 +264,19 @@ export default function Home() {
<div className="flex items-center gap-4">
{/* Play button */}
<button
onClick={() => setCurrentTrack(track)}
onClick={() => {
if (currentTrack?.id === track.id) {
// Toggle play/pause for current track
setIsPlaying(!isPlaying)
} else {
// Switch to new track and start playing
setCurrentTrack(track)
setIsPlaying(true)
}
}}
className="flex-shrink-0 w-12 h-12 flex items-center justify-center bg-orange-500 hover:bg-orange-600 rounded-full transition-colors shadow-sm"
>
{currentTrack?.id === track.id ? (
{currentTrack?.id === track.id && isPlaying ? (
<svg className="w-5 h-5 text-white" fill="currentColor" viewBox="0 0 24 24">
<path d="M6 4h4v16H6V4zm8 0h4v16h-4V4z"/>
</svg>
@@ -347,7 +387,11 @@ export default function Home() {
{/* Fixed Audio Player at bottom */}
<div className="fixed bottom-0 left-0 right-0 z-50">
<AudioPlayer track={currentTrack} />
<AudioPlayer
track={currentTrack}
isPlaying={isPlaying}
onPlayingChange={setIsPlaying}
/>
</div>
</div>
)

View File

@@ -2,13 +2,15 @@
import { useState, useRef, useEffect } from "react"
import type { Track } from "@/lib/types"
import { getApiUrl } from "@/lib/api"
interface AudioPlayerProps {
track: Track | null
isPlaying: boolean
onPlayingChange: (playing: boolean) => void
}
export default function AudioPlayer({ track }: AudioPlayerProps) {
const [isPlaying, setIsPlaying] = useState(false)
export default function AudioPlayer({ track, isPlaying, onPlayingChange }: AudioPlayerProps) {
const [currentTime, setCurrentTime] = useState(0)
const [duration, setDuration] = useState(0)
const [volume, setVolume] = useState(1)
@@ -22,7 +24,7 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
// Load audio and waveform when track changes
useEffect(() => {
if (!track) {
setIsPlaying(false)
onPlayingChange(false)
setCurrentTime(0)
setWaveformPeaks([])
return
@@ -33,14 +35,14 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
if (audioRef.current) {
audioRef.current.load()
// Autoplay when track loads
audioRef.current.play().then(() => {
setIsPlaying(true)
}).catch((error: unknown) => {
// Autoplay when track loads if isPlaying is true
if (isPlaying) {
audioRef.current.play().catch((error: unknown) => {
console.error("Autoplay failed:", error)
setIsPlaying(false)
onPlayingChange(false)
})
}
}
}, [track?.id])
// Update current time as audio plays
@@ -54,7 +56,7 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
setDuration(audio.duration)
}
}
const handleEnded = () => setIsPlaying(false)
const handleEnded = () => onPlayingChange(false)
audio.addEventListener("timeupdate", updateTime)
audio.addEventListener("loadedmetadata", updateDuration)
@@ -77,8 +79,15 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
const loadWaveform = async (trackId: string) => {
setIsLoadingWaveform(true)
try {
const token = localStorage.getItem('access_token')
const headers: HeadersInit = {}
if (token) {
headers['Authorization'] = `Bearer ${token}`
}
const response = await fetch(
`${process.env.NEXT_PUBLIC_API_URL}/api/audio/waveform/${trackId}`
`${getApiUrl()}/api/audio/waveform/${trackId}`,
{ headers }
)
if (response.ok) {
const data = await response.json()
@@ -91,15 +100,24 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
}
}
const togglePlay = () => {
if (!audioRef.current || !track) return
// Sync playing state with audio element
useEffect(() => {
const audio = audioRef.current
if (!audio) return
if (isPlaying) {
audioRef.current.pause()
audio.play().catch((error: unknown) => {
console.error("Play failed:", error)
onPlayingChange(false)
})
} else {
audioRef.current.play()
audio.pause()
}
setIsPlaying(!isPlaying)
}, [isPlaying, onPlayingChange])
const togglePlay = () => {
if (!audioRef.current || !track) return
onPlayingChange(!isPlaying)
}
const handleVolumeChange = (e: React.ChangeEvent<HTMLInputElement>) => {
@@ -148,10 +166,19 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
const progress = duration > 0 ? (currentTime / duration) * 100 : 0
const getAuthenticatedStreamUrl = (trackId: string) => {
const token = localStorage.getItem('access_token')
const baseUrl = `${getApiUrl()}/api/audio/stream/${trackId}`
if (token) {
return `${baseUrl}?token=${encodeURIComponent(token)}`
}
return baseUrl
}
return (
<div className="bg-gray-50 border-t border-gray-300 shadow-lg" style={{ height: '80px' }}>
{/* Hidden audio element */}
{track && <audio ref={audioRef} src={`${process.env.NEXT_PUBLIC_API_URL}/api/audio/stream/${track.id}`} />}
{track && <audio ref={audioRef} src={getAuthenticatedStreamUrl(track.id)} />}
<div className="h-full flex items-center gap-3 px-4">
{/* Play/Pause button */}
@@ -290,7 +317,7 @@ export default function AudioPlayer({ track }: AudioPlayerProps) {
{/* Download button */}
{track && (
<a
href={`${process.env.NEXT_PUBLIC_API_URL}/api/audio/download/${track.id}`}
href={getAuthenticatedStreamUrl(track.id).replace('/stream/', '/download/')}
download
className="w-8 h-8 flex items-center justify-center text-gray-600 hover:text-gray-900 transition-colors rounded hover:bg-gray-200 flex-shrink-0"
aria-label="Download"

View File

@@ -0,0 +1,37 @@
"use client"
import { useEffect, useState } from "react"
import { useRouter, usePathname } from "next/navigation"
import { isAuthenticated } from "@/lib/auth"
export default function AuthGuard({ children }: { children: React.ReactNode }) {
const router = useRouter()
const pathname = usePathname()
const [isChecking, setIsChecking] = useState(true)
useEffect(() => {
// Skip auth check for login page
if (pathname === "/login") {
setIsChecking(false)
return
}
// Check if user is authenticated
if (!isAuthenticated()) {
router.push("/login")
} else {
setIsChecking(false)
}
}, [pathname, router])
// Show loading while checking auth
if (isChecking && pathname !== "/login") {
return (
<div className="min-h-screen flex items-center justify-center bg-gray-900">
<div className="text-white">Loading...</div>
</div>
)
}
return <>{children}</>
}

View File

@@ -93,23 +93,40 @@ export default function FilterPanel({
</select>
</div>
{/* Instrument Filter */}
{/* Instrument Filter - Multiple Selection */}
<div>
<label className="block text-sm font-semibold text-slate-700 mb-2">
Instrument
Instruments
</label>
<select
value={localFilters.instrument || ""}
onChange={(e) => handleFilterChange("instrument", e.target.value || undefined)}
className="w-full px-3 py-2 bg-slate-50 border border-slate-300 rounded-lg text-sm focus:outline-none focus:ring-2 focus:ring-orange-500 focus:border-transparent"
<div className="space-y-2 max-h-48 overflow-y-auto bg-slate-50 border border-slate-300 rounded-lg p-2">
{availableInstruments.length === 0 ? (
<p className="text-xs text-slate-500 p-2">Aucun instrument disponible</p>
) : (
availableInstruments.map((instrument) => {
const isSelected = localFilters.instruments?.includes(instrument) || false
return (
<label
key={instrument}
className="flex items-center gap-2 p-2 hover:bg-slate-100 rounded cursor-pointer transition-colors"
>
<option value="">Tous les instruments</option>
{availableInstruments.map((instrument) => (
<option key={instrument} value={instrument}>
{instrument}
</option>
))}
</select>
<input
type="checkbox"
checked={isSelected}
onChange={(e) => {
const currentInstruments = localFilters.instruments || []
const newInstruments = e.target.checked
? [...currentInstruments, instrument]
: currentInstruments.filter(i => i !== instrument)
handleFilterChange("instruments", newInstruments.length > 0 ? newInstruments : undefined)
}}
className="w-4 h-4 text-orange-500 border-slate-300 rounded focus:ring-2 focus:ring-orange-500 focus:ring-offset-0"
/>
<span className="text-sm text-slate-700">{instrument}</span>
</label>
)
})
)}
</div>
</div>
{/* Key Filter */}
@@ -165,10 +182,16 @@ export default function FilterPanel({
<span className="font-medium text-slate-800">{localFilters.mood}</span>
</div>
)}
{localFilters.instrument && (
<div className="flex items-center justify-between text-xs">
<span className="text-slate-600">Instrument:</span>
<span className="font-medium text-slate-800">{localFilters.instrument}</span>
{localFilters.instruments && localFilters.instruments.length > 0 && (
<div className="text-xs">
<span className="text-slate-600">Instruments:</span>
<div className="flex flex-wrap gap-1 mt-1">
{localFilters.instruments.map((instrument) => (
<span key={instrument} className="inline-flex items-center px-2 py-0.5 rounded text-xs bg-emerald-50 text-emerald-700">
{instrument}
</span>
))}
</div>
</div>
)}
{localFilters.key && (

View File

@@ -0,0 +1,15 @@
#!/bin/sh
# Generate runtime configuration file
echo "Generating runtime configuration..."
echo "API URL: ${NEXT_PUBLIC_API_URL:-http://localhost:8001}"
cat > /app/public/config.js << EOF
// Runtime configuration generated at container startup
window.__RUNTIME_CONFIG__ = {
API_URL: '${NEXT_PUBLIC_API_URL:-http://localhost:8001}'
};
EOF
echo "Configuration generated successfully!"
cat /app/public/config.js

View File

@@ -14,28 +14,63 @@ import type {
FilterParams,
} from './types'
const API_BASE_URL = process.env.NEXT_PUBLIC_API_URL || 'http://localhost:8000'
// Get API URL from runtime config (injected at container startup) or fallback to env var
export function getApiUrl(): string {
if (typeof window !== 'undefined' && (window as any).__RUNTIME_CONFIG__) {
return (window as any).__RUNTIME_CONFIG__.API_URL
}
return process.env.NEXT_PUBLIC_API_URL || 'http://localhost:8000'
}
const apiClient = axios.create({
baseURL: API_BASE_URL,
// Create axios instance dynamically to use runtime config
function getApiClient() {
const client = axios.create({
baseURL: getApiUrl(),
headers: {
'Content-Type': 'application/json',
},
})
// Add JWT token to requests if available
client.interceptors.request.use((config) => {
if (typeof window !== 'undefined') {
const token = localStorage.getItem('access_token')
if (token) {
config.headers.Authorization = `Bearer ${token}`
}
}
return config
})
// Handle 401 errors (redirect to login)
client.interceptors.response.use(
(response) => response,
(error) => {
if (error.response?.status === 401 && typeof window !== 'undefined') {
localStorage.removeItem('access_token')
localStorage.removeItem('user')
window.location.href = '/login'
}
return Promise.reject(error)
}
)
return client
}
// Tracks
export async function getTracks(params: FilterParams & { skip?: number; limit?: number }): Promise<TracksResponse> {
const response = await apiClient.get('/api/tracks', { params })
const response = await getApiClient().get('/api/tracks', { params })
return response.data
}
export async function getTrack(id: string): Promise<Track> {
const response = await apiClient.get(`/api/tracks/${id}`)
const response = await getApiClient().get(`/api/tracks/${id}`)
return response.data
}
export async function deleteTrack(id: string): Promise<void> {
await apiClient.delete(`/api/tracks/${id}`)
await getApiClient().delete(`/api/tracks/${id}`)
}
// Search
@@ -43,7 +78,7 @@ export async function searchTracks(
query: string,
filters?: { genre?: string; mood?: string; limit?: number }
): Promise<SearchResponse> {
const response = await apiClient.get('/api/search', {
const response = await getApiClient().get('/api/search', {
params: { q: query, ...filters },
})
return response.data
@@ -51,7 +86,7 @@ export async function searchTracks(
// Similar
export async function getSimilarTracks(id: string, limit: number = 10): Promise<SimilarTracksResponse> {
const response = await apiClient.get(`/api/tracks/${id}/similar`, {
const response = await getApiClient().get(`/api/tracks/${id}/similar`, {
params: { limit },
})
return response.data
@@ -59,30 +94,30 @@ export async function getSimilarTracks(id: string, limit: number = 10): Promise<
// Analysis
export async function analyzeFolder(request: AnalyzeFolderRequest): Promise<{ job_id: string }> {
const response = await apiClient.post('/api/analyze/folder', request)
const response = await getApiClient().post('/api/analyze/folder', request)
return response.data
}
export async function getAnalyzeStatus(jobId: string): Promise<JobStatus> {
const response = await apiClient.get(`/api/analyze/status/${jobId}`)
const response = await getApiClient().get(`/api/analyze/status/${jobId}`)
return response.data
}
export async function deleteJob(jobId: string): Promise<void> {
await apiClient.delete(`/api/analyze/job/${jobId}`)
await getApiClient().delete(`/api/analyze/job/${jobId}`)
}
// Audio
export function getStreamUrl(trackId: string): string {
return `${API_BASE_URL}/api/audio/stream/${trackId}`
return `${getApiUrl()}/api/audio/stream/${trackId}`
}
export function getDownloadUrl(trackId: string): string {
return `${API_BASE_URL}/api/audio/download/${trackId}`
return `${getApiUrl()}/api/audio/download/${trackId}`
}
export async function getWaveform(trackId: string, numPeaks: number = 800): Promise<WaveformData> {
const response = await apiClient.get(`/api/audio/waveform/${trackId}`, {
const response = await getApiClient().get(`/api/audio/waveform/${trackId}`, {
params: { num_peaks: numPeaks },
})
return response.data
@@ -90,14 +125,12 @@ export async function getWaveform(trackId: string, numPeaks: number = 800): Prom
// Stats
export async function getStats(): Promise<Stats> {
const response = await apiClient.get('/api/stats')
const response = await getApiClient().get('/api/stats')
return response.data
}
// Health
export async function healthCheck(): Promise<{ status: string }> {
const response = await apiClient.get('/health')
const response = await getApiClient().get('/health')
return response.data
}
export default apiClient

34
frontend/lib/auth.ts Normal file
View File

@@ -0,0 +1,34 @@
/**
* Authentication utilities
*/
export function getToken(): string | null {
if (typeof window === "undefined") return null
return localStorage.getItem("access_token")
}
export function setToken(token: string): void {
localStorage.setItem("access_token", token)
}
export function removeToken(): void {
localStorage.removeItem("access_token")
localStorage.removeItem("user")
}
export function getUser(): any | null {
if (typeof window === "undefined") return null
const user = localStorage.getItem("user")
return user ? JSON.parse(user) : null
}
export function isAuthenticated(): boolean {
return getToken() !== null
}
export function logout(): void {
removeToken()
if (typeof window !== "undefined") {
window.location.href = "/login"
}
}

View File

@@ -60,6 +60,7 @@ export interface FilterParams {
has_vocals?: boolean
key?: string
instrument?: string
instruments?: string[] // Multiple instruments filter
tempo_range?: 'slow' | 'medium' | 'fast' // Lent (<100), Moyen (100-140), Rapide (>140)
sort_by?: 'analyzed_at' | 'tempo_bpm' | 'duration_seconds' | 'filename' | 'energy'
sort_desc?: boolean

20
frontend/middleware.ts Normal file
View File

@@ -0,0 +1,20 @@
import { NextResponse } from 'next/server'
import type { NextRequest } from 'next/server'
export function middleware(request: NextRequest) {
// Middleware runs on server, can't access localStorage
// Auth check will be done client-side in layout.tsx
return NextResponse.next()
}
export const config = {
matcher: [
/*
* Match all request paths except for the ones starting with:
* - _next/static (static files)
* - _next/image (image optimization files)
* - favicon.ico (favicon file)
*/
'/((?!_next/static|_next/image|favicon.ico).*)',
],
}

View File

@@ -0,0 +1,4 @@
// This file will be overwritten at container startup
window.__RUNTIME_CONFIG__ = {
API_URL: 'http://localhost:8001'
};