diff --git a/CLAUDE.md b/CLAUDE.md
new file mode 100644
index 0000000..e7be5d2
--- /dev/null
+++ b/CLAUDE.md
@@ -0,0 +1,63 @@
+# CLAUDE.md
+
+This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
+
+## Commands
+
+```bash
+# Build
+go build -o importer .
+
+# Build with version baked in
+go build -ldflags="-X main.version=v1.0.0" -o importer .
+
+# Run locally (requires IMPORT_DIR and LIBRARY_DIR env vars)
+IMPORT_DIR=/path/to/import LIBRARY_DIR=/path/to/library ./importer
+
+# Build Docker image
+docker build -t music-importer .
+
+# Build Docker image with version
+docker build --build-arg VERSION=v1.0.0 -t music-importer .
+```
+
+There are no tests in this codebase.
+
+## Architecture
+
+This is a single-package Go web app (`package main`) that runs as a web server on port 8080. Users trigger an import via the web UI, which runs the import pipeline in a background goroutine.
+
+**Pipeline flow** (`importer.go: RunImporter`):
+1. **Cluster** — loose audio files at the top of `IMPORT_DIR` are grouped into subdirectories by album tag (`files.go: cluster`)
+2. For each album directory:
+ - **Clean tags** — removes COMMENT/DESCRIPTION tags via `metaflac` (`audio.go`)
+ - **Tag metadata** — tries `beets` first; falls back to reading existing file tags, then MusicBrainz API (`metadata.go: getAlbumMetadata`)
+ - **Lyrics** — fetches synced LRC lyrics from LRClib API; falls back to plain lyrics formatted as LRC (`lrc.go`)
+ - **ReplayGain** — runs `rsgain easy` on the directory (`audio.go`)
+ - **Cover art** — looks for existing image files, downloads from Cover Art Archive via MusicBrainz if missing, then embeds into tracks (`media.go`)
+ - **Move** — moves tracks, .lrc files, and cover image into `LIBRARY_DIR/{Artist}/[{Date}] {Album} [{Quality}]/` (`files.go: moveToLibrary`)
+
+**Key types** (`importer.go`):
+- `AlbumResult` — tracks per-step success/failure/skip for one album
+- `ImportSession` — holds all `AlbumResult`s for one run; stored in `lastSession` global
+- `MusicMetadata` — artist/album/title/date/quality used throughout the pipeline
+
+**Web layer** (`main.go`):
+- `GET /` — renders `index.html.tmpl` with the last session's results
+- `POST /run` — starts `RunImporter()` in a goroutine; prevents concurrent runs via `importerMu` mutex
+
+**External tool dependencies** (must be present in PATH at runtime):
+- `ffprobe` — reads audio tags and stream info
+- `beet` — metadata tagging via MusicBrainz (primary metadata source)
+- `rsgain` — ReplayGain calculation
+- `metaflac` — FLAC tag manipulation and cover embedding
+- `curl` — MusicBrainz API fallback queries
+
+**Environment variables**:
+- `IMPORT_DIR` — source directory scanned for albums
+- `LIBRARY_DIR` — destination library root
+- `COPYMODE=true` — copies files instead of moving (still destructive on the destination)
+- `SLSKD_URL` — base URL of the slskd instance (e.g. `http://localhost:5030`)
+- `SLSKD_API_KEY` — slskd API key (sent as `X-API-Key` header)
+
+**Releases**: Docker image `gabehf/music-importer` is built and pushed to Docker Hub via GitHub Actions on `v*` tags.
diff --git a/README.md b/README.md
index c203fb1..52ac42a 100644
--- a/README.md
+++ b/README.md
@@ -1,5 +1,9 @@
# music-importer
+Goes through a folder with a bunch of loose .flac/.mp3 files, or with album folders containing music files, then
+fetches metadata with beets/musicbrainz, downloads lyrics via LRClib, embeds discovered cover art, and moves them
+into the library with the format {Artist}/[{Year}] {Title} [{Format-Quality}]
+
this thing is like 95% AI code. use at your own risk
i didn't feel like spending the time to do it all right and i figured its simple enough that chatgpt couldn't possible screw it up *that* bad
@@ -21,6 +25,7 @@ services:
environment:
IMPORT_DIR: /import
LIBRARY_DIR: /library
+ COPYMODE: true # copies files instead of moving. NOT NON-DESTRUCTIVE!!
```
diff --git a/discover.go b/discover.go
new file mode 100644
index 0000000..ddd38e6
--- /dev/null
+++ b/discover.go
@@ -0,0 +1,547 @@
+package main
+
+import (
+ "encoding/json"
+ "fmt"
+ "log"
+ "net/http"
+ "net/url"
+ "strings"
+ "sync"
+ "time"
+)
+
+// ── MusicBrainz types ─────────────────────────────────────────────────────────
+
+type mbArtistCredit struct {
+ Name string `json:"name"`
+ Artist struct {
+ ID string `json:"id"`
+ Name string `json:"name"`
+ } `json:"artist"`
+}
+
+type mbMedia struct {
+ Format string `json:"format"`
+ TrackCount int `json:"track-count"`
+}
+
+type mbRelease struct {
+ ID string `json:"id"`
+ Title string `json:"title"`
+ Date string `json:"date"`
+ Country string `json:"country"`
+ Disambiguation string `json:"disambiguation"`
+ TextRepresentation struct {
+ Language string `json:"language"`
+ } `json:"text-representation"`
+ Media []mbMedia `json:"media"`
+ ArtistCredit []mbArtistCredit `json:"artist-credit"`
+ ReleaseGroup struct {
+ PrimaryType string `json:"primary-type"`
+ } `json:"release-group"`
+}
+
+type mbArtist struct {
+ ID string `json:"id"`
+ Name string `json:"name"`
+ Country string `json:"country"`
+ Disambiguation string `json:"disambiguation"`
+}
+
+type mbReleaseGroup struct {
+ ID string `json:"id"`
+ Title string `json:"title"`
+ PrimaryType string `json:"primary-type"`
+ FirstReleaseDate string `json:"first-release-date"`
+}
+
+// releaseTrackCount returns the total number of tracks across all media in a release.
+func releaseTrackCount(r mbRelease) int {
+ total := 0
+ for _, m := range r.Media {
+ total += m.TrackCount
+ }
+ return total
+}
+
+// getMBRelease fetches a single release by MBID (with media/track-count included).
+func getMBRelease(mbid string) (*mbRelease, error) {
+ var r mbRelease
+ err := mbGet(fmt.Sprintf("/ws/2/release/%s?fmt=json&inc=media", url.QueryEscape(mbid)), &r)
+ return &r, err
+}
+
+func mbGet(path string, out interface{}) error {
+ req, err := http.NewRequest("GET", "https://musicbrainz.org"+path, nil)
+ if err != nil {
+ return err
+ }
+ req.Header.Set("User-Agent", "music-importer/1.0 (https://github.com/gabehf/music-importer)")
+
+ resp, err := http.DefaultClient.Do(req)
+ if err != nil {
+ return err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusOK {
+ return fmt.Errorf("MusicBrainz returned %d", resp.StatusCode)
+ }
+ return json.NewDecoder(resp.Body).Decode(out)
+}
+
+func searchMBReleases(query string) ([]mbRelease, error) {
+ var result struct {
+ Releases []mbRelease `json:"releases"`
+ }
+ err := mbGet("/ws/2/release/?query="+url.QueryEscape(query)+"&fmt=json&limit=20&inc=media", &result)
+ return result.Releases, err
+}
+
+func searchMBArtists(query string) ([]mbArtist, error) {
+ var result struct {
+ Artists []mbArtist `json:"artists"`
+ }
+ err := mbGet("/ws/2/artist/?query="+url.QueryEscape(query)+"&fmt=json&limit=20", &result)
+ return result.Artists, err
+}
+
+// releaseFormatScore returns a preference score for a release's media format.
+// Higher is better. CD=2, Digital Media=1, anything else=0.
+func releaseFormatScore(r mbRelease) int {
+ for _, m := range r.Media {
+ switch m.Format {
+ case "Digital Media":
+ return 2
+ case "CD":
+ return 1
+ }
+ }
+ return 0
+}
+
+// releaseCountryScore returns a preference score for a release's country.
+// Higher is better. KR=3, JP=2, XW=1, anything else=0.
+func releaseCountryScore(r mbRelease) int {
+ switch r.Country {
+ case "XW":
+ return 2
+ case "KR":
+ return 1
+ }
+ return 0
+}
+
+// returns true if strings formatted 'YYYY-MM-DD" ts1 is before ts2
+func timeStringIsBefore(ts1, ts2 string) (bool, error) {
+ datefmt := "2006-02-01"
+ t1, err := time.Parse(datefmt, ts1)
+ if err != nil {
+ return false, err
+ }
+ t2, err := time.Parse(datefmt, ts2)
+ if err != nil {
+ return false, err
+ }
+ return t1.Unix() <= t2.Unix(), nil
+}
+
+// pickBestRelease selects the preferred release from a list.
+// No disambiguation (canonical release) is the primary sort key;
+// format (CD > Digital Media > *) is secondary; country (KR > XW > *) breaks ties.
+func pickBestRelease(releases []mbRelease) *mbRelease {
+ if len(releases) == 0 {
+ return nil
+ }
+ best := &releases[0]
+ for i := 1; i < len(releases); i++ {
+ r := &releases[i]
+
+ rNoDisamb := r.Disambiguation == ""
+ bestNoDisamb := best.Disambiguation == ""
+
+ // Prefer releases with no disambiguation — they are the canonical default.
+ if rNoDisamb && !bestNoDisamb {
+ best = r
+ continue
+ }
+ if !rNoDisamb && bestNoDisamb {
+ continue
+ }
+
+ // Both have the same disambiguation status; use date/format/country.
+ if before, err := timeStringIsBefore(r.Date, best.Date); before && err == nil {
+ rf, bf := releaseFormatScore(*r), releaseFormatScore(*best)
+ if rf > bf || (rf == bf && releaseCountryScore(*r) > releaseCountryScore(*best)) {
+ best = r
+ }
+ }
+ }
+ return best
+}
+
+// pickBestReleaseForGroup fetches all releases for a release group via the
+// MusicBrainz browse API (with media info) and returns the preferred release.
+// Returns nil on error or when the group has no releases.
+func pickBestReleaseForGroup(rgMBID string) *mbRelease {
+ var result struct {
+ Releases []mbRelease `json:"releases"`
+ }
+ path := fmt.Sprintf("/ws/2/release?release-group=%s&fmt=json&inc=media&limit=100", url.QueryEscape(rgMBID))
+ if err := mbGet(path, &result); err != nil || len(result.Releases) == 0 {
+ return nil
+ }
+ return pickBestRelease(result.Releases)
+}
+
+// getMBArtistReleaseGroups returns all Album and EP release groups for an artist,
+// paginating through the MusicBrainz browse API with the required 1 req/s rate limit.
+func getMBArtistReleaseGroups(artistMBID string) ([]mbReleaseGroup, error) {
+ const limit = 100
+ var all []mbReleaseGroup
+
+ for offset := 0; ; offset += limit {
+ path := fmt.Sprintf(
+ "/ws/2/release-group?artist=%s&type=album%%7Cep&fmt=json&limit=%d&offset=%d",
+ url.QueryEscape(artistMBID), limit, offset,
+ )
+
+ var result struct {
+ ReleaseGroups []mbReleaseGroup `json:"release-groups"`
+ Count int `json:"release-group-count"`
+ }
+ if err := mbGet(path, &result); err != nil {
+ return all, err
+ }
+
+ for _, rg := range result.ReleaseGroups {
+ t := strings.ToLower(rg.PrimaryType)
+ if t == "album" || t == "ep" {
+ all = append(all, rg)
+ }
+ }
+
+ if offset+limit >= result.Count {
+ break
+ }
+ time.Sleep(time.Second) // MusicBrainz rate limit
+ }
+
+ return all, nil
+}
+
+// fetchArtist fetches every Album and EP release group for an artist by running
+// fetchRelease for each one sequentially, then registers each for monitoring.
+func fetchArtist(artistMBID, artistName string, logf func(string)) error {
+ log.Printf("[discover] artist fetch started: %s (%s)", artistName, artistMBID)
+ logf(fmt.Sprintf("Looking up discography for %s on MusicBrainz…", artistName))
+
+ groups, err := getMBArtistReleaseGroups(artistMBID)
+ if err != nil {
+ return fmt.Errorf("MusicBrainz discography lookup failed: %w", err)
+ }
+ if len(groups) == 0 {
+ return fmt.Errorf("no albums or EPs found for %s on MusicBrainz", artistName)
+ }
+
+ log.Printf("[discover] found %d release groups for %s", len(groups), artistName)
+ logf(fmt.Sprintf("Found %d albums/EPs", len(groups)))
+
+ failed := 0
+ for i, rg := range groups {
+ logf(fmt.Sprintf("[%d/%d] %s", i+1, len(groups), rg.Title))
+ // Pick the best release for this group. beets --search-id requires a
+ // release MBID; release group MBIDs are not accepted.
+ time.Sleep(time.Second) // MusicBrainz rate limit
+ rel := pickBestReleaseForGroup(rg.ID)
+ releaseMBID := ""
+ trackCount := 0
+ if rel == nil {
+ logf(fmt.Sprintf(" ↳ warning: could not resolve release for group %s, beets will search by name", rg.ID))
+ } else {
+ releaseMBID = rel.ID
+ trackCount = releaseTrackCount(*rel)
+ format := ""
+ if len(rel.Media) > 0 {
+ format = rel.Media[0].Format
+ }
+ logf(fmt.Sprintf(" ↳ selected release: %s [%s / %s / %d tracks]", releaseMBID, format, rel.Country, trackCount))
+ }
+
+ folder, err := fetchRelease(artistName, rg.Title, releaseMBID, trackCount, logf)
+ if err != nil {
+ log.Printf("[discover] fetch failed for %q by %s: %v", rg.Title, artistName, err)
+ logf(fmt.Sprintf(" ↳ failed: %v", err))
+ failed++
+ continue
+ }
+ // Key the pending download by release group ID for dedup; beets uses releaseMBID.
+ registerDownload(rg.ID, releaseMBID, artistName, rg.Title, trackCount, folder, nil)
+ logf(fmt.Sprintf(" ↳ registered for import (release mbid: %s)", releaseMBID))
+ }
+
+ if failed > 0 {
+ logf(fmt.Sprintf("Done — %d/%d queued, %d failed", len(groups)-failed, len(groups), failed))
+ } else {
+ logf(fmt.Sprintf("Done — all %d downloads queued, monitoring for import", len(groups)))
+ }
+ log.Printf("[discover] artist fetch complete: %s (%d/%d succeeded)", artistName, len(groups)-failed, len(groups))
+ return nil
+}
+
+// ── Fetch state ───────────────────────────────────────────────────────────────
+
+type fetchEntry struct {
+ mu sync.Mutex
+ ID string `json:"id"`
+ Artist string `json:"artist"`
+ Album string `json:"album"`
+ Log []string `json:"log"`
+ Done bool `json:"done"`
+ Success bool `json:"success"`
+ ErrMsg string `json:"error,omitempty"`
+}
+
+var (
+ fetchesMu sync.Mutex
+ fetchMap = make(map[string]*fetchEntry)
+)
+
+func newFetchEntry(id, artist, album string) *fetchEntry {
+ e := &fetchEntry{ID: id, Artist: artist, Album: album}
+ fetchesMu.Lock()
+ fetchMap[id] = e
+ fetchesMu.Unlock()
+ return e
+}
+
+func (e *fetchEntry) appendLog(msg string) {
+ e.mu.Lock()
+ e.Log = append(e.Log, msg)
+ e.mu.Unlock()
+}
+
+func (e *fetchEntry) finish(err error) {
+ e.mu.Lock()
+ e.Done = true
+ if err != nil {
+ e.ErrMsg = err.Error()
+ } else {
+ e.Success = true
+ }
+ e.mu.Unlock()
+}
+
+func (e *fetchEntry) snapshot() fetchEntry {
+ e.mu.Lock()
+ defer e.mu.Unlock()
+ cp := *e
+ cp.Log = append([]string(nil), e.Log...)
+ return cp
+}
+
+// ── HTTP handlers ─────────────────────────────────────────────────────────────
+
+// handleDiscoverSearch handles GET /discover/search?q=...&type=release|artist
+func handleDiscoverSearch(w http.ResponseWriter, r *http.Request) {
+ q := r.URL.Query().Get("q")
+ if q == "" {
+ http.Error(w, "missing q", http.StatusBadRequest)
+ return
+ }
+ searchType := r.URL.Query().Get("type")
+ if searchType == "" {
+ searchType = "release"
+ }
+ log.Printf("[discover] search: type=%s q=%q", searchType, q)
+
+ w.Header().Set("Content-Type", "application/json")
+
+ switch searchType {
+ case "artist":
+ artists, err := searchMBArtists(q)
+ if err != nil {
+ log.Printf("[discover] artist search error: %v", err)
+ http.Error(w, err.Error(), http.StatusInternalServerError)
+ return
+ }
+ log.Printf("[discover] artist search returned %d results", len(artists))
+ json.NewEncoder(w).Encode(artists)
+
+ default: // "release"
+ releases, err := searchMBReleases(q)
+ if err != nil {
+ log.Printf("[discover] release search error: %v", err)
+ http.Error(w, err.Error(), http.StatusInternalServerError)
+ return
+ }
+ log.Printf("[discover] release search returned %d results", len(releases))
+ json.NewEncoder(w).Encode(releases)
+ }
+}
+
+// handleDiscoverFetch handles POST /discover/fetch
+// Body: {"id":"mbid","artist":"...","album":"..."}
+func handleDiscoverFetch(w http.ResponseWriter, r *http.Request) {
+ if r.Method != http.MethodPost {
+ http.Error(w, "POST only", http.StatusMethodNotAllowed)
+ return
+ }
+
+ var body struct {
+ ID string `json:"id"`
+ Artist string `json:"artist"`
+ Album string `json:"album"`
+ }
+ if err := json.NewDecoder(r.Body).Decode(&body); err != nil || body.ID == "" || body.Artist == "" || body.Album == "" {
+ http.Error(w, "id, artist and album are required", http.StatusBadRequest)
+ return
+ }
+
+ // If a fetch for this ID is already in progress, return its ID without starting a new one.
+ fetchesMu.Lock()
+ existing := fetchMap[body.ID]
+ fetchesMu.Unlock()
+ if existing != nil {
+ existing.mu.Lock()
+ done := existing.Done
+ existing.mu.Unlock()
+ if !done {
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(map[string]string{"id": body.ID})
+ return
+ }
+ }
+
+ log.Printf("[discover] starting fetch: %q by %s (mbid: %s)", body.Album, body.Artist, body.ID)
+ entry := newFetchEntry(body.ID, body.Artist, body.Album)
+
+ // Look up the expected track count from MusicBrainz so the folder-selection
+ // logic can prefer results that match the release we intend to import.
+ trackCount := 0
+ if rel, err := getMBRelease(body.ID); err == nil {
+ trackCount = releaseTrackCount(*rel)
+ log.Printf("[discover] release %s has %d tracks", body.ID, trackCount)
+ } else {
+ log.Printf("[discover] could not fetch release track count for %s: %v", body.ID, err)
+ }
+
+ go func() {
+ folder, err := fetchRelease(body.Artist, body.Album, body.ID, trackCount, entry.appendLog)
+ if err != nil {
+ log.Printf("[discover] fetch failed for %q by %s: %v", body.Album, body.Artist, err)
+ entry.finish(err)
+ return
+ }
+ log.Printf("[discover] fetch complete for %q by %s, registering for import", body.Album, body.Artist)
+ registerDownload(body.ID, body.ID, body.Artist, body.Album, trackCount, folder, entry)
+ // entry.finish is called by the monitor when import completes
+ }()
+
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(map[string]string{"id": body.ID})
+}
+
+// handleDiscoverFetchArtist handles POST /discover/fetch/artist
+// Body: {"id":"artist-mbid","name":"Artist Name"}
+func handleDiscoverFetchArtist(w http.ResponseWriter, r *http.Request) {
+ if r.Method != http.MethodPost {
+ http.Error(w, "POST only", http.StatusMethodNotAllowed)
+ return
+ }
+
+ var body struct {
+ ID string `json:"id"`
+ Name string `json:"name"`
+ }
+ if err := json.NewDecoder(r.Body).Decode(&body); err != nil || body.ID == "" || body.Name == "" {
+ http.Error(w, "id and name are required", http.StatusBadRequest)
+ return
+ }
+
+ fetchesMu.Lock()
+ existing := fetchMap[body.ID]
+ fetchesMu.Unlock()
+ if existing != nil {
+ existing.mu.Lock()
+ done := existing.Done
+ existing.mu.Unlock()
+ if !done {
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(map[string]string{"id": body.ID})
+ return
+ }
+ }
+
+ log.Printf("[discover] starting artist fetch: %s (%s)", body.Name, body.ID)
+ entry := newFetchEntry(body.ID, body.Name, "")
+ go func() {
+ err := fetchArtist(body.ID, body.Name, entry.appendLog)
+ if err != nil {
+ log.Printf("[discover] artist fetch failed for %s: %v", body.Name, err)
+ } else {
+ log.Printf("[discover] artist fetch complete for %s", body.Name)
+ }
+ entry.finish(err)
+ }()
+
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(map[string]string{"id": body.ID})
+}
+
+// handleDiscoverFetchStatus handles GET /discover/fetch/status?id=...
+func handleDiscoverFetchStatus(w http.ResponseWriter, r *http.Request) {
+ id := r.URL.Query().Get("id")
+ if id == "" {
+ http.Error(w, "missing id", http.StatusBadRequest)
+ return
+ }
+
+ fetchesMu.Lock()
+ entry := fetchMap[id]
+ fetchesMu.Unlock()
+
+ if entry == nil {
+ http.Error(w, "not found", http.StatusNotFound)
+ return
+ }
+
+ snap := entry.snapshot()
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(snap)
+}
+
+// fetchListItem is a summary of a fetch entry for the list endpoint.
+type fetchListItem struct {
+ ID string `json:"id"`
+ Title string `json:"title"`
+ Done bool `json:"done"`
+ Success bool `json:"success"`
+}
+
+// handleDiscoverFetchList handles GET /discover/fetch/list
+// Returns a summary of all known fetch entries so the frontend can discover
+// entries created server-side (e.g. per-album entries from an artist fetch).
+func handleDiscoverFetchList(w http.ResponseWriter, r *http.Request) {
+ fetchesMu.Lock()
+ items := make([]fetchListItem, 0, len(fetchMap))
+ for _, e := range fetchMap {
+ e.mu.Lock()
+ title := e.Artist
+ if e.Album != "" {
+ title = e.Artist + " \u2014 " + e.Album
+ }
+ items = append(items, fetchListItem{
+ ID: e.ID,
+ Title: title,
+ Done: e.Done,
+ Success: e.Success,
+ })
+ e.mu.Unlock()
+ }
+ fetchesMu.Unlock()
+
+ w.Header().Set("Content-Type", "application/json")
+ json.NewEncoder(w).Encode(items)
+}
diff --git a/files.go b/files.go
index 5155997..75a388c 100644
--- a/files.go
+++ b/files.go
@@ -2,22 +2,41 @@ package main
import (
"fmt"
+ "io"
"os"
"path"
"path/filepath"
"strings"
)
-// moveToLibrary moves a file to {libDir}/{artist}/[{year}] {album}/filename.
+// albumTargetDir returns the destination directory for an album without
+// creating it. Use this to check for an existing import before moving files.
+func albumTargetDir(libDir string, md *MusicMetadata) string {
+ date := md.Date
+ if date == "" {
+ date = md.Year
+ }
+ albumDir := fmt.Sprintf("[%s] %s", date, md.Album)
+ if md.Quality != "" {
+ albumDir += fmt.Sprintf(" [%s]", md.Quality)
+ }
+ return filepath.Join(libDir, sanitize(md.Artist), sanitize(albumDir))
+}
+
+// moveToLibrary moves a file to {libDir}/{artist}/[{date}] {album} [{quality}]/filename.
func moveToLibrary(libDir string, md *MusicMetadata, srcPath string) error {
- targetDir := filepath.Join(libDir, sanitize(md.Artist), sanitize(fmt.Sprintf("[%s] %s", md.Year, md.Album)))
+ targetDir := albumTargetDir(libDir, md)
if err := os.MkdirAll(targetDir, 0755); err != nil {
return err
}
dst := filepath.Join(targetDir, filepath.Base(srcPath))
fmt.Println("→ Moving:", srcPath, "→", dst)
- return os.Rename(srcPath, dst)
+ if strings.ToLower(os.Getenv("COPYMODE")) == "true" {
+ return copy(srcPath, dst)
+ } else {
+ return os.Rename(srcPath, dst)
+ }
}
// cluster moves all top-level audio files in dir into subdirectories named
@@ -101,3 +120,63 @@ func sanitize(s string) string {
)
return r.Replace(s)
}
+
+// CopyFile copies a file from src to dst. If src and dst files exist, and are
+// the same, then return success. Otherise, attempt to create a hard link
+// between the two files. If that fail, copy the file contents from src to dst.
+func copy(src, dst string) (err error) {
+ sfi, err := os.Stat(src)
+ if err != nil {
+ return
+ }
+ if !sfi.Mode().IsRegular() {
+ // cannot copy non-regular files (e.g., directories,
+ // symlinks, devices, etc.)
+ return fmt.Errorf("CopyFile: non-regular source file %s (%q)", sfi.Name(), sfi.Mode().String())
+ }
+ dfi, err := os.Stat(dst)
+ if err != nil {
+ if !os.IsNotExist(err) {
+ return
+ }
+ } else {
+ if !(dfi.Mode().IsRegular()) {
+ return fmt.Errorf("CopyFile: non-regular destination file %s (%q)", dfi.Name(), dfi.Mode().String())
+ }
+ if os.SameFile(sfi, dfi) {
+ return
+ }
+ }
+ if err = os.Link(src, dst); err == nil {
+ return
+ }
+ err = copyFileContents(src, dst)
+ return
+}
+
+// copyFileContents copies the contents of the file named src to the file named
+// by dst. The file will be created if it does not already exist. If the
+// destination file exists, all it's contents will be replaced by the contents
+// of the source file.
+func copyFileContents(src, dst string) (err error) {
+ in, err := os.Open(src)
+ if err != nil {
+ return
+ }
+ defer in.Close()
+ out, err := os.Create(dst)
+ if err != nil {
+ return
+ }
+ defer func() {
+ cerr := out.Close()
+ if err == nil {
+ err = cerr
+ }
+ }()
+ if _, err = io.Copy(out, in); err != nil {
+ return
+ }
+ err = out.Sync()
+ return
+}
diff --git a/importer.go b/importer.go
index ac449b5..d108f16 100644
--- a/importer.go
+++ b/importer.go
@@ -5,8 +5,116 @@ import (
"log"
"os"
"path/filepath"
+ "time"
)
+// StepStatus records the outcome of a single pipeline step for an album.
+type StepStatus struct {
+ Skipped bool
+ Err error
+}
+
+func (s StepStatus) Failed() bool { return s.Err != nil }
+
+// MetadataSource identifies which backend resolved the album metadata.
+type MetadataSource string
+
+const (
+ MetadataSourceBeets MetadataSource = "beets"
+ MetadataSourceMusicBrainz MetadataSource = "musicbrainz"
+ MetadataSourceFileTags MetadataSource = "file_tags"
+ MetadataSourceUnknown MetadataSource = ""
+)
+
+// LyricsStats summarises per-track lyric discovery for an album.
+type LyricsStats struct {
+ Total int // total audio tracks examined
+ Synced int // tracks with synced (timestamped) LRC lyrics downloaded
+ Plain int // tracks with plain (un-timestamped) lyrics downloaded
+ AlreadyHad int // tracks that already had an .lrc file, skipped
+ NotFound int // tracks for which no lyrics could be found
+}
+
+func (l LyricsStats) Downloaded() int { return l.Synced + l.Plain }
+
+// CoverArtStats records what happened with cover art for an album.
+type CoverArtStats struct {
+ Found bool // a cover image file was found in the folder
+ Embedded bool // cover was successfully embedded into tracks
+ Source string // filename of the cover image, e.g. "cover.jpg"
+}
+
+// AlbumResult holds the outcome of every pipeline step for one imported album.
+type AlbumResult struct {
+ Name string
+ Path string
+ Metadata *MusicMetadata
+
+ MetadataSource MetadataSource
+ LyricsStats LyricsStats
+ CoverArtStats CoverArtStats
+ TrackCount int
+
+ CleanTags StepStatus
+ TagMetadata StepStatus
+ Lyrics StepStatus
+ ReplayGain StepStatus
+ CoverArt StepStatus
+ Move StepStatus
+
+ // FatalStep is the name of the step that caused the album to be skipped
+ // entirely, or empty if the album completed the full pipeline.
+ FatalStep string
+}
+
+func (a *AlbumResult) skippedAt(step string) {
+ a.FatalStep = step
+}
+
+func (a *AlbumResult) Succeeded() bool { return a.FatalStep == "" }
+func (a *AlbumResult) HasWarnings() bool {
+ if a.CleanTags.Failed() ||
+ a.TagMetadata.Failed() ||
+ a.Lyrics.Failed() ||
+ a.ReplayGain.Failed() ||
+ a.CoverArt.Failed() ||
+ a.Move.Failed() {
+ return true
+ } else {
+ return false
+ }
+}
+
+// ImportSession holds the results of a single importer run.
+type ImportSession struct {
+ StartedAt time.Time
+ FinishedAt time.Time
+ Albums []*AlbumResult
+}
+
+func (s *ImportSession) Failed() []*AlbumResult {
+ var out []*AlbumResult
+ for _, a := range s.Albums {
+ if !a.Succeeded() {
+ out = append(out, a)
+ }
+ }
+ return out
+}
+
+func (s *ImportSession) WithWarnings() []*AlbumResult {
+ var out []*AlbumResult
+ for _, a := range s.Albums {
+ if a.Succeeded() && a.HasWarnings() {
+ out = append(out, a)
+ }
+ }
+ return out
+}
+
+// lastSession is populated at the end of each RunImporter call.
+var lastSession *ImportSession
+
func RunImporter() {
importDir := os.Getenv("IMPORT_DIR")
libraryDir := os.Getenv("LIBRARY_DIR")
@@ -29,6 +137,12 @@ func RunImporter() {
return
}
+ session := &ImportSession{StartedAt: time.Now()}
+ defer func() {
+ session.FinishedAt = time.Now()
+ lastSession = session
+ }()
+
fmt.Println("=== Starting Import ===")
if err := cluster(importDir); err != nil {
@@ -60,59 +174,102 @@ func RunImporter() {
fmt.Println("\n===== Album:", e.Name(), "=====")
+ result := &AlbumResult{Name: e.Name(), Path: albumPath}
+ session.Albums = append(session.Albums, result)
+ result.TrackCount = len(tracks)
+
fmt.Println("→ Cleaning album tags:")
- if err = cleanAlbumTags(albumPath); err != nil {
- fmt.Println("Cleaning album tags failed:", err)
+ result.CleanTags.Err = cleanAlbumTags(albumPath)
+ if result.CleanTags.Failed() {
+ fmt.Println("Cleaning album tags failed:", result.CleanTags.Err)
}
fmt.Println("→ Tagging album metadata:")
- md, err := getAlbumMetadata(albumPath, tracks[0])
+ md, src, err := getAlbumMetadata(albumPath, tracks[0], "")
+ result.TagMetadata.Err = err
+ result.MetadataSource = src
if err != nil {
fmt.Println("Metadata failed, skipping album:", err)
+ result.skippedAt("TagMetadata")
continue
}
+ result.Metadata = md
fmt.Println("→ Fetching synced lyrics from LRCLIB:")
- if err := DownloadAlbumLyrics(albumPath); err != nil {
+ lyricsStats, err := DownloadAlbumLyrics(albumPath)
+ result.Lyrics.Err = err
+ result.LyricsStats = lyricsStats
+ if result.Lyrics.Failed() {
fmt.Println("Failed to download synced lyrics.")
}
fmt.Println("→ Applying ReplayGain to album:", albumPath)
- if err := applyReplayGain(albumPath); err != nil {
- fmt.Println("ReplayGain failed, skipping album:", err)
+ result.ReplayGain.Err = applyReplayGain(albumPath)
+ if result.ReplayGain.Failed() {
+ fmt.Println("ReplayGain failed, skipping album:", result.ReplayGain.Err)
+ result.skippedAt("ReplayGain")
continue
}
+ fmt.Println("→ Downloading cover art for album:", albumPath)
+ if _, err := FindCoverImage(albumPath); err != nil {
+ if err := DownloadCoverArt(albumPath, md, ""); err != nil {
+ fmt.Println("Cover art download failed:", err)
+ }
+ }
+
+ if err := NormalizeCoverArt(albumPath); err != nil {
+ fmt.Println("Cover art normalization warning:", err)
+ }
+
fmt.Println("→ Embedding cover art for album:", albumPath)
- if err := EmbedAlbumArtIntoFolder(albumPath); err != nil {
- fmt.Println("Cover embed failed, skipping album:", err)
+ result.CoverArt.Err = EmbedAlbumArtIntoFolder(albumPath)
+ if coverImg, err := FindCoverImage(albumPath); err == nil {
+ result.CoverArtStats.Found = true
+ result.CoverArtStats.Source = filepath.Base(coverImg)
+ if result.CoverArt.Err == nil {
+ result.CoverArtStats.Embedded = true
+ }
+ }
+ if result.CoverArt.Failed() {
+ fmt.Println("Cover embed failed, skipping album:", result.CoverArt.Err)
+ result.skippedAt("CoverArt")
continue
}
- fmt.Println("→ Moving tracks into library for album:", albumPath)
- for _, track := range tracks {
- if err := moveToLibrary(libraryDir, md, track); err != nil {
- fmt.Println("Failed to move track:", track, err)
+ targetDir := albumTargetDir(libraryDir, md)
+ if _, err := os.Stat(targetDir); err == nil {
+ fmt.Println("→ Album already exists in library, skipping move:", targetDir)
+ result.Move.Skipped = true
+ } else {
+ fmt.Println("→ Moving tracks into library for album:", albumPath)
+ for _, track := range tracks {
+ if err := moveToLibrary(libraryDir, md, track); err != nil {
+ fmt.Println("Failed to move track:", track, err)
+ result.Move.Err = err // retains last error; all attempts are still made
+ }
}
- }
- lyrics, _ := getLyricFiles(albumPath)
+ lyrics, _ := getLyricFiles(albumPath)
- fmt.Println("→ Moving lyrics into library for album:", albumPath)
- for _, file := range lyrics {
- if err := moveToLibrary(libraryDir, md, file); err != nil {
- fmt.Println("Failed to move lyrics:", file, err)
+ fmt.Println("→ Moving lyrics into library for album:", albumPath)
+ for _, file := range lyrics {
+ if err := moveToLibrary(libraryDir, md, file); err != nil {
+ fmt.Println("Failed to move lyrics:", file, err)
+ result.Move.Err = err
+ }
}
- }
- fmt.Println("→ Moving album cover into library for album:", albumPath)
- if coverImg, err := FindCoverImage(albumPath); err == nil {
- if err := moveToLibrary(libraryDir, md, coverImg); err != nil {
- fmt.Println("Failed to cover image:", coverImg, err)
+ fmt.Println("→ Moving album cover into library for album:", albumPath)
+ if coverImg, err := FindCoverImage(albumPath); err == nil {
+ if err := moveToLibrary(libraryDir, md, coverImg); err != nil {
+ fmt.Println("Failed to cover image:", coverImg, err)
+ result.Move.Err = err
+ }
}
- }
- os.Remove(albumPath)
+ os.Remove(albumPath)
+ }
}
fmt.Println("\n=== Import Complete ===")
diff --git a/index.html.tmpl b/index.html.tmpl
new file mode 100644
index 0000000..f89241c
--- /dev/null
+++ b/index.html.tmpl
@@ -0,0 +1,143 @@
+
+
+
+
+
+ Music Importer
+
+
+
+ Music Importer
+
+
+
+
+
+
+
+ {{with .Session}}
+
+
+
+ {{range .Albums}}{{$album := .}}
+
+
+
+ {{with .Metadata}}
+
+ {{.Artist}} — {{.Album}}{{if .Year}} ({{.Year}}){{end}}
+ {{if $album.MetadataSource}}
+
+ via
+ {{if eq (print $album.MetadataSource) "beets"}}
+ beets
+ {{else if eq (print $album.MetadataSource) "musicbrainz"}}
+ MusicBrainz
+ {{else if eq (print $album.MetadataSource) "file_tags"}}
+ file tags
+ {{else}}
+ unknown
+ {{end}}
+
+ {{end}}
+
+ {{end}}
+
+
+
+
Tracks
+
{{.TrackCount}}
+
+
+
+
Lyrics
+ {{if eq .LyricsStats.Total 0}}
+
n/a
+ {{else}}
+
+ {{.LyricsStats.Downloaded}} / {{.LyricsStats.Total}}
+
+
+ {{if gt .LyricsStats.Synced 0}}{{.LyricsStats.Synced}} synced{{end}}
+ {{if and (gt .LyricsStats.Synced 0) (gt .LyricsStats.Plain 0)}} · {{end}}
+ {{if gt .LyricsStats.Plain 0}}{{.LyricsStats.Plain}} plain{{end}}
+ {{if gt .LyricsStats.AlreadyHad 0}} {{.LyricsStats.AlreadyHad}} existing{{end}}
+ {{if gt .LyricsStats.NotFound 0}} {{.LyricsStats.NotFound}} missing{{end}}
+
+ {{end}}
+
+
+
+
Cover Art
+ {{if .CoverArtStats.Found}}
+ {{if .CoverArtStats.Embedded}}
+
Embedded
+
{{.CoverArtStats.Source}}
+ {{else}}
+
Found, not embedded
+
{{.CoverArtStats.Source}}
+ {{end}}
+ {{else}}
+
Not found
+ {{end}}
+
+
+
+ Pipeline
+
+ {{stepCell "Clean Tags" .CleanTags ""}}
+ {{stepCell "Metadata" .TagMetadata .FatalStep}}
+ {{stepCell "Lyrics" .Lyrics ""}}
+ {{stepCell "ReplayGain" .ReplayGain .FatalStep}}
+ {{stepCell "Cover Art" .CoverArt .FatalStep}}
+ {{stepCell "Move" .Move ""}}
+
+
+ {{end}}
+
+ {{end}}
+
+
+
+
+
+
+
+
+
+
diff --git a/lrc.go b/lrc.go
index 94b1233..bdfae86 100644
--- a/lrc.go
+++ b/lrc.go
@@ -51,7 +51,8 @@ func TrackDuration(path string) (int, error) {
// DownloadAlbumLyrics downloads synced lyrics (LRC format) for each track in the album directory.
// Assumes metadata is already final (tags complete).
-func DownloadAlbumLyrics(albumDir string) error {
+func DownloadAlbumLyrics(albumDir string) (LyricsStats, error) {
+ var stats LyricsStats
err := filepath.Walk(albumDir, func(path string, info os.FileInfo, err error) error {
if err != nil {
return err
@@ -64,10 +65,12 @@ func DownloadAlbumLyrics(albumDir string) error {
if ext != ".mp3" && ext != ".flac" {
return nil
}
+ stats.Total++
// Skip if LRC already exists next to the file
lrcPath := strings.TrimSuffix(path, ext) + ".lrc"
if _, err := os.Stat(lrcPath); err == nil {
+ stats.AlreadyHad++
fmt.Println("→ Skipping (already has lyrics):", filepath.Base(path))
return nil
}
@@ -75,18 +78,21 @@ func DownloadAlbumLyrics(albumDir string) error {
// Read metadata
md, err := readTags(path)
if err != nil {
+ stats.NotFound++
fmt.Println("Skipping (unable to read tags):", path, "error:", err)
return nil
}
if md.Title == "" || md.Artist == "" || md.Album == "" {
+ stats.NotFound++
fmt.Println("Skipping (missing metadata):", path)
return nil
}
duration, _ := TrackDuration(path)
- lyrics, err := fetchLRCLibLyrics(md.Artist, md.Title, md.Album, duration)
+ lyrics, synced, err := fetchLRCLibLyrics(md.Artist, md.Title, md.Album, duration)
if err != nil {
+ stats.NotFound++
fmt.Println("No lyrics found:", md.Artist, "-", md.Title)
return nil
}
@@ -96,15 +102,20 @@ func DownloadAlbumLyrics(albumDir string) error {
return fmt.Errorf("writing lrc file for %s: %w", path, err)
}
+ if synced {
+ stats.Synced++
+ } else {
+ stats.Plain++
+ }
fmt.Println("→ Downloaded lyrics:", filepath.Base(lrcPath))
return nil
})
- return err
+ return stats, err
}
// fetchLRCLibLyrics calls the LRCLIB API and returns synced lyrics if available.
-func fetchLRCLibLyrics(artist, title, album string, duration int) (string, error) {
+func fetchLRCLibLyrics(artist, title, album string, duration int) (string, bool, error) {
url := fmt.Sprintf(
"https://lrclib.net/api/get?artist_name=%s&track_name=%s&album_name=%s&duration=%d",
@@ -113,35 +124,35 @@ func fetchLRCLibLyrics(artist, title, album string, duration int) (string, error
resp, err := http.Get(url)
if err != nil {
- return "", fmt.Errorf("lrclib fetch error: %w", err)
+ return "", false, fmt.Errorf("lrclib fetch error: %w", err)
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
- return "", fmt.Errorf("lrclib returned status %d", resp.StatusCode)
+ return "", false, fmt.Errorf("lrclib returned status %d", resp.StatusCode)
}
bodyBytes, err := io.ReadAll(resp.Body)
if err != nil {
- return "", fmt.Errorf("reading lrclib response: %w", err)
+ return "", false, fmt.Errorf("reading lrclib response: %w", err)
}
var out LRCLibResponse
if err := json.Unmarshal(bodyBytes, &out); err != nil {
- return "", fmt.Errorf("parsing lrclib json: %w", err)
+ return "", false, fmt.Errorf("parsing lrclib json: %w", err)
}
if out.SyncedLyrics != "" {
- return out.SyncedLyrics, nil
+ return out.SyncedLyrics, true, nil
}
// If no syncedLyrics, fallback to plain
if out.PlainLyrics != "" {
// Convert plain text to a fake LRC wrapper
- return plainToLRC(out.PlainLyrics), nil
+ return plainToLRC(out.PlainLyrics), false, nil
}
- return "", fmt.Errorf("no lyrics found")
+ return "", false, fmt.Errorf("no lyrics found")
}
// URL escape helper
diff --git a/main.go b/main.go
index a8f2cb5..64a0ec9 100644
--- a/main.go
+++ b/main.go
@@ -1,10 +1,13 @@
package main
import (
+ "embed"
+ "fmt"
+ "html/template"
"log"
"net/http"
"sync"
- "text/template"
+ "time"
)
// version is set at build time via -ldflags="-X main.version=..."
@@ -13,62 +16,89 @@ var version = "dev"
var importerMu sync.Mutex
var importerRunning bool
-var tmpl = template.Must(template.New("index").Parse(`
-
-
-
- Music Importer
-
-
-
- Music Importer
-
-
-
-
-`))
+//go:embed index.html.tmpl
+var tmplFS embed.FS
+
+//go:embed static
+var staticFS embed.FS
+var tmpl = template.Must(
+ template.New("index.html.tmpl").
+ Funcs(template.FuncMap{
+ // duration formats the elapsed time between two timestamps.
+ "duration": func(start, end time.Time) string {
+ if end.IsZero() {
+ return ""
+ }
+ d := end.Sub(start).Round(time.Second)
+ if d < time.Minute {
+ return d.String()
+ }
+ return fmt.Sprintf("%dm %ds", int(d.Minutes()), int(d.Seconds())%60)
+ },
+ // not is needed because Go templates have no built-in boolean negation.
+ "not": func(b bool) bool { return !b },
+ // stepCell renders a uniform step status cell.
+ // fatalStep is AlbumResult.FatalStep; when it matches the step's key
+ // the cell is marked fatal rather than a warning.
+ "stepCell": func(label string, s StepStatus, fatalStep string) template.HTML {
+ var statusClass, statusText, errHTML string
+ switch {
+ case s.Err != nil && fatalStep != "" && stepKey(label) == fatalStep:
+ statusClass = "step-fatal"
+ statusText = "✗ fatal"
+ errHTML = `` + template.HTMLEscapeString(s.Err.Error()) + ``
+ case s.Err != nil:
+ statusClass = "step-warn"
+ statusText = "⚠ error"
+ errHTML = `` + template.HTMLEscapeString(s.Err.Error()) + ``
+ case s.Skipped:
+ statusClass = "step-warn"
+ statusText = "– skipped"
+ default:
+ statusClass = "step-ok"
+ statusText = "✓ ok"
+ }
+ return template.HTML(`` +
+ `` + template.HTMLEscapeString(label) + `` +
+ `` + statusText + `` +
+ errHTML +
+ `
`)
+ },
+ }).
+ ParseFS(tmplFS, "index.html.tmpl"),
+)
+
+// stepKey maps a human-readable step label to the FatalStep identifier used in
+// AlbumResult so the template can highlight the step that caused the abort.
+func stepKey(label string) string {
+ switch label {
+ case "Metadata":
+ return "TagMetadata"
+ case "Cover Art":
+ return "CoverArt"
+ default:
+ return label
+ }
+}
+
+type templateData struct {
+ Running bool
+ Version string
+ Session *ImportSession
+}
func handleHome(w http.ResponseWriter, r *http.Request) {
importerMu.Lock()
running := importerRunning
importerMu.Unlock()
- tmpl.Execute(w, struct {
- Running bool
- Version string
- }{Running: running, Version: version})
+ if err := tmpl.Execute(w, templateData{
+ Running: running,
+ Version: version,
+ Session: lastSession,
+ }); err != nil {
+ log.Println("Template error:", err)
+ }
}
func handleRun(w http.ResponseWriter, r *http.Request) {
@@ -93,8 +123,15 @@ func handleRun(w http.ResponseWriter, r *http.Request) {
func main() {
log.Printf("Music Importer %s starting on http://localhost:8080", version)
+ startMonitor()
+ http.Handle("/static/", http.FileServer(http.FS(staticFS)))
http.HandleFunc("/", handleHome)
http.HandleFunc("/run", handleRun)
+ http.HandleFunc("/discover/search", handleDiscoverSearch)
+ http.HandleFunc("/discover/fetch", handleDiscoverFetch)
+ http.HandleFunc("/discover/fetch/artist", handleDiscoverFetchArtist)
+ http.HandleFunc("/discover/fetch/status", handleDiscoverFetchStatus)
+ http.HandleFunc("/discover/fetch/list", handleDiscoverFetchList)
log.Fatal(http.ListenAndServe(":8080", nil))
}
diff --git a/media.go b/media.go
index 60711a2..e0274a6 100644
--- a/media.go
+++ b/media.go
@@ -2,7 +2,11 @@ package main
import (
"bytes"
+ "encoding/json"
"fmt"
+ "io"
+ "net/http"
+ "net/url"
"os"
"os/exec"
"path/filepath"
@@ -53,6 +57,158 @@ func EmbedAlbumArtIntoFolder(albumDir string) error {
return err
}
+// DownloadCoverArt downloads the front cover from the Cover Art Archive and
+// saves it as cover.jpg/cover.png inside albumDir.
+// If mbid is non-empty it is used directly, bypassing the MusicBrainz search.
+// Otherwise, a search is performed using md's artist and album.
+func DownloadCoverArt(albumDir string, md *MusicMetadata, mbid string) error {
+ if mbid == "" {
+ var err error
+ mbid, err = searchMusicBrainzRelease(md.Artist, md.Album)
+ if err != nil {
+ return fmt.Errorf("MusicBrainz release search failed: %w", err)
+ }
+ }
+
+ data, ext, err := fetchCoverArtArchiveFront(mbid)
+ if err != nil {
+ return fmt.Errorf("Cover Art Archive fetch failed: %w", err)
+ }
+
+ dest := filepath.Join(albumDir, "cover."+ext)
+ if err := os.WriteFile(dest, data, 0644); err != nil {
+ return fmt.Errorf("writing cover image: %w", err)
+ }
+
+ fmt.Println("→ Downloaded cover art:", filepath.Base(dest))
+ return nil
+}
+
+// searchMusicBrainzRelease queries the MusicBrainz API for a release matching
+// the given artist and album and returns its MBID.
+func searchMusicBrainzRelease(artist, album string) (string, error) {
+ q := fmt.Sprintf(`release:"%s" AND artist:"%s"`,
+ strings.ReplaceAll(album, `"`, `\"`),
+ strings.ReplaceAll(artist, `"`, `\"`),
+ )
+ apiURL := "https://musicbrainz.org/ws/2/release/?query=" + url.QueryEscape(q) + "&fmt=json&limit=1"
+
+ req, err := http.NewRequest("GET", apiURL, nil)
+ if err != nil {
+ return "", err
+ }
+ req.Header.Set("User-Agent", "music-importer/1.0 (https://github.com/example/music-importer)")
+
+ resp, err := http.DefaultClient.Do(req)
+ if err != nil {
+ return "", err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusOK {
+ return "", fmt.Errorf("MusicBrainz returned status %d", resp.StatusCode)
+ }
+
+ var result struct {
+ Releases []struct {
+ ID string `json:"id"`
+ } `json:"releases"`
+ }
+ if err := json.NewDecoder(resp.Body).Decode(&result); err != nil {
+ return "", err
+ }
+ if len(result.Releases) == 0 {
+ return "", fmt.Errorf("no MusicBrainz release found for %q by %q", album, artist)
+ }
+ return result.Releases[0].ID, nil
+}
+
+// fetchCoverArtArchiveFront fetches the front cover image for the given
+// MusicBrainz release MBID from coverartarchive.org. It follows the 307
+// redirect to the actual image and returns the raw bytes plus the file
+// extension (e.g. "jpg" or "png").
+func fetchCoverArtArchiveFront(mbid string) ([]byte, string, error) {
+ apiURL := "https://coverartarchive.org/release/" + mbid + "/front"
+
+ resp, err := http.Get(apiURL)
+ if err != nil {
+ return nil, "", err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusOK {
+ return nil, "", fmt.Errorf("Cover Art Archive returned status %d for MBID %s", resp.StatusCode, mbid)
+ }
+
+ data, err := io.ReadAll(resp.Body)
+ if err != nil {
+ return nil, "", err
+ }
+
+ // Derive the extension from the final URL after redirect, falling back to
+ // sniffing the magic bytes.
+ ext := "jpg"
+ if finalURL := resp.Request.URL.String(); strings.HasSuffix(strings.ToLower(finalURL), ".png") {
+ ext = "png"
+ } else if bytes.HasPrefix(data, []byte{0x89, 0x50, 0x4E, 0x47}) {
+ ext = "png"
+ }
+
+ return data, ext, nil
+}
+
+const coverMaxBytes = 5 * 1024 * 1024 // 5 MB
+
+// NormalizeCoverArt checks whether the cover image in albumDir is a large
+// non-JPEG (>5 MB). If so, it converts it to JPEG and resizes it to at most
+// 2000×2000 pixels using ffmpeg, replacing the original file with cover.jpg.
+// The function is a no-op when no cover is found, the cover is already JPEG,
+// or the file is ≤5 MB.
+func NormalizeCoverArt(albumDir string) error {
+ cover, err := FindCoverImage(albumDir)
+ if err != nil {
+ return nil // no cover present, nothing to do
+ }
+
+ // Already JPEG — no conversion needed regardless of size.
+ ext := strings.ToLower(filepath.Ext(cover))
+ if ext == ".jpg" || ext == ".jpeg" {
+ return nil
+ }
+
+ info, err := os.Stat(cover)
+ if err != nil {
+ return fmt.Errorf("stat cover: %w", err)
+ }
+ if info.Size() <= coverMaxBytes {
+ return nil // small enough, leave as-is
+ }
+
+ dest := filepath.Join(albumDir, "cover.jpg")
+ fmt.Printf("→ Cover art is %.1f MB %s; converting to JPEG (max 2000×2000)…\n",
+ float64(info.Size())/(1024*1024), strings.ToUpper(strings.TrimPrefix(ext, ".")))
+
+ // scale=2000:2000:force_original_aspect_ratio=decrease fits the image within
+ // 2000×2000 while preserving aspect ratio, and never upscales smaller images.
+ cmd := exec.Command("ffmpeg", "-y", "-i", cover,
+ "-vf", "scale=2000:2000:force_original_aspect_ratio=decrease",
+ "-q:v", "2",
+ dest,
+ )
+ if out, err := cmd.CombinedOutput(); err != nil {
+ return fmt.Errorf("ffmpeg cover conversion failed: %w\n%s", err, out)
+ }
+
+ if cover != dest {
+ if err := os.Remove(cover); err != nil {
+ fmt.Println("Warning: could not remove original cover:", err)
+ }
+ }
+
+ fmt.Println("→ Converted cover art to JPEG:", filepath.Base(dest))
+ return nil
+}
+
// -------------------------
// Find cover image
// -------------------------
diff --git a/metadata.go b/metadata.go
index 0861567..da146a8 100644
--- a/metadata.go
+++ b/metadata.go
@@ -1,19 +1,25 @@
package main
import (
+ "bufio"
"encoding/json"
"errors"
"fmt"
+ "math"
+ "os"
"os/exec"
"path/filepath"
+ "strconv"
"strings"
)
type MusicMetadata struct {
- Artist string
- Album string
- Title string
- Year string
+ Artist string
+ Album string
+ Title string
+ Year string // four-digit year, kept for backward compat
+ Date string // normalised as YYYY.MM.DD (or YYYY.MM or YYYY)
+ Quality string // e.g. "FLAC-24bit-96kHz" or "MP3-320kbps"
}
// Read embedded tags from an audio file using ffprobe.
@@ -39,18 +45,243 @@ func readTags(path string) (*MusicMetadata, error) {
return &MusicMetadata{}, nil
}
+ rawDate := firstNonEmpty(t["date"], t["DATE"], t["year"], t["YEAR"], t["ORIGINALYEAR"])
+ date := parseDate(rawDate)
+ year := ""
+ if len(date) >= 4 {
+ year = date[:4]
+ }
+
return &MusicMetadata{
Artist: firstNonEmpty(t["artist"], t["ARTIST"]),
Album: firstNonEmpty(t["album"], t["ALBUM"]),
Title: firstNonEmpty(t["title"], t["TITLE"]),
- Year: firstNonEmpty(t["year"], t["YEAR"], t["ORIGINALYEAR"]),
+ Year: year,
+ Date: date,
}, nil
}
+// parseDate normalises a raw DATE/date tag value into YYYY.MM.DD (or YYYY.MM
+// or YYYY) dot-separated format, or returns the input unchanged if it cannot
+// be recognised.
+//
+// Supported input formats:
+// - YYYY
+// - YYYY-MM
+// - YYYY-MM-DD
+// - YYYYMMDD
+func parseDate(raw string) string {
+ raw = strings.TrimSpace(raw)
+ if raw == "" {
+ return ""
+ }
+
+ // YYYYMMDD (exactly 8 digits, no separators)
+ if len(raw) == 8 && isAllDigits(raw) {
+ return raw[0:4] + "." + raw[4:6] + "." + raw[6:8]
+ }
+
+ // YYYY-MM-DD, YYYY-MM, or YYYY (with dashes)
+ parts := strings.Split(raw, "-")
+ switch len(parts) {
+ case 1:
+ if len(parts[0]) == 4 && isAllDigits(parts[0]) {
+ return parts[0]
+ }
+ case 2:
+ if len(parts[0]) == 4 && isAllDigits(parts[0]) && len(parts[1]) == 2 && isAllDigits(parts[1]) {
+ return parts[0] + "." + parts[1]
+ }
+ case 3:
+ if len(parts[0]) == 4 && isAllDigits(parts[0]) &&
+ len(parts[1]) == 2 && isAllDigits(parts[1]) &&
+ len(parts[2]) == 2 && isAllDigits(parts[2]) {
+ return parts[0] + "." + parts[1] + "." + parts[2]
+ }
+ }
+
+ // Unrecognised — return as-is so we don't silently drop it.
+ return raw
+}
+
+func isAllDigits(s string) bool {
+ for _, c := range s {
+ if c < '0' || c > '9' {
+ return false
+ }
+ }
+ return len(s) > 0
+}
+
+// readAudioQuality probes the first audio stream of path and returns a
+// quality label such as "FLAC-24bit-96kHz" or "MP3-320kbps".
+func readAudioQuality(path string) (string, error) {
+ out, err := exec.Command(
+ "ffprobe", "-v", "quiet", "-print_format", "json",
+ "-show_streams", "-select_streams", "a:0",
+ path,
+ ).Output()
+ if err != nil {
+ return "", err
+ }
+
+ var data struct {
+ Streams []struct {
+ CodecName string `json:"codec_name"`
+ SampleRate string `json:"sample_rate"`
+ BitRate string `json:"bit_rate"`
+ BitsPerRawSample string `json:"bits_per_raw_sample"`
+ } `json:"streams"`
+ }
+
+ if err := json.Unmarshal(out, &data); err != nil {
+ return "", err
+ }
+ if len(data.Streams) == 0 {
+ return "", fmt.Errorf("no audio streams found in %s", path)
+ }
+
+ s := data.Streams[0]
+ codec := strings.ToUpper(s.CodecName) // e.g. "FLAC", "MP3"
+
+ switch strings.ToLower(s.CodecName) {
+ case "flac":
+ bits := s.BitsPerRawSample
+ if bits == "" || bits == "0" {
+ bits = "16" // safe fallback
+ }
+ khz := sampleRateToKHz(s.SampleRate)
+ return fmt.Sprintf("%s-%sbit-%s", codec, bits, khz), nil
+
+ case "mp3":
+ kbps := snapMP3Bitrate(s.BitRate)
+ return fmt.Sprintf("%s-%dkbps", codec, kbps), nil
+
+ default:
+ // Generic fallback: codec + bitrate if available.
+ if s.BitRate != "" && s.BitRate != "0" {
+ kbps := snapMP3Bitrate(s.BitRate)
+ return fmt.Sprintf("%s-%dkbps", codec, kbps), nil
+ }
+ return codec, nil
+ }
+}
+
+// sampleRateToKHz converts a sample-rate string in Hz (e.g. "44100") to a
+// human-friendly kHz string (e.g. "44.1kHz").
+func sampleRateToKHz(hz string) string {
+ n, err := strconv.Atoi(strings.TrimSpace(hz))
+ if err != nil || n == 0 {
+ return "?kHz"
+ }
+ if n%1000 == 0 {
+ return fmt.Sprintf("%dkHz", n/1000)
+ }
+ return fmt.Sprintf("%.1fkHz", float64(n)/1000.0)
+}
+
+// commonMP3Bitrates lists the standard MPEG audio bitrates in kbps.
+var commonMP3Bitrates = []int{32, 40, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 256, 320}
+
+// snapMP3Bitrate rounds a raw bitrate string (in bps) to the nearest standard
+// MP3 bitrate (in kbps). For example "318731" → 320.
+func snapMP3Bitrate(bpsStr string) int {
+ bps, err := strconv.Atoi(strings.TrimSpace(bpsStr))
+ if err != nil || bps <= 0 {
+ return 128 // safe fallback
+ }
+ kbps := float64(bps) / 1000.0
+ best := commonMP3Bitrates[0]
+ bestDiff := math.Abs(kbps - float64(best))
+ for _, candidate := range commonMP3Bitrates[1:] {
+ if d := math.Abs(kbps - float64(candidate)); d < bestDiff {
+ bestDiff = d
+ best = candidate
+ }
+ }
+ return best
+}
+
// Use beets to fetch metadata and tag all files in a directory.
-func tagWithBeets(path string) error {
+// A temp log file is passed to beets via -l so that skipped albums
+// (which exit 0 but produce a "skip" log entry) are detected and
+// returned as errors, triggering the MusicBrainz fallback.
+// If mbid is non-empty it is passed as --search-id to pin beets to a specific
+// MusicBrainz release. In that case, quiet mode is skipped and newlines are
+// piped to stdin so beets auto-accepts the pinned release regardless of
+// confidence score.
+func tagWithBeets(path, mbid string) error {
fmt.Println("→ Tagging with beets:", path)
- return runCmd("beet", "import", "-Cq", path)
+
+ logFile, err := os.CreateTemp("", "beets-log-*.txt")
+ if err != nil {
+ return fmt.Errorf("beets: could not create temp log file: %w", err)
+ }
+ logPath := logFile.Name()
+ logFile.Close()
+ defer os.Remove(logPath)
+
+ args := []string{"import", "-C", "-l", logPath}
+ if mbid != "" {
+ // Drop -q so beets doesn't skip on low confidence. Pipe newlines to
+ // auto-accept the interactive prompt for the MBID-pinned release.
+ args = append(args, "--search-id", mbid, path)
+ cmd := exec.Command("beet", args...)
+ cmd.Stdout = os.Stdout
+ cmd.Stderr = os.Stderr
+ cmd.Stdin = strings.NewReader(strings.Repeat("A\n", 20))
+ if err := cmd.Run(); err != nil {
+ return err
+ }
+ } else {
+ args = append(args, "-q", path)
+ if err := runCmd("beet", args...); err != nil {
+ return err
+ }
+ }
+
+ // Even on exit 0, beets may have skipped the album in quiet mode.
+ // The log format is one entry per line: " "
+ // We treat any "skip" line as a failure so the caller falls through
+ // to the MusicBrainz lookup.
+ skipped, err := beetsLogHasSkip(logPath)
+ if err != nil {
+ // If we can't read the log, assume beets succeeded.
+ fmt.Println("beets: could not read log file:", err)
+ return nil
+ }
+ if skipped {
+ return errors.New("beets skipped album (no confident match found)")
+ }
+ return nil
+}
+
+// beetsLogHasSkip reads a beets import log file and reports whether any
+// entry has the action "skip". The log format is:
+//
+// # beets import log
+//
+// ...
+func beetsLogHasSkip(logPath string) (bool, error) {
+ f, err := os.Open(logPath)
+ if err != nil {
+ return false, err
+ }
+ defer f.Close()
+
+ scanner := bufio.NewScanner(f)
+ for scanner.Scan() {
+ line := strings.TrimSpace(scanner.Text())
+ // Skip blank lines and the header comment.
+ if line == "" || strings.HasPrefix(line, "#") {
+ continue
+ }
+ action, _, found := strings.Cut(line, " ")
+ if found && strings.EqualFold(action, "skip") {
+ return true, nil
+ }
+ }
+ return false, scanner.Err()
}
// Fallback: query MusicBrainz API manually if beets fails.
@@ -99,26 +330,44 @@ func fetchMusicBrainzInfo(filename string) (*MusicMetadata, error) {
// getAlbumMetadata attempts beets tagging on the album directory, reads tags
// back from the first track, and falls back to MusicBrainz if tags are missing.
-func getAlbumMetadata(albumPath, trackPath string) (*MusicMetadata, error) {
+// If mbid is non-empty it is forwarded to beets as --search-id.
+func getAlbumMetadata(albumPath, trackPath, mbid string) (*MusicMetadata, MetadataSource, error) {
fmt.Println("→ Tagging track with beets:", trackPath)
- if err := tagWithBeets(albumPath); err != nil {
- fmt.Println("Beets tagging failed; fallback to manual MusicBrainz lookup:", err)
+ beetsErr := tagWithBeets(albumPath, mbid)
+ if beetsErr != nil {
+ fmt.Println("Beets tagging failed; fallback to manual MusicBrainz lookup:", beetsErr)
}
md, err := readTags(trackPath)
if err == nil && md.Artist != "" && md.Album != "" {
- return md, nil
+ attachQuality(md, trackPath)
+ if beetsErr == nil {
+ return md, MetadataSourceBeets, nil
+ }
+ return md, MetadataSourceFileTags, nil
}
fmt.Println("→ Missing tags, attempting MusicBrainz manual lookup...")
md, err = fetchMusicBrainzInfo(trackPath)
if err != nil {
- return nil, fmt.Errorf("metadata lookup failed: %w", err)
+ return nil, MetadataSourceUnknown, fmt.Errorf("metadata lookup failed: %w", err)
}
- return md, nil
+ attachQuality(md, trackPath)
+ return md, MetadataSourceMusicBrainz, nil
+}
+
+// attachQuality probes trackPath for audio quality and sets md.Quality.
+// Errors are logged but not returned — a missing quality label is non-fatal.
+func attachQuality(md *MusicMetadata, trackPath string) {
+ q, err := readAudioQuality(trackPath)
+ if err != nil {
+ fmt.Println("Could not determine audio quality:", err)
+ return
+ }
+ md.Quality = q
}
func firstNonEmpty(vals ...string) string {
diff --git a/monitor.go b/monitor.go
new file mode 100644
index 0000000..c09438b
--- /dev/null
+++ b/monitor.go
@@ -0,0 +1,304 @@
+package main
+
+import (
+ "fmt"
+ "log"
+ "os"
+ "path/filepath"
+ "strings"
+ "sync"
+ "time"
+)
+
+// pendingDownload tracks a queued slskd download that should be auto-imported
+// once all files have transferred successfully.
+type pendingDownload struct {
+ ID string // dedup key (release MBID for single fetches; release group MBID for artist fetches)
+ BeetsMBID string // release MBID passed to beets --search-id (may differ from ID)
+ Artist string
+ Album string
+ Username string // slskd peer username
+ Dir string // remote directory path on the peer
+ Files []slskdFile // files that were queued for download
+ Entry *fetchEntry // fetch card to update with import progress
+ TrackCount int // expected number of audio tracks (0 = unknown, skip check)
+}
+
+var (
+ pendingMu sync.Mutex
+ pendingDownloads = make(map[string]*pendingDownload) // keyed by MBID
+)
+
+// registerDownload records a queued slskd download for monitoring and eventual
+// auto-import. id is used as the dedup key; beetsMBID is the release MBID
+// forwarded to beets --search-id (may be empty or differ from id).
+// trackCount is the expected number of audio tracks from MusicBrainz; 0 means
+// unknown and the sanity check will be skipped at import time.
+// If entry is nil a new fetchEntry is created so the frontend can discover it
+// via /discover/fetch/list.
+func registerDownload(id, beetsMBID, artist, album string, trackCount int, folder *albumFolder, entry *fetchEntry) {
+ pd := &pendingDownload{
+ ID: id,
+ BeetsMBID: beetsMBID,
+ Artist: artist,
+ Album: album,
+ Username: folder.Username,
+ Dir: folder.Dir,
+ Files: folder.Files,
+ Entry: entry,
+ TrackCount: trackCount,
+ }
+
+ if entry == nil {
+ e := newFetchEntry(id, artist, album)
+ e.appendLog(fmt.Sprintf("Queued %d files from %s — waiting for download",
+ len(folder.Files), folder.Username))
+ pd.Entry = e
+ }
+
+ pendingMu.Lock()
+ pendingDownloads[id] = pd
+ pendingMu.Unlock()
+
+ log.Printf("[monitor] registered: %q by %s (id: %s, beets mbid: %s, peer: %s, %d files, expected tracks: %d)",
+ album, artist, id, beetsMBID, folder.Username, len(folder.Files), trackCount)
+}
+
+// startMonitor launches a background goroutine that periodically checks whether
+// pending downloads have completed and triggers import when they have.
+func startMonitor() {
+ go func() {
+ for {
+ time.Sleep(15 * time.Second)
+ checkPendingDownloads()
+ }
+ }()
+ log.Println("[monitor] started")
+}
+
+// checkPendingDownloads polls slskd transfer state for every pending download
+// and kicks off importPendingRelease for any that are fully complete.
+func checkPendingDownloads() {
+ pendingMu.Lock()
+ if len(pendingDownloads) == 0 {
+ pendingMu.Unlock()
+ return
+ }
+ snapshot := make(map[string]*pendingDownload, len(pendingDownloads))
+ for k, v := range pendingDownloads {
+ snapshot[k] = v
+ }
+ pendingMu.Unlock()
+
+ log.Printf("[monitor] checking %d pending download(s)", len(snapshot))
+
+ // Group by username to minimise API calls.
+ byUser := make(map[string][]*pendingDownload)
+ for _, pd := range snapshot {
+ byUser[pd.Username] = append(byUser[pd.Username], pd)
+ }
+
+ for username, pds := range byUser {
+ dirs, err := getSlskdTransfers(username)
+ if err != nil {
+ log.Printf("[monitor] failed to get transfers for %s: %v", username, err)
+ continue
+ }
+
+ // Index transfer dirs by normalised path.
+ transfersByDir := make(map[string][]slskdTransferFile, len(dirs))
+ for _, d := range dirs {
+ norm := strings.ReplaceAll(d.Directory, "\\", "/")
+ transfersByDir[norm] = d.Files
+ }
+
+ for _, pd := range pds {
+ normDir := strings.ReplaceAll(pd.Dir, "\\", "/")
+ files, ok := transfersByDir[normDir]
+ if !ok {
+ log.Printf("[monitor] transfer dir not found yet for %q (peer: %s)", pd.Dir, username)
+ continue
+ }
+
+ if !allFilesCompleted(files) {
+ log.Printf("[monitor] %q by %s: download still in progress", pd.Album, pd.Artist)
+ continue
+ }
+
+ localDir := localDirForDownload(pd, files)
+ if localDir == "" {
+ log.Printf("[monitor] cannot determine local dir for %q by %s", pd.Album, pd.Artist)
+ pd.Entry.appendLog("Error: could not determine local download path from transfer info")
+ continue
+ }
+
+ log.Printf("[monitor] download complete: %q by %s → %s", pd.Album, pd.Artist, localDir)
+
+ // Remove from pending before starting import to avoid double-import.
+ pendingMu.Lock()
+ delete(pendingDownloads, pd.ID)
+ pendingMu.Unlock()
+
+ go importPendingRelease(pd, localDir)
+ }
+ }
+}
+
+// allFilesCompleted reports whether every file in a transfer directory has
+// reached a terminal Completed state. Returns false if files is empty.
+func allFilesCompleted(files []slskdTransferFile) bool {
+ if len(files) == 0 {
+ return false
+ }
+ for _, f := range files {
+ if !strings.Contains(f.State, "Completed") {
+ return false
+ }
+ }
+ return true
+}
+
+// localDirForDownload resolves the local filesystem path for a completed download.
+//
+// Strategy 1 — localFilename from transfer metadata: slskd sets this field to
+// the absolute path of the downloaded file. Works when paths are consistent
+// across containers (same volume mount point).
+//
+// Strategy 2 — SLSKD_DOWNLOAD_DIR reconstruction: slskd stores files under
+// {downloadDir}/{username}/{sanitized_remote_dir}/. Used when localFilename is
+// empty or when SLSKD_DOWNLOAD_DIR is explicitly set to override.
+func localDirForDownload(pd *pendingDownload, files []slskdTransferFile) string {
+ // Strategy 1: use localFilename from transfer response.
+ for _, f := range files {
+ if f.LocalFilename != "" {
+ dir := filepath.Dir(f.LocalFilename)
+ log.Printf("[monitor] local dir from localFilename: %s", dir)
+ return dir
+ }
+ }
+
+ // Strategy 2: reconstruct from SLSKD_DOWNLOAD_DIR.
+ // slskd places downloaded files directly into {downloadDir}/{album_folder_name}/,
+ // where the folder name is the last path component of the remote directory.
+ dlDir := os.Getenv("SLSKD_DOWNLOAD_DIR")
+ if dlDir == "" {
+ log.Printf("[monitor] localFilename empty and SLSKD_DOWNLOAD_DIR not set — cannot determine local dir for %q", pd.Album)
+ return ""
+ }
+
+ dir := filepath.Join(dlDir, filepath.Base(filepath.FromSlash(pd.Dir)))
+ log.Printf("[monitor] local dir reconstructed from SLSKD_DOWNLOAD_DIR: %s", dir)
+ return dir
+}
+
+// importPendingRelease runs the full import pipeline on a completed download.
+// It mirrors RunImporter's per-album logic but uses the MBID for beets tagging.
+func importPendingRelease(pd *pendingDownload, localDir string) {
+ entry := pd.Entry
+ logf := func(msg string) {
+ entry.appendLog("[import] " + msg)
+ log.Printf("[monitor/import %s] %s", pd.ID, msg)
+ }
+
+ logf(fmt.Sprintf("Starting import from %s", localDir))
+
+ libraryDir := os.Getenv("LIBRARY_DIR")
+ if libraryDir == "" {
+ entry.finish(fmt.Errorf("LIBRARY_DIR is not set"))
+ return
+ }
+
+ tracks, err := getAudioFiles(localDir)
+ if err != nil {
+ entry.finish(fmt.Errorf("scanning audio files: %w", err))
+ return
+ }
+ if len(tracks) == 0 {
+ entry.finish(fmt.Errorf("no audio files found in %s", localDir))
+ return
+ }
+ logf(fmt.Sprintf("Found %d tracks", len(tracks)))
+
+ if pd.TrackCount > 0 && len(tracks) != pd.TrackCount {
+ entry.finish(fmt.Errorf(
+ "track count mismatch: downloaded %d tracks but release expects %d — aborting to avoid importing wrong edition",
+ len(tracks), pd.TrackCount,
+ ))
+ return
+ }
+
+ if err := cleanAlbumTags(localDir); err != nil {
+ logf(fmt.Sprintf("Clean tags warning: %v", err))
+ }
+
+ md, src, err := getAlbumMetadata(localDir, tracks[0], pd.BeetsMBID)
+ if err != nil {
+ entry.finish(fmt.Errorf("metadata failed: %w", err))
+ return
+ }
+ logf(fmt.Sprintf("Tagged via %s: %s — %s", src, md.Artist, md.Album))
+
+ if _, err := DownloadAlbumLyrics(localDir); err != nil {
+ logf(fmt.Sprintf("Lyrics warning: %v", err))
+ }
+
+ if err := applyReplayGain(localDir); err != nil {
+ entry.finish(fmt.Errorf("ReplayGain failed: %w", err))
+ return
+ }
+ logf("ReplayGain applied")
+
+ if _, err := FindCoverImage(localDir); err != nil {
+ if err := DownloadCoverArt(localDir, md, pd.BeetsMBID); err != nil {
+ logf(fmt.Sprintf("Cover art download warning: %v", err))
+ }
+ }
+
+ if err := NormalizeCoverArt(localDir); err != nil {
+ logf(fmt.Sprintf("Cover art normalization warning: %v", err))
+ }
+
+ if err := EmbedAlbumArtIntoFolder(localDir); err != nil {
+ entry.finish(fmt.Errorf("cover embed failed: %w", err))
+ return
+ }
+ logf("Cover art embedded")
+
+ targetDir := albumTargetDir(libraryDir, md)
+ if _, err := os.Stat(targetDir); err == nil {
+ logf(fmt.Sprintf("Album already exists in library, skipping move: %s", targetDir))
+ entry.finish(nil)
+ return
+ }
+
+ var moveErr error
+ for _, track := range tracks {
+ if err := moveToLibrary(libraryDir, md, track); err != nil {
+ logf(fmt.Sprintf("Move warning: %v", err))
+ moveErr = err
+ }
+ }
+
+ lyrics, _ := getLyricFiles(localDir)
+ for _, file := range lyrics {
+ if err := moveToLibrary(libraryDir, md, file); err != nil {
+ logf(fmt.Sprintf("Move lyrics warning: %v", err))
+ }
+ }
+
+ if coverImg, err := FindCoverImage(localDir); err == nil {
+ if err := moveToLibrary(libraryDir, md, coverImg); err != nil {
+ logf(fmt.Sprintf("Move cover warning: %v", err))
+ }
+ }
+
+ os.Remove(localDir)
+
+ if moveErr != nil {
+ entry.finish(fmt.Errorf("import completed with move errors: %w", moveErr))
+ return
+ }
+
+ logf("Import complete")
+ entry.finish(nil)
+}
diff --git a/slskd.go b/slskd.go
new file mode 100644
index 0000000..70fcad9
--- /dev/null
+++ b/slskd.go
@@ -0,0 +1,460 @@
+package main
+
+import (
+ "bytes"
+ "encoding/json"
+ "fmt"
+ "io"
+ "log"
+ "net/http"
+ "os"
+ "path"
+ "strings"
+ "time"
+)
+
+// slskdAttr is a Soulseek file attribute (bitrate, sample rate, bit depth, etc.).
+// Attribute types: 0 = bitrate (kbps), 1 = duration (s), 2 = VBR flag,
+//
+// 4 = sample rate (Hz), 5 = bit depth.
+type slskdAttr struct {
+ Type int `json:"type"`
+ Value int `json:"value"`
+}
+
+// slskdFile is a single file in a slskd search response.
+type slskdFile struct {
+ Filename string `json:"filename"`
+ Size int64 `json:"size"`
+ Extension string `json:"extension"`
+ Attributes []slskdAttr `json:"attributes"`
+}
+
+// slskdPeerResponse is one peer's response to a search.
+type slskdPeerResponse struct {
+ Username string `json:"username"`
+ Files []slskdFile `json:"files"`
+}
+
+// slskdSearch is the search-state object returned by GET /api/v0/searches/{id}.
+// File responses are not included here; fetch them from /searches/{id}/responses.
+type slskdSearch struct {
+ ID string `json:"id"`
+ State string `json:"state"`
+}
+
+// Quality tiers; higher value = more preferred.
+const (
+ qualityUnknown = 0
+ qualityMP3Any = 1
+ qualityMP3_320 = 2
+ qualityFLACOther = 3 // FLAC at unspecified or uncommon specs
+ qualityFLAC24_96 = 4
+ qualityFLAC16_44 = 5 // most preferred: standard CD-quality lossless
+)
+
+// albumFolder groups audio files from the same peer and directory path.
+type albumFolder struct {
+ Username string
+ Dir string
+ Files []slskdFile
+ Quality int
+}
+
+func slskdBaseURL() string {
+ return strings.TrimRight(os.Getenv("SLSKD_URL"), "/")
+}
+
+// slskdDo performs an authenticated HTTP request against the slskd API.
+func slskdDo(method, endpoint string, body interface{}) (*http.Response, error) {
+ base := slskdBaseURL()
+ if base == "" {
+ return nil, fmt.Errorf("SLSKD_URL is not configured")
+ }
+
+ var br io.Reader
+ if body != nil {
+ data, err := json.Marshal(body)
+ if err != nil {
+ return nil, err
+ }
+ br = bytes.NewReader(data)
+ }
+
+ req, err := http.NewRequest(method, base+endpoint, br)
+ if err != nil {
+ return nil, err
+ }
+ if key := os.Getenv("SLSKD_API_KEY"); key != "" {
+ req.Header.Set("X-API-Key", key)
+ }
+ if body != nil {
+ req.Header.Set("Content-Type", "application/json")
+ }
+
+ return http.DefaultClient.Do(req)
+}
+
+// createSlskdSearch starts a new slskd search and returns its ID.
+func createSlskdSearch(searchText string) (string, error) {
+ payload := map[string]interface{}{
+ "searchText": searchText,
+ "fileLimit": 10000,
+ "filterResponses": true,
+ "maximumPeerQueueLength": 1000,
+ "minimumPeerUploadSpeed": 0,
+ "responseLimit": 100,
+ "timeout": 15000,
+ }
+
+ resp, err := slskdDo("POST", "/api/v0/searches", payload)
+ if err != nil {
+ return "", err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusCreated && resp.StatusCode != http.StatusOK {
+ b, _ := io.ReadAll(resp.Body)
+ return "", fmt.Errorf("slskd search failed (%d): %s", resp.StatusCode, strings.TrimSpace(string(b)))
+ }
+
+ var s slskdSearch
+ if err := json.NewDecoder(resp.Body).Decode(&s); err != nil {
+ return "", err
+ }
+ return s.ID, nil
+}
+
+// slskdSearchIsTerminal reports whether a slskd SearchStates string has reached
+// a terminal state. slskd serialises its [Flags] enum as a comma-separated list
+// (e.g. "Completed, TimedOut"), so we check for containment rather than equality.
+func slskdSearchIsTerminal(state string) bool {
+ for _, term := range []string{"Completed", "TimedOut", "Errored", "Cancelled"} {
+ if strings.Contains(state, term) {
+ return true
+ }
+ }
+ return false
+}
+
+// pollSlskdSearch waits up to 30 s for a search to reach a terminal state,
+// then returns the responses from the dedicated /responses sub-endpoint.
+// Each poll check-in is reported via logf.
+func pollSlskdSearch(id string, logf func(string)) ([]slskdPeerResponse, error) {
+ deadline := time.Now().Add(60 * time.Second)
+ for {
+ resp, err := slskdDo("GET", "/api/v0/searches/"+id, nil)
+ if err != nil {
+ return nil, err
+ }
+ var s slskdSearch
+ err = json.NewDecoder(resp.Body).Decode(&s)
+ resp.Body.Close()
+ if err != nil {
+ return nil, err
+ }
+
+ logf(fmt.Sprintf("Search state: %s", s.State))
+
+ if slskdSearchIsTerminal(s.State) {
+ return fetchSlskdResponses(id, logf)
+ }
+
+ if time.Now().After(deadline) {
+ logf("Poll deadline reached, fetching current results")
+ return fetchSlskdResponses(id, logf)
+ }
+ time.Sleep(2 * time.Second)
+ }
+}
+
+// fetchSlskdResponses fetches file responses from the dedicated sub-endpoint.
+// The main GET /searches/{id} endpoint only returns metadata; responses live at
+// /searches/{id}/responses.
+func fetchSlskdResponses(id string, logf func(string)) ([]slskdPeerResponse, error) {
+ resp, err := slskdDo("GET", "/api/v0/searches/"+id+"/responses", nil)
+ if err != nil {
+ return nil, err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusOK {
+ b, _ := io.ReadAll(resp.Body)
+ return nil, fmt.Errorf("fetching responses failed (%d): %s", resp.StatusCode, strings.TrimSpace(string(b)))
+ }
+
+ var responses []slskdPeerResponse
+ if err := json.NewDecoder(resp.Body).Decode(&responses); err != nil {
+ return nil, fmt.Errorf("decoding responses: %w", err)
+ }
+ logf(fmt.Sprintf("Fetched %d peer responses", len(responses)))
+ return responses, nil
+}
+
+// deleteSlskdSearch removes a search from slskd (best-effort cleanup).
+func deleteSlskdSearch(id string) {
+ resp, err := slskdDo("DELETE", "/api/v0/searches/"+id, nil)
+ if err == nil {
+ resp.Body.Close()
+ }
+}
+
+// fileDir returns the directory portion of a Soulseek filename,
+// normalising backslashes to forward slashes first.
+func fileDir(filename string) string {
+ return path.Dir(strings.ReplaceAll(filename, "\\", "/"))
+}
+
+// normaliseExt returns a lower-case extension that always starts with ".".
+func normaliseExt(raw string) string {
+ ext := strings.ToLower(raw)
+ if ext != "" && !strings.HasPrefix(ext, ".") {
+ ext = "." + ext
+ }
+ return ext
+}
+
+// fileQuality scores a single file by the preferred quality tier.
+func fileQuality(f slskdFile) int {
+ ext := normaliseExt(f.Extension)
+ if ext == "." || ext == "" {
+ ext = strings.ToLower(path.Ext(strings.ReplaceAll(f.Filename, "\\", "/")))
+ }
+
+ switch ext {
+ case ".flac":
+ var depth, rate int
+ for _, a := range f.Attributes {
+ switch a.Type {
+ case 4:
+ rate = a.Value
+ case 5:
+ depth = a.Value
+ }
+ }
+ if depth == 16 && rate == 44100 {
+ return qualityFLAC16_44
+ }
+ if depth == 24 && rate == 96000 {
+ return qualityFLAC24_96
+ }
+ return qualityFLACOther
+
+ case ".mp3":
+ for _, a := range f.Attributes {
+ if a.Type == 0 && a.Value >= 315 {
+ return qualityMP3_320
+ }
+ }
+ return qualityMP3Any
+ }
+
+ return qualityUnknown
+}
+
+// groupAlbumFolders groups audio files by (username, directory) and scores each group.
+func groupAlbumFolders(responses []slskdPeerResponse) []albumFolder {
+ type key struct{ user, dir string }
+ m := make(map[key]*albumFolder)
+
+ for _, r := range responses {
+ for _, f := range r.Files {
+ ext := normaliseExt(f.Extension)
+ if ext == "." || ext == "" {
+ ext = strings.ToLower(path.Ext(strings.ReplaceAll(f.Filename, "\\", "/")))
+ }
+ if ext != ".flac" && ext != ".mp3" {
+ continue
+ }
+
+ k := key{r.Username, fileDir(f.Filename)}
+ if m[k] == nil {
+ m[k] = &albumFolder{Username: r.Username, Dir: k.dir}
+ }
+ m[k].Files = append(m[k].Files, f)
+ if q := fileQuality(f); q > m[k].Quality {
+ m[k].Quality = q
+ }
+ }
+ }
+
+ out := make([]albumFolder, 0, len(m))
+ for _, af := range m {
+ out = append(out, *af)
+ }
+ return out
+}
+
+// bestAlbumFolder picks the highest-quality folder; file count breaks ties.
+func bestAlbumFolder(folders []albumFolder) *albumFolder {
+ if len(folders) == 0 {
+ return nil
+ }
+ best := &folders[0]
+ for i := 1; i < len(folders); i++ {
+ a := &folders[i]
+ if a.Quality > best.Quality || (a.Quality == best.Quality && len(a.Files) > len(best.Files)) {
+ best = a
+ }
+ }
+ return best
+}
+
+// queueSlskdDownload sends a batch download request to slskd for all files in folder.
+func queueSlskdDownload(folder *albumFolder) error {
+ type dlFile struct {
+ Filename string `json:"filename"`
+ Size int64 `json:"size"`
+ }
+ files := make([]dlFile, len(folder.Files))
+ for i, f := range folder.Files {
+ files[i] = dlFile{Filename: f.Filename, Size: f.Size}
+ }
+
+ resp, err := slskdDo("POST", "/api/v0/transfers/downloads/"+folder.Username, files)
+ if err != nil {
+ return err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusCreated && resp.StatusCode != http.StatusOK {
+ b, _ := io.ReadAll(resp.Body)
+ return fmt.Errorf("slskd download request failed (%d): %s", resp.StatusCode, strings.TrimSpace(string(b)))
+ }
+ return nil
+}
+
+// qualityLabel returns a human-readable label for a quality tier constant.
+func qualityLabel(q int) string {
+ switch q {
+ case qualityFLAC16_44:
+ return "FLAC 16bit/44.1kHz"
+ case qualityFLAC24_96:
+ return "FLAC 24bit/96kHz"
+ case qualityFLACOther:
+ return "FLAC"
+ case qualityMP3_320:
+ return "MP3 320kbps"
+ case qualityMP3Any:
+ return "MP3"
+ default:
+ return "unknown"
+ }
+}
+
+// slskdTransferFile is one file entry in a slskd transfers response.
+type slskdTransferFile struct {
+ Filename string `json:"filename"`
+ LocalFilename string `json:"localFilename"`
+ State string `json:"state"`
+ Size int64 `json:"size"`
+}
+
+// slskdTransferDir groups transfer files by remote directory.
+type slskdTransferDir struct {
+ Directory string `json:"directory"`
+ Files []slskdTransferFile `json:"files"`
+}
+
+// slskdUserTransfers is the object returned by GET /api/v0/transfers/downloads/{username}.
+type slskdUserTransfers struct {
+ Directories []slskdTransferDir `json:"directories"`
+}
+
+// getSlskdTransfers returns all active/pending download transfer directories for a peer.
+func getSlskdTransfers(username string) ([]slskdTransferDir, error) {
+ resp, err := slskdDo("GET", "/api/v0/transfers/downloads/"+username, nil)
+ if err != nil {
+ return nil, err
+ }
+ defer resp.Body.Close()
+
+ if resp.StatusCode != http.StatusOK {
+ b, _ := io.ReadAll(resp.Body)
+ return nil, fmt.Errorf("slskd transfers (%d): %s", resp.StatusCode, strings.TrimSpace(string(b)))
+ }
+
+ var ut slskdUserTransfers
+ if err := json.NewDecoder(resp.Body).Decode(&ut); err != nil {
+ return nil, err
+ }
+ return ut.Directories, nil
+}
+
+// fetchRelease searches slskd for an album, queues the best-quality match for
+// download, and returns the chosen folder so the caller can monitor completion.
+// mbid, if non-empty, will be stored for use during import (beets --search-id).
+// trackCount, if > 0, filters candidate folders to those whose audio file count
+// matches the expected number of tracks on the release, so alternate editions
+// with different track counts are not accidentally selected.
+func fetchRelease(artist, album, mbid string, trackCount int, logf func(string)) (*albumFolder, error) {
+ query := artist + " " + album
+ log.Printf("[discover] fetch started: %q by %s (expected tracks: %d)", album, artist, trackCount)
+ logf("Starting fetch for: " + query)
+
+ logf("Creating slskd search…")
+ id, err := createSlskdSearch(query)
+ if err != nil {
+ return nil, fmt.Errorf("create search: %w", err)
+ }
+ log.Printf("[discover] slskd search created: %s", id)
+ logf(fmt.Sprintf("Search created (id: %s)", id))
+ defer func() {
+ log.Printf("[discover] deleting slskd search %s", id)
+ deleteSlskdSearch(id)
+ }()
+
+ logf("Polling for results…")
+ responses, err := pollSlskdSearch(id, logf)
+ if err != nil {
+ return nil, fmt.Errorf("poll search: %w", err)
+ }
+ log.Printf("[discover] search %s finished: %d peer responses", id, len(responses))
+ logf(fmt.Sprintf("Search finished: %d peer responses received", len(responses)))
+
+ logf("Grouping results into album folders…")
+ folders := groupAlbumFolders(responses)
+ log.Printf("[discover] grouped into %d candidate album folders", len(folders))
+ logf(fmt.Sprintf("Found %d candidate album folders", len(folders)))
+
+ if len(folders) == 0 {
+ return nil, fmt.Errorf("no audio files found for %q by %s", album, artist)
+ }
+
+ // When we know the expected track count, prefer folders that match exactly
+ // so we don't accidentally grab a bonus-track edition or a different version
+ // that won't align with the release MBID we pass to beets.
+ candidates := folders
+ if trackCount > 0 {
+ var matched []albumFolder
+ for _, f := range folders {
+ if len(f.Files) == trackCount {
+ matched = append(matched, f)
+ }
+ }
+ if len(matched) > 0 {
+ log.Printf("[discover] %d/%d folders match expected track count (%d)", len(matched), len(folders), trackCount)
+ logf(fmt.Sprintf("Filtered to %d/%d folders matching expected track count (%d)",
+ len(matched), len(folders), trackCount))
+ candidates = matched
+ } else {
+ log.Printf("[discover] no folders matched expected track count (%d); using best available", trackCount)
+ logf(fmt.Sprintf("Warning: no folders matched expected track count (%d); using best available", trackCount))
+ }
+ }
+
+ best := bestAlbumFolder(candidates)
+ log.Printf("[discover] selected folder: %s from %s (%s, %d files)",
+ best.Dir, best.Username, qualityLabel(best.Quality), len(best.Files))
+ logf(fmt.Sprintf("Selected folder: %s", best.Dir))
+ logf(fmt.Sprintf(" Peer: %s | Quality: %s | Files: %d",
+ best.Username, qualityLabel(best.Quality), len(best.Files)))
+
+ logf(fmt.Sprintf("Queuing %d files for download…", len(best.Files)))
+ if err := queueSlskdDownload(best); err != nil {
+ return nil, fmt.Errorf("queue download: %w", err)
+ }
+ log.Printf("[discover] download queued: %d files from %s", len(best.Files), best.Username)
+ logf("Download queued — waiting for completion before import")
+ return best, nil
+}
diff --git a/static/app.js b/static/app.js
new file mode 100644
index 0000000..e4e4104
--- /dev/null
+++ b/static/app.js
@@ -0,0 +1,313 @@
+"use strict";
+
+// IDs of fetch cards we've already created, so we don't duplicate them.
+const knownFetchIds = new Set();
+
+document.addEventListener("DOMContentLoaded", () => {
+ initTabs();
+ initSearch();
+ initFetchList();
+});
+
+// ── Tabs ───────────────────────────────────────────────────────────────────────
+
+function initTabs() {
+ document.querySelector(".tabs").addEventListener("click", (e) => {
+ const btn = e.target.closest(".tab-btn");
+ if (!btn) return;
+ showTab(btn.dataset.tab);
+ });
+}
+
+function showTab(name) {
+ document
+ .querySelectorAll(".tab-pane")
+ .forEach((p) => p.classList.remove("active"));
+ document
+ .querySelectorAll(".tab-btn")
+ .forEach((b) => b.classList.remove("active"));
+ document.getElementById("tab-" + name).classList.add("active");
+ document
+ .querySelector(`.tab-btn[data-tab="${name}"]`)
+ .classList.add("active");
+}
+
+// ── Search ─────────────────────────────────────────────────────────────────────
+
+let searchType = "release";
+
+function initSearch() {
+ document.querySelector(".type-toggle").addEventListener("click", (e) => {
+ const btn = e.target.closest(".type-btn");
+ if (btn) setSearchType(btn.dataset.type);
+ });
+
+ const searchBtn = document.getElementById("search-btn");
+ const searchInput = document.getElementById("search-q");
+ searchBtn.addEventListener("click", doSearch);
+ searchInput.addEventListener("keydown", (e) => {
+ if (e.key === "Enter") doSearch();
+ });
+
+ // Event delegation for dynamically rendered result buttons
+ document.getElementById("search-results").addEventListener("click", (e) => {
+ const btn = e.target.closest(".fetch-btn");
+ if (!btn || btn.disabled) return;
+ if (btn.dataset.fetchType === "artist") startArtistFetch(btn);
+ else startReleaseFetch(btn);
+ });
+}
+
+function setSearchType(type) {
+ searchType = type;
+ document.querySelectorAll(".type-btn").forEach((b) => {
+ b.classList.toggle("active", b.dataset.type === type);
+ });
+}
+
+function doSearch() {
+ const q = document.getElementById("search-q").value.trim();
+ if (!q) return;
+
+ const btn = document.getElementById("search-btn");
+ const resultsEl = document.getElementById("search-results");
+
+ btn.disabled = true;
+ btn.textContent = "Searching\u2026";
+ resultsEl.innerHTML = 'Searching MusicBrainz\u2026
';
+
+ fetch(`/discover/search?q=${encodeURIComponent(q)}&type=${searchType}`)
+ .then((r) => {
+ if (!r.ok)
+ return r.text().then((t) => {
+ throw new Error(t || r.statusText);
+ });
+ return r.json();
+ })
+ .then((data) => renderResults(data))
+ .catch((err) => {
+ resultsEl.innerHTML = `Error: ${esc(err.message)}
`;
+ })
+ .finally(() => {
+ btn.disabled = false;
+ btn.textContent = "Search";
+ });
+}
+
+// ── Results rendering ──────────────────────────────────────────────────────────
+
+function renderResults(data) {
+ const el = document.getElementById("search-results");
+ if (!data || data.length === 0) {
+ el.innerHTML = 'No results found.
';
+ return;
+ }
+ const renderer = searchType === "artist" ? renderArtist : renderRelease;
+ el.innerHTML = data.map(renderer).join("");
+}
+
+function renderRelease(r) {
+ const credits = r["artist-credit"] ?? [];
+ const artist =
+ credits.map((c) => c.name || c.artist?.name || "").join("") ||
+ "Unknown Artist";
+ const year = r.date?.substring(0, 4) ?? "";
+ const type = r["release-group"]?.["primary-type"] ?? "";
+ const country = r.country ?? "";
+ const formats = [
+ ...new Set((r.media ?? []).map((m) => m.format).filter(Boolean)),
+ ].join("+");
+ const lang = r["text-representation"]?.language ?? "";
+ const meta = [year, type, formats, country, lang]
+ .filter(Boolean)
+ .join(" \u00b7 ");
+ const dis = r.disambiguation ? ` (${esc(r.disambiguation)})` : "";
+ const coverUrl = `https://coverartarchive.org/release/${r.id}/front-250`;
+
+ return `
+
+

+
+ ${esc(artist)} \u2014 ${esc(r.title)}${dis}
+ ${meta ? `${esc(meta)}` : ""}
+
+
+
`;
+}
+
+function renderArtist(a) {
+ const dis = a.disambiguation ? ` (${esc(a.disambiguation)})` : "";
+ return `
+
+
+ ${esc(a.name)}${dis}
+ ${a.country ? `${esc(a.country)}` : ""}
+
+
+
`;
+}
+
+// ── Fetch operations ───────────────────────────────────────────────────────────
+
+function startReleaseFetch(btn) {
+ const { id, artist, album } = btn.dataset;
+ btn.disabled = true;
+ btn.textContent = "Fetching\u2026";
+
+ fetch("/discover/fetch", {
+ method: "POST",
+ headers: { "Content-Type": "application/json" },
+ body: JSON.stringify({ id, artist, album }),
+ })
+ .then((r) => {
+ if (!r.ok)
+ return r.text().then((t) => {
+ throw new Error(t || r.statusText);
+ });
+ return r.json();
+ })
+ .then(() => {
+ addFetchCard(id, `${artist} \u2014 ${album}`);
+ pollFetch(id);
+ })
+ .catch((err) => {
+ btn.disabled = false;
+ btn.textContent = "Fetch";
+ showFetchError(err.message);
+ });
+}
+
+function startArtistFetch(btn) {
+ const { id, name } = btn.dataset;
+ btn.disabled = true;
+ btn.textContent = "Fetching\u2026";
+
+ fetch("/discover/fetch/artist", {
+ method: "POST",
+ headers: { "Content-Type": "application/json" },
+ body: JSON.stringify({ id, name }),
+ })
+ .then((r) => {
+ if (!r.ok)
+ return r.text().then((t) => {
+ throw new Error(t || r.statusText);
+ });
+ return r.json();
+ })
+ .then(() => {
+ addFetchCard(id, `${name} \u2014 full discography`);
+ pollFetch(id);
+ })
+ .catch((err) => {
+ btn.disabled = false;
+ btn.textContent = "Fetch All";
+ showFetchError(err.message);
+ });
+}
+
+// ── Fetch cards ────────────────────────────────────────────────────────────────
+
+function addFetchCard(id, title) {
+ knownFetchIds.add(id);
+ const list = document.getElementById("fetch-list");
+ const card = document.createElement("div");
+ card.className = "fetch-card";
+ card.id = `fetch-${id}`;
+ card.innerHTML = `
+
+ `;
+ list.prepend(card);
+}
+
+function pollFetch(id) {
+ fetch(`/discover/fetch/status?id=${encodeURIComponent(id)}`)
+ .then((r) => r.json())
+ .then((data) => {
+ const logEl = document.getElementById(`flog-${id}`);
+ const statusEl = document.getElementById(`fstatus-${id}`);
+ const card = document.getElementById(`fetch-${id}`);
+
+ if (logEl && data.log) {
+ logEl.innerHTML = data.log
+ .map((l) => `${esc(l)}
`)
+ .join("");
+ logEl.scrollTop = logEl.scrollHeight;
+ }
+
+ if (data.done) {
+ if (data.success) {
+ statusEl?.setAttribute("class", "fetch-status fetch-status-ok");
+ if (statusEl) statusEl.textContent = "\u2713 done";
+ card?.classList.add("fetch-card-ok");
+ } else {
+ statusEl?.setAttribute("class", "fetch-status fetch-status-err");
+ if (statusEl) statusEl.textContent = "\u2717 failed";
+ card?.classList.add("fetch-card-err");
+ if (data.error && logEl) {
+ logEl.innerHTML += `${esc(data.error)}
`;
+ logEl.scrollTop = logEl.scrollHeight;
+ }
+ }
+ } else {
+ setTimeout(() => pollFetch(id), 2000);
+ }
+ })
+ .catch(() => setTimeout(() => pollFetch(id), 3000));
+}
+
+// ── Fetch list polling ─────────────────────────────────────────────────────────
+
+// Polls /discover/fetch/list every 5 s to discover server-created fetch entries
+// (e.g. per-album cards spawned during an artist fetch) and create cards for them.
+function initFetchList() {
+ pollFetchList();
+}
+
+function pollFetchList() {
+ fetch("/discover/fetch/list")
+ .then((r) => (r.ok ? r.json() : null))
+ .then((items) => {
+ if (!items) return;
+ for (const item of items) {
+ if (!knownFetchIds.has(item.id)) {
+ knownFetchIds.add(item.id);
+ addFetchCard(item.id, item.title);
+ if (!item.done) pollFetch(item.id);
+ }
+ }
+ })
+ .catch(() => {})
+ .finally(() => setTimeout(pollFetchList, 5000));
+}
+
+// ── Utilities ──────────────────────────────────────────────────────────────────
+
+function showFetchError(msg) {
+ const list = document.getElementById("fetch-list");
+ const el = document.createElement("div");
+ el.className = "fetch-card fetch-card-err";
+ el.innerHTML = `
+ `;
+ list.prepend(el);
+}
+
+function esc(s) {
+ return String(s ?? "")
+ .replace(/&/g, "&")
+ .replace(//g, ">")
+ .replace(/"/g, """);
+}
diff --git a/static/style.css b/static/style.css
new file mode 100644
index 0000000..4ab9e1a
--- /dev/null
+++ b/static/style.css
@@ -0,0 +1,585 @@
+/* ── Custom properties ────────────────────────────────────────────────────── */
+
+:root {
+ --bg: #111;
+ --surface: #1a1a1a;
+ --surface-hi: #222;
+ --border: #2a2a2a;
+ --border-focus: #555;
+
+ --text: #eee;
+ --text-secondary: #aaa;
+ --text-muted: #777;
+ --text-dim: #555;
+
+ --green: #4caf50;
+ --green-bg: #1e4d2b;
+ --green-hover: #1e3d1e;
+ --green-border: #3a7a3a;
+ --amber: #f0a500;
+ --amber-bg: #4d3a00;
+ --red: #e05050;
+ --red-bg: #4d1a1a;
+ --red-text: #c0392b;
+
+ --pill-beets: #7ec8e3;
+ --pill-mb: #c084fc;
+ --pill-tags: #f0a500;
+
+ --radius-lg: 8px;
+ --radius: 6px;
+ --radius-sm: 5px;
+ --radius-xs: 4px;
+
+ --max-w: 860px;
+ --pad-x: 24px;
+}
+
+/* ── Reset & base ─────────────────────────────────────────────────────────── */
+
+*,
+*::before,
+*::after {
+ box-sizing: border-box;
+}
+
+body {
+ font-family:
+ system-ui,
+ -apple-system,
+ sans-serif;
+ background: var(--bg);
+ color: var(--text);
+ margin: 0;
+ padding: 48px var(--pad-x) 80px;
+ text-align: center;
+}
+
+h1 {
+ margin: 0 0 24px;
+ font-size: clamp(20px, 4vw, 28px);
+}
+
+/* ── Tabs ─────────────────────────────────────────────────────────────────── */
+
+.tabs {
+ display: inline-flex;
+ gap: 4px;
+ background: var(--surface);
+ border: 1px solid var(--border);
+ border-radius: var(--radius-lg);
+ padding: 4px;
+ margin-bottom: 36px;
+}
+
+.tab-btn {
+ font-size: 14px;
+ min-height: 36px;
+ padding: 0 24px;
+ border-radius: var(--radius);
+ border: none;
+ cursor: pointer;
+ background: transparent;
+ color: var(--text-muted);
+ transition:
+ background 0.15s,
+ color 0.15s;
+}
+.tab-btn.active {
+ background: var(--surface-hi);
+ color: var(--text);
+}
+
+.tab-pane {
+ display: none;
+}
+.tab-pane.active {
+ display: block;
+}
+
+/* ── Shared card / content container ─────────────────────────────────────── */
+
+.content-box {
+ max-width: var(--max-w);
+ margin: 0 auto;
+ text-align: left;
+}
+
+/* ── Import tab — run button ─────────────────────────────────────────────── */
+
+.run-btn {
+ font-size: clamp(18px, 4vw, 28px);
+ padding: 18px 40px;
+ border-radius: 10px;
+ border: none;
+ cursor: pointer;
+ background: var(--green);
+ color: #fff;
+ transition: opacity 0.15s;
+}
+.run-btn:hover:not(:disabled) {
+ opacity: 0.88;
+}
+.run-btn:disabled {
+ background: #555;
+ cursor: not-allowed;
+}
+
+/* ── Import tab — session summary ────────────────────────────────────────── */
+
+.session {
+ margin-top: 48px;
+}
+
+.session-header {
+ display: flex;
+ justify-content: space-between;
+ align-items: baseline;
+ flex-wrap: wrap;
+ gap: 4px;
+ border-bottom: 1px solid #333;
+ padding-bottom: 8px;
+ margin-bottom: 20px;
+}
+.session-header h2 {
+ margin: 0;
+ font-size: 16px;
+ color: var(--text-secondary);
+}
+.session-header .duration {
+ font-size: 13px;
+ color: var(--text-dim);
+}
+
+/* ── Album card ───────────────────────────────────────────────────────────── */
+
+.album {
+ background: var(--surface);
+ border: 1px solid var(--border);
+ border-radius: var(--radius-lg);
+ padding: 16px 20px;
+ margin-bottom: 12px;
+}
+
+.album-header {
+ display: flex;
+ align-items: center;
+ gap: 10px;
+ margin-bottom: 10px;
+ flex-wrap: wrap;
+}
+.album-name {
+ font-weight: 600;
+ font-size: 15px;
+ flex: 1;
+ min-width: 0;
+ white-space: nowrap;
+ overflow: hidden;
+ text-overflow: ellipsis;
+}
+
+.badge {
+ font-size: 11px;
+ font-weight: 700;
+ padding: 2px 8px;
+ border-radius: var(--radius-xs);
+ white-space: nowrap;
+ flex-shrink: 0;
+}
+.badge-ok {
+ background: var(--green-bg);
+ color: var(--green);
+}
+.badge-warn {
+ background: var(--amber-bg);
+ color: var(--amber);
+}
+.badge-fatal {
+ background: var(--red-bg);
+ color: var(--red);
+}
+
+/* ── Metadata row ─────────────────────────────────────────────────────────── */
+
+.metadata {
+ display: flex;
+ align-items: baseline;
+ flex-wrap: wrap;
+ gap: 10px;
+ font-size: 12px;
+ color: var(--text-muted);
+ margin-bottom: 12px;
+}
+.metadata-title {
+ color: var(--text-secondary);
+ font-size: 13px;
+}
+
+.metadata-pill {
+ display: inline-flex;
+ align-items: center;
+ gap: 4px;
+ background: var(--surface-hi);
+ border-radius: var(--radius-xs);
+ padding: 2px 7px;
+ font-size: 11px;
+}
+.pill-label {
+ color: var(--text-dim);
+}
+.pill-beets {
+ color: var(--pill-beets);
+}
+.pill-musicbrainz {
+ color: var(--pill-mb);
+}
+.pill-file_tags {
+ color: var(--pill-tags);
+}
+.pill-unknown {
+ color: #888;
+}
+
+/* ── Info grid ────────────────────────────────────────────────────────────── */
+
+.info-grid {
+ display: grid;
+ grid-template-columns: repeat(auto-fill, minmax(160px, 1fr));
+ gap: 6px;
+ margin-bottom: 12px;
+}
+.info-card {
+ background: var(--surface-hi);
+ border-radius: var(--radius);
+ padding: 8px 12px;
+ font-size: 12px;
+}
+.info-card-label {
+ font-size: 10px;
+ text-transform: uppercase;
+ letter-spacing: 0.06em;
+ color: var(--text-dim);
+ margin-bottom: 4px;
+}
+.info-card-value {
+ color: var(--text-secondary);
+ font-size: 13px;
+ font-weight: 600;
+}
+.info-card-sub {
+ margin-top: 3px;
+ color: var(--text-dim);
+ font-size: 11px;
+ line-height: 1.4;
+}
+
+.info-ok {
+ color: var(--green);
+}
+.info-warn {
+ color: var(--amber);
+}
+.info-dim {
+ color: var(--text-dim);
+}
+
+/* ── Pipeline steps ───────────────────────────────────────────────────────── */
+
+.steps-label {
+ font-size: 10px;
+ text-transform: uppercase;
+ letter-spacing: 0.06em;
+ color: #444;
+ margin-bottom: 6px;
+}
+.steps {
+ display: grid;
+ grid-template-columns: repeat(auto-fill, minmax(180px, 1fr));
+ gap: 6px;
+}
+.step {
+ font-size: 12px;
+ padding: 5px 10px;
+ border-radius: var(--radius-sm);
+ background: var(--surface-hi);
+ display: flex;
+ flex-direction: column;
+ gap: 2px;
+}
+.step-label {
+ color: #888;
+}
+.step-ok {
+ color: var(--green);
+}
+.step-warn {
+ color: var(--amber);
+}
+.step-fatal {
+ color: var(--red);
+}
+.step-err {
+ font-size: 11px;
+ color: var(--red-text);
+ margin-top: 2px;
+ word-break: break-word;
+}
+
+/* ── Discover tab — search form ───────────────────────────────────────────── */
+
+.search-form {
+ display: flex;
+ gap: 8px;
+ align-items: stretch;
+ margin-bottom: 20px;
+}
+
+.type-toggle {
+ display: flex;
+ border: 1px solid #333;
+ border-radius: var(--radius);
+ overflow: hidden;
+ flex-shrink: 0;
+}
+.type-btn {
+ font-size: 13px;
+ padding: 0 16px;
+ border: none;
+ background: var(--surface);
+ color: var(--text-muted);
+ cursor: pointer;
+ transition:
+ background 0.15s,
+ color 0.15s;
+ white-space: nowrap;
+}
+.type-btn.active {
+ background: var(--surface-hi);
+ color: var(--text);
+}
+
+.search-input {
+ flex: 1;
+ min-width: 0;
+ font-size: 14px;
+ padding: 0 12px;
+ height: 38px;
+ background: var(--surface);
+ border: 1px solid #333;
+ border-radius: var(--radius);
+ color: var(--text);
+ outline: none;
+ transition: border-color 0.15s;
+}
+.search-input:focus {
+ border-color: var(--border-focus);
+}
+
+.search-btn {
+ font-size: 14px;
+ padding: 0 20px;
+ height: 38px;
+ border-radius: var(--radius);
+ border: none;
+ background: var(--green);
+ color: #fff;
+ cursor: pointer;
+ flex-shrink: 0;
+ transition: opacity 0.15s;
+}
+.search-btn:hover:not(:disabled) {
+ opacity: 0.88;
+}
+.search-btn:disabled {
+ background: #555;
+ cursor: not-allowed;
+}
+
+/* ── Discover tab — search results ───────────────────────────────────────── */
+
+.search-msg {
+ text-align: center;
+ color: var(--text-dim);
+ font-size: 14px;
+ padding: 32px 0;
+}
+.search-msg.error {
+ color: var(--red);
+}
+
+.result-row {
+ display: flex;
+ align-items: center;
+ gap: 12px;
+ background: var(--surface);
+ border: 1px solid var(--border);
+ border-radius: var(--radius-lg);
+ padding: 12px 16px;
+ margin-bottom: 8px;
+}
+.result-cover {
+ width: 48px;
+ height: 48px;
+ object-fit: cover;
+ border-radius: var(--radius-sm);
+ flex-shrink: 0;
+ background: var(--surface-hi);
+}
+.result-info {
+ flex: 1;
+ min-width: 0;
+}
+.result-title {
+ display: block;
+ font-size: 14px;
+ color: #ddd;
+ white-space: nowrap;
+ overflow: hidden;
+ text-overflow: ellipsis;
+}
+.result-dis {
+ color: var(--text-dim);
+}
+.result-meta {
+ display: block;
+ font-size: 12px;
+ color: var(--text-dim);
+ margin-top: 2px;
+}
+
+.fetch-btn {
+ font-size: 12px;
+ padding: 5px 14px;
+ border-radius: var(--radius-sm);
+ border: 1px solid var(--green-border);
+ background: transparent;
+ color: var(--green);
+ cursor: pointer;
+ flex-shrink: 0;
+ white-space: nowrap;
+ transition: background 0.15s;
+}
+.fetch-btn:hover:not(:disabled) {
+ background: var(--green-hover);
+}
+.fetch-btn:disabled {
+ border-color: #333;
+ color: var(--text-dim);
+ cursor: not-allowed;
+}
+
+/* ── Discover tab — fetch log cards ───────────────────────────────────────── */
+
+.fetch-list {
+ margin-top: 32px;
+}
+
+.fetch-card {
+ background: var(--surface);
+ border: 1px solid var(--border);
+ border-radius: var(--radius-lg);
+ padding: 14px 16px;
+ margin-bottom: 10px;
+ transition: border-color 0.3s;
+}
+.fetch-card-ok {
+ border-color: var(--green-bg);
+}
+.fetch-card-err {
+ border-color: var(--red-bg);
+}
+
+.fetch-header {
+ display: flex;
+ justify-content: space-between;
+ align-items: center;
+ flex-wrap: wrap;
+ gap: 6px;
+ margin-bottom: 8px;
+}
+.fetch-title {
+ font-size: 14px;
+ font-weight: 600;
+ color: var(--text-secondary);
+ min-width: 0;
+ overflow: hidden;
+ text-overflow: ellipsis;
+ white-space: nowrap;
+}
+.fetch-status {
+ font-size: 12px;
+ color: var(--text-dim);
+ flex-shrink: 0;
+}
+.fetch-status-ok {
+ color: var(--green);
+}
+.fetch-status-err {
+ color: var(--red);
+}
+
+.fetch-log {
+ font-size: 12px;
+ font-family: ui-monospace, "Cascadia Code", "Fira Mono", monospace;
+ color: var(--text-muted);
+ max-height: 260px;
+ overflow-y: auto;
+ scrollbar-width: thin;
+ scrollbar-color: #333 transparent;
+}
+.log-line {
+ padding: 1px 0;
+ line-height: 1.5;
+}
+.log-line-err {
+ color: var(--red-text);
+}
+
+/* ── Footer ───────────────────────────────────────────────────────────────── */
+
+footer {
+ position: fixed;
+ bottom: 14px;
+ left: 0;
+ width: 100%;
+ font-size: 12px;
+ color: #444;
+ text-align: center;
+ pointer-events: none;
+}
+
+/* ── Responsive ───────────────────────────────────────────────────────────── */
+
+@media (max-width: 600px) {
+ body {
+ padding: 32px 16px 72px;
+ }
+
+ .tabs {
+ display: flex;
+ width: 100%;
+ }
+ .tab-btn {
+ flex: 1;
+ padding: 0;
+ min-height: 40px;
+ }
+
+ .search-form {
+ flex-wrap: wrap;
+ }
+ .type-toggle {
+ width: 100%;
+ }
+ .type-btn {
+ flex: 1;
+ min-height: 38px;
+ }
+ .search-btn {
+ width: 100%;
+ }
+
+ .result-title {
+ white-space: normal;
+ }
+}