mirror of
https://github.com/navidrome/navidrome.git
synced 2026-02-02 06:24:14 +00:00
The CacheWarmer was failing with data not found errors because PreCache was being called inside the database transaction before the data was committed. The CacheWarmer runs in a separate goroutine with its own database context and could not access the uncommitted data due to transaction isolation. Changed the persistChanges method in phase_1_folders.go to collect artwork IDs during the transaction and only call PreCache after the transaction successfully commits. This ensures the artwork data is visible to the CacheWarmer when it attempts to retrieve and cache the images. The fix eliminates the data not found errors and allows the cache warmer to properly pre-cache album and artist artwork during library scanning. Signed-off-by: Deluan <deluan@navidrome.org>
502 lines
18 KiB
Go
502 lines
18 KiB
Go
package scanner
|
|
|
|
import (
|
|
"cmp"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"maps"
|
|
"path"
|
|
"slices"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/Masterminds/squirrel"
|
|
ppl "github.com/google/go-pipeline/pkg/pipeline"
|
|
"github.com/navidrome/navidrome/conf"
|
|
"github.com/navidrome/navidrome/consts"
|
|
"github.com/navidrome/navidrome/core/artwork"
|
|
"github.com/navidrome/navidrome/core/storage"
|
|
"github.com/navidrome/navidrome/log"
|
|
"github.com/navidrome/navidrome/model"
|
|
"github.com/navidrome/navidrome/model/metadata"
|
|
"github.com/navidrome/navidrome/utils"
|
|
"github.com/navidrome/navidrome/utils/pl"
|
|
"github.com/navidrome/navidrome/utils/slice"
|
|
)
|
|
|
|
func createPhaseFolders(ctx context.Context, state *scanState, ds model.DataStore, cw artwork.CacheWarmer) *phaseFolders {
|
|
var jobs []*scanJob
|
|
|
|
// Create scan jobs for all libraries
|
|
for _, lib := range state.libraries {
|
|
// Get target folders for this library if selective scan
|
|
var targetFolders []string
|
|
if state.isSelectiveScan() {
|
|
targetFolders = state.targets[lib.ID]
|
|
}
|
|
|
|
job, err := newScanJob(ctx, ds, cw, lib, state.fullScan, targetFolders)
|
|
if err != nil {
|
|
log.Error(ctx, "Scanner: Error creating scan context", "lib", lib.Name, err)
|
|
state.sendWarning(err.Error())
|
|
continue
|
|
}
|
|
jobs = append(jobs, job)
|
|
}
|
|
|
|
return &phaseFolders{jobs: jobs, ctx: ctx, ds: ds, state: state}
|
|
}
|
|
|
|
type scanJob struct {
|
|
lib model.Library
|
|
fs storage.MusicFS
|
|
cw artwork.CacheWarmer
|
|
lastUpdates map[string]model.FolderUpdateInfo // Holds last update info for all (DB) folders in this library
|
|
targetFolders []string // Specific folders to scan (including all descendants)
|
|
lock sync.Mutex
|
|
numFolders atomic.Int64
|
|
}
|
|
|
|
func newScanJob(ctx context.Context, ds model.DataStore, cw artwork.CacheWarmer, lib model.Library, fullScan bool, targetFolders []string) (*scanJob, error) {
|
|
// Get folder updates, optionally filtered to specific target folders
|
|
lastUpdates, err := ds.Folder(ctx).GetFolderUpdateInfo(lib, targetFolders...)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("getting last updates: %w", err)
|
|
}
|
|
|
|
fileStore, err := storage.For(lib.Path)
|
|
if err != nil {
|
|
log.Error(ctx, "Error getting storage for library", "library", lib.Name, "path", lib.Path, err)
|
|
return nil, fmt.Errorf("getting storage for library: %w", err)
|
|
}
|
|
fsys, err := fileStore.FS()
|
|
if err != nil {
|
|
log.Error(ctx, "Error getting fs for library", "library", lib.Name, "path", lib.Path, err)
|
|
return nil, fmt.Errorf("getting fs for library: %w", err)
|
|
}
|
|
return &scanJob{
|
|
lib: lib,
|
|
fs: fsys,
|
|
cw: cw,
|
|
lastUpdates: lastUpdates,
|
|
targetFolders: targetFolders,
|
|
}, nil
|
|
}
|
|
|
|
// popLastUpdate retrieves and removes the last update info for the given folder ID
|
|
// This is used to track which folders have been found during the walk_dir_tree
|
|
func (j *scanJob) popLastUpdate(folderID string) model.FolderUpdateInfo {
|
|
j.lock.Lock()
|
|
defer j.lock.Unlock()
|
|
|
|
lastUpdate := j.lastUpdates[folderID]
|
|
delete(j.lastUpdates, folderID)
|
|
return lastUpdate
|
|
}
|
|
|
|
// createFolderEntry creates a new folderEntry for the given path, using the last update info from the job
|
|
// to populate the previous update time and hash. It also removes the folder from the job's lastUpdates map.
|
|
// This is used to track which folders have been found during the walk_dir_tree.
|
|
func (j *scanJob) createFolderEntry(path string) *folderEntry {
|
|
id := model.FolderID(j.lib, path)
|
|
info := j.popLastUpdate(id)
|
|
return newFolderEntry(j, id, path, info.UpdatedAt, info.Hash)
|
|
}
|
|
|
|
// phaseFolders represents the first phase of the scanning process, which is responsible
|
|
// for scanning all libraries and importing new or updated files. This phase involves
|
|
// traversing the directory tree of each library, identifying new or modified media files,
|
|
// and updating the database with the relevant information.
|
|
//
|
|
// The phaseFolders struct holds the context, data store, and jobs required for the scanning
|
|
// process. Each job represents a library being scanned, and contains information about the
|
|
// library, file system, and the last updates of the folders.
|
|
//
|
|
// The phaseFolders struct implements the phase interface, providing methods to produce
|
|
// folder entries, process folders, persist changes to the database, and log the results.
|
|
type phaseFolders struct {
|
|
jobs []*scanJob
|
|
ds model.DataStore
|
|
ctx context.Context
|
|
state *scanState
|
|
prevAlbumPIDConf string
|
|
}
|
|
|
|
func (p *phaseFolders) description() string {
|
|
return "Scan all libraries and import new/updated files"
|
|
}
|
|
|
|
func (p *phaseFolders) producer() ppl.Producer[*folderEntry] {
|
|
return ppl.NewProducer(func(put func(entry *folderEntry)) error {
|
|
var err error
|
|
p.prevAlbumPIDConf, err = p.ds.Property(p.ctx).DefaultGet(consts.PIDAlbumKey, "")
|
|
if err != nil {
|
|
return fmt.Errorf("getting album PID conf: %w", err)
|
|
}
|
|
|
|
// TODO Parallelize multiple job when we have multiple libraries
|
|
var total int64
|
|
var totalChanged int64
|
|
for _, job := range p.jobs {
|
|
if utils.IsCtxDone(p.ctx) {
|
|
break
|
|
}
|
|
|
|
outputChan, err := walkDirTree(p.ctx, job, job.targetFolders...)
|
|
if err != nil {
|
|
log.Warn(p.ctx, "Scanner: Error scanning library", "lib", job.lib.Name, err)
|
|
}
|
|
for folder := range pl.ReadOrDone(p.ctx, outputChan) {
|
|
job.numFolders.Add(1)
|
|
p.state.sendProgress(&ProgressInfo{
|
|
LibID: job.lib.ID,
|
|
FileCount: uint32(len(folder.audioFiles)),
|
|
Path: folder.path,
|
|
Phase: "1",
|
|
})
|
|
|
|
// Log folder info
|
|
log.Trace(p.ctx, "Scanner: Checking folder state", " folder", folder.path, "_updTime", folder.updTime,
|
|
"_modTime", folder.modTime, "_lastScanStartedAt", folder.job.lib.LastScanStartedAt,
|
|
"numAudioFiles", len(folder.audioFiles), "numImageFiles", len(folder.imageFiles),
|
|
"numPlaylists", folder.numPlaylists, "numSubfolders", folder.numSubFolders)
|
|
|
|
// Check if folder is outdated
|
|
if folder.isOutdated() {
|
|
if !p.state.fullScan {
|
|
if folder.hasNoFiles() && folder.isNew() {
|
|
log.Trace(p.ctx, "Scanner: Skipping new folder with no files", "folder", folder.path, "lib", job.lib.Name)
|
|
continue
|
|
}
|
|
log.Debug(p.ctx, "Scanner: Detected changes in folder", "folder", folder.path, "lastUpdate", folder.modTime, "lib", job.lib.Name)
|
|
}
|
|
totalChanged++
|
|
folder.elapsed.Stop()
|
|
put(folder)
|
|
} else {
|
|
log.Trace(p.ctx, "Scanner: Skipping up-to-date folder", "folder", folder.path, "lastUpdate", folder.modTime, "lib", job.lib.Name)
|
|
}
|
|
}
|
|
total += job.numFolders.Load()
|
|
}
|
|
log.Debug(p.ctx, "Scanner: Finished loading all folders", "numFolders", total, "numChanged", totalChanged)
|
|
return nil
|
|
}, ppl.Name("traverse filesystem"))
|
|
}
|
|
|
|
func (p *phaseFolders) measure(entry *folderEntry) func() time.Duration {
|
|
entry.elapsed.Start()
|
|
return func() time.Duration { return entry.elapsed.Stop() }
|
|
}
|
|
|
|
func (p *phaseFolders) stages() []ppl.Stage[*folderEntry] {
|
|
return []ppl.Stage[*folderEntry]{
|
|
ppl.NewStage(p.processFolder, ppl.Name("process folder"), ppl.Concurrency(conf.Server.DevScannerThreads)),
|
|
ppl.NewStage(p.persistChanges, ppl.Name("persist changes")),
|
|
ppl.NewStage(p.logFolder, ppl.Name("log results")),
|
|
}
|
|
}
|
|
|
|
func (p *phaseFolders) processFolder(entry *folderEntry) (*folderEntry, error) {
|
|
defer p.measure(entry)()
|
|
|
|
// Load children mediafiles from DB
|
|
cursor, err := p.ds.MediaFile(p.ctx).GetCursor(model.QueryOptions{
|
|
Filters: squirrel.And{squirrel.Eq{"folder_id": entry.id}},
|
|
})
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error loading mediafiles from DB", "folder", entry.path, err)
|
|
return entry, err
|
|
}
|
|
dbTracks := make(map[string]*model.MediaFile)
|
|
for mf, err := range cursor {
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error loading mediafiles from DB", "folder", entry.path, err)
|
|
return entry, err
|
|
}
|
|
dbTracks[mf.Path] = &mf
|
|
}
|
|
|
|
// Get list of files to import, based on modtime (or all if fullScan),
|
|
// leave in dbTracks only tracks that are missing (not found in the FS)
|
|
filesToImport := make(map[string]*model.MediaFile, len(entry.audioFiles))
|
|
for afPath, af := range entry.audioFiles {
|
|
fullPath := path.Join(entry.path, afPath)
|
|
dbTrack, foundInDB := dbTracks[fullPath]
|
|
if !foundInDB || p.state.fullScan {
|
|
filesToImport[fullPath] = dbTrack
|
|
} else {
|
|
info, err := af.Info()
|
|
if err != nil {
|
|
log.Warn(p.ctx, "Scanner: Error getting file info", "folder", entry.path, "file", af.Name(), err)
|
|
p.state.sendWarning(fmt.Sprintf("Error getting file info for %s/%s: %v", entry.path, af.Name(), err))
|
|
return entry, nil
|
|
}
|
|
if info.ModTime().After(dbTrack.UpdatedAt) || dbTrack.Missing {
|
|
filesToImport[fullPath] = dbTrack
|
|
}
|
|
}
|
|
delete(dbTracks, fullPath)
|
|
}
|
|
|
|
// Remaining dbTracks are tracks that were not found in the FS, so they should be marked as missing
|
|
entry.missingTracks = slices.Collect(maps.Values(dbTracks))
|
|
|
|
// Load metadata from files that need to be imported
|
|
if len(filesToImport) > 0 {
|
|
err = p.loadTagsFromFiles(entry, filesToImport)
|
|
if err != nil {
|
|
log.Warn(p.ctx, "Scanner: Error loading tags from files. Skipping", "folder", entry.path, err)
|
|
p.state.sendWarning(fmt.Sprintf("Error loading tags from files in %s: %v", entry.path, err))
|
|
return entry, nil
|
|
}
|
|
|
|
p.createAlbumsFromMediaFiles(entry)
|
|
p.createArtistsFromMediaFiles(entry)
|
|
}
|
|
|
|
return entry, nil
|
|
}
|
|
|
|
const filesBatchSize = 200
|
|
|
|
// loadTagsFromFiles reads metadata from the files in the given list and populates
|
|
// the entry's tracks and tags with the results.
|
|
func (p *phaseFolders) loadTagsFromFiles(entry *folderEntry, toImport map[string]*model.MediaFile) error {
|
|
tracks := make([]model.MediaFile, 0, len(toImport))
|
|
uniqueTags := make(map[string]model.Tag, len(toImport))
|
|
for chunk := range slice.CollectChunks(maps.Keys(toImport), filesBatchSize) {
|
|
allInfo, err := entry.job.fs.ReadTags(chunk...)
|
|
if err != nil {
|
|
log.Warn(p.ctx, "Scanner: Error extracting metadata from files. Skipping", "folder", entry.path, err)
|
|
return err
|
|
}
|
|
for filePath, info := range allInfo {
|
|
md := metadata.New(filePath, info)
|
|
track := md.ToMediaFile(entry.job.lib.ID, entry.id)
|
|
tracks = append(tracks, track)
|
|
for _, t := range track.Tags.FlattenAll() {
|
|
uniqueTags[t.ID] = t
|
|
}
|
|
|
|
// Keep track of any album ID changes, to reassign annotations later
|
|
prevAlbumID := ""
|
|
if prev := toImport[filePath]; prev != nil {
|
|
prevAlbumID = prev.AlbumID
|
|
} else {
|
|
prevAlbumID = md.AlbumID(track, p.prevAlbumPIDConf)
|
|
}
|
|
_, ok := entry.albumIDMap[track.AlbumID]
|
|
if prevAlbumID != track.AlbumID && !ok {
|
|
entry.albumIDMap[track.AlbumID] = prevAlbumID
|
|
}
|
|
}
|
|
}
|
|
entry.tracks = tracks
|
|
entry.tags = slices.Collect(maps.Values(uniqueTags))
|
|
return nil
|
|
}
|
|
|
|
// createAlbumsFromMediaFiles groups the entry's tracks by album ID and creates albums
|
|
func (p *phaseFolders) createAlbumsFromMediaFiles(entry *folderEntry) {
|
|
grouped := slice.Group(entry.tracks, func(mf model.MediaFile) string { return mf.AlbumID })
|
|
albums := make(model.Albums, 0, len(grouped))
|
|
for _, group := range grouped {
|
|
songs := model.MediaFiles(group)
|
|
album := songs.ToAlbum()
|
|
albums = append(albums, album)
|
|
}
|
|
entry.albums = albums
|
|
}
|
|
|
|
// createArtistsFromMediaFiles creates artists from the entry's tracks
|
|
func (p *phaseFolders) createArtistsFromMediaFiles(entry *folderEntry) {
|
|
participants := make(model.Participants, len(entry.tracks)*3) // preallocate ~3 artists per track
|
|
for _, track := range entry.tracks {
|
|
participants.Merge(track.Participants)
|
|
}
|
|
entry.artists = participants.AllArtists()
|
|
}
|
|
|
|
func (p *phaseFolders) persistChanges(entry *folderEntry) (*folderEntry, error) {
|
|
defer p.measure(entry)()
|
|
p.state.changesDetected.Store(true)
|
|
|
|
// Collect artwork IDs to pre-cache after the transaction commits
|
|
var artworkIDs []model.ArtworkID
|
|
|
|
err := p.ds.WithTx(func(tx model.DataStore) error {
|
|
// Instantiate all repositories just once per folder
|
|
folderRepo := tx.Folder(p.ctx)
|
|
tagRepo := tx.Tag(p.ctx)
|
|
artistRepo := tx.Artist(p.ctx)
|
|
libraryRepo := tx.Library(p.ctx)
|
|
albumRepo := tx.Album(p.ctx)
|
|
mfRepo := tx.MediaFile(p.ctx)
|
|
|
|
// Save folder to DB
|
|
folder := entry.toFolder()
|
|
err := folderRepo.Put(folder)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting folder to DB", "folder", entry.path, err)
|
|
return err
|
|
}
|
|
|
|
// Save all tags to DB
|
|
err = tagRepo.Add(entry.job.lib.ID, entry.tags...)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting tags to DB", "folder", entry.path, err)
|
|
return err
|
|
}
|
|
|
|
// Save all new/modified artists to DB. Their information will be incomplete, but they will be refreshed later
|
|
for i := range entry.artists {
|
|
err = artistRepo.Put(&entry.artists[i], "name",
|
|
"mbz_artist_id", "sort_artist_name", "order_artist_name", "full_text", "updated_at")
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting artist to DB", "folder", entry.path, "artist", entry.artists[i].Name, err)
|
|
return err
|
|
}
|
|
err = libraryRepo.AddArtist(entry.job.lib.ID, entry.artists[i].ID)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error adding artist to library", "lib", entry.job.lib.ID, "artist", entry.artists[i].Name, err)
|
|
return err
|
|
}
|
|
if entry.artists[i].Name != consts.UnknownArtist && entry.artists[i].Name != consts.VariousArtists {
|
|
artworkIDs = append(artworkIDs, entry.artists[i].CoverArtID())
|
|
}
|
|
}
|
|
|
|
// Save all new/modified albums to DB. Their information will be incomplete, but they will be refreshed later
|
|
for i := range entry.albums {
|
|
err = p.persistAlbum(albumRepo, &entry.albums[i], entry.albumIDMap)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting album to DB", "folder", entry.path, "album", entry.albums[i], err)
|
|
return err
|
|
}
|
|
if entry.albums[i].Name != consts.UnknownAlbum {
|
|
artworkIDs = append(artworkIDs, entry.albums[i].CoverArtID())
|
|
}
|
|
}
|
|
|
|
// Save all tracks to DB
|
|
for i := range entry.tracks {
|
|
err = mfRepo.Put(&entry.tracks[i])
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting mediafile to DB", "folder", entry.path, "track", entry.tracks[i], err)
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Mark all missing tracks as not available
|
|
if len(entry.missingTracks) > 0 {
|
|
err = mfRepo.MarkMissing(true, entry.missingTracks...)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error marking missing tracks", "folder", entry.path, err)
|
|
return err
|
|
}
|
|
|
|
// Touch all albums that have missing tracks, so they get refreshed in later phases
|
|
groupedMissingTracks := slice.ToMap(entry.missingTracks, func(mf *model.MediaFile) (string, struct{}) {
|
|
return mf.AlbumID, struct{}{}
|
|
})
|
|
albumsToUpdate := slices.Collect(maps.Keys(groupedMissingTracks))
|
|
err = albumRepo.Touch(albumsToUpdate...)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error touching album", "folder", entry.path, "albums", albumsToUpdate, err)
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}, "scanner: persist changes")
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error persisting changes to DB", "folder", entry.path, err)
|
|
}
|
|
|
|
// Pre-cache artwork after the transaction commits successfully
|
|
if err == nil {
|
|
for _, artID := range artworkIDs {
|
|
entry.job.cw.PreCache(artID)
|
|
}
|
|
}
|
|
|
|
return entry, err
|
|
}
|
|
|
|
// persistAlbum persists the given album to the database, and reassigns annotations from the previous album ID
|
|
func (p *phaseFolders) persistAlbum(repo model.AlbumRepository, a *model.Album, idMap map[string]string) error {
|
|
prevID := idMap[a.ID]
|
|
log.Trace(p.ctx, "Persisting album", "album", a.Name, "albumArtist", a.AlbumArtist, "id", a.ID, "prevID", cmp.Or(prevID, "nil"))
|
|
if err := repo.Put(a); err != nil {
|
|
return fmt.Errorf("persisting album %s: %w", a.ID, err)
|
|
}
|
|
if prevID == "" {
|
|
return nil
|
|
}
|
|
|
|
// Reassign annotation from previous album to new album
|
|
log.Trace(p.ctx, "Reassigning album annotations", "from", prevID, "to", a.ID, "album", a.Name)
|
|
if err := repo.ReassignAnnotation(prevID, a.ID); err != nil {
|
|
log.Warn(p.ctx, "Scanner: Could not reassign annotations", "from", prevID, "to", a.ID, "album", a.Name, err)
|
|
p.state.sendWarning(fmt.Sprintf("Could not reassign annotations from %s to %s ('%s'): %v", prevID, a.ID, a.Name, err))
|
|
}
|
|
|
|
// Keep created_at field from previous instance of the album
|
|
if err := repo.CopyAttributes(prevID, a.ID, "created_at"); err != nil {
|
|
// Silently ignore when the previous album is not found
|
|
if !errors.Is(err, model.ErrNotFound) {
|
|
log.Warn(p.ctx, "Scanner: Could not copy fields", "from", prevID, "to", a.ID, "album", a.Name, err)
|
|
p.state.sendWarning(fmt.Sprintf("Could not copy fields from %s to %s ('%s'): %v", prevID, a.ID, a.Name, err))
|
|
}
|
|
}
|
|
// Don't keep track of this mapping anymore
|
|
delete(idMap, a.ID)
|
|
return nil
|
|
}
|
|
|
|
func (p *phaseFolders) logFolder(entry *folderEntry) (*folderEntry, error) {
|
|
logCall := log.Info
|
|
if entry.isEmpty() {
|
|
logCall = log.Trace
|
|
}
|
|
logCall(p.ctx, "Scanner: Completed processing folder",
|
|
"audioCount", len(entry.audioFiles), "imageCount", len(entry.imageFiles), "plsCount", entry.numPlaylists,
|
|
"elapsed", entry.elapsed.Elapsed(), "tracksMissing", len(entry.missingTracks),
|
|
"tracksImported", len(entry.tracks), "library", entry.job.lib.Name, consts.Zwsp+"folder", entry.path)
|
|
return entry, nil
|
|
}
|
|
|
|
func (p *phaseFolders) finalize(err error) error {
|
|
errF := p.ds.WithTx(func(tx model.DataStore) error {
|
|
for _, job := range p.jobs {
|
|
// Mark all folders that were not updated as missing
|
|
if len(job.lastUpdates) == 0 {
|
|
continue
|
|
}
|
|
folderIDs := slices.Collect(maps.Keys(job.lastUpdates))
|
|
err := tx.Folder(p.ctx).MarkMissing(true, folderIDs...)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error marking missing folders", "lib", job.lib.Name, err)
|
|
return err
|
|
}
|
|
err = tx.MediaFile(p.ctx).MarkMissingByFolder(true, folderIDs...)
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error marking tracks in missing folders", "lib", job.lib.Name, err)
|
|
return err
|
|
}
|
|
// Touch all albums that have missing folders, so they get refreshed in later phases
|
|
_, err = tx.Album(p.ctx).TouchByMissingFolder()
|
|
if err != nil {
|
|
log.Error(p.ctx, "Scanner: Error touching albums with missing folders", "lib", job.lib.Name, err)
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}, "scanner: finalize phaseFolders")
|
|
return errors.Join(err, errF)
|
|
}
|
|
|
|
var _ phase[*folderEntry] = (*phaseFolders)(nil)
|