mirror of
https://github.com/kemko/icecast-ripper.git
synced 2026-01-01 15:55:42 +03:00
183 lines
4.7 KiB
Go
183 lines
4.7 KiB
Go
package rss
|
|
|
|
import (
|
|
"fmt"
|
|
"io/fs"
|
|
"log/slog"
|
|
"path/filepath"
|
|
"regexp"
|
|
"sort"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/gorilla/feeds"
|
|
"github.com/kemko/icecast-ripper/internal/config"
|
|
"github.com/kemko/icecast-ripper/internal/hash"
|
|
)
|
|
|
|
// RecordingInfo contains metadata about a recording
|
|
type RecordingInfo struct {
|
|
Filename string
|
|
Hash string
|
|
FileSize int64
|
|
Duration time.Duration
|
|
RecordedAt time.Time
|
|
}
|
|
|
|
// Generator creates RSS feeds
|
|
type Generator struct {
|
|
feedBaseURL string
|
|
recordingsPath string
|
|
feedTitle string
|
|
feedDesc string
|
|
streamName string
|
|
}
|
|
|
|
// New creates a new RSS Generator instance
|
|
func New(cfg *config.Config, title, description, streamName string) *Generator {
|
|
baseURL := cfg.PublicUrl
|
|
if baseURL == "" {
|
|
slog.Warn("RSS_FEED_URL not set, using default")
|
|
baseURL = "http://localhost:8080/recordings/"
|
|
}
|
|
|
|
// Ensure base URL ends with a slash
|
|
if !strings.HasSuffix(baseURL, "/") {
|
|
baseURL += "/"
|
|
}
|
|
|
|
return &Generator{
|
|
feedBaseURL: baseURL,
|
|
recordingsPath: cfg.RecordingsPath,
|
|
feedTitle: title,
|
|
feedDesc: description,
|
|
streamName: streamName,
|
|
}
|
|
}
|
|
|
|
// Pattern to extract timestamp from recording filename (stream.somesite.com_20240907_195622.mp3)
|
|
var recordingPattern = regexp.MustCompile(`([^_]+)_(\d{8}_\d{6})\.mp3$`)
|
|
|
|
// GenerateFeed produces the RSS feed XML as a byte slice
|
|
func (g *Generator) GenerateFeed(maxItems int) ([]byte, error) {
|
|
recordings, err := g.scanRecordings(maxItems)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to scan recordings: %w", err)
|
|
}
|
|
|
|
feed := &feeds.Feed{
|
|
Title: g.feedTitle,
|
|
Link: &feeds.Link{Href: g.feedBaseURL},
|
|
Description: g.feedDesc,
|
|
Created: time.Now(),
|
|
}
|
|
|
|
feed.Items = make([]*feeds.Item, 0, len(recordings))
|
|
|
|
baseURL := g.feedBaseURL
|
|
baseURL = strings.TrimSuffix(baseURL, "/")
|
|
|
|
for _, rec := range recordings {
|
|
fileURL := fmt.Sprintf("%s/recordings/%s", baseURL, rec.Filename)
|
|
|
|
item := &feeds.Item{
|
|
Title: fmt.Sprintf("Recording %s", rec.RecordedAt.Format("2006-01-02 15:04")),
|
|
Link: &feeds.Link{Href: fileURL},
|
|
Description: fmt.Sprintf("Icecast stream recording from %s. Duration: %s",
|
|
rec.RecordedAt.Format(time.RFC1123), rec.Duration.String()),
|
|
Created: rec.RecordedAt,
|
|
Id: rec.Hash,
|
|
Enclosure: &feeds.Enclosure{
|
|
Url: fileURL,
|
|
Length: fmt.Sprintf("%d", rec.FileSize),
|
|
Type: "audio/mpeg",
|
|
},
|
|
}
|
|
feed.Items = append(feed.Items, item)
|
|
}
|
|
|
|
rssFeed, err := feed.ToRss()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to generate RSS feed: %w", err)
|
|
}
|
|
|
|
slog.Debug("RSS feed generated", "itemCount", len(feed.Items))
|
|
return []byte(rssFeed), nil
|
|
}
|
|
|
|
// scanRecordings scans the recordings directory and returns metadata about the files
|
|
func (g *Generator) scanRecordings(maxItems int) ([]RecordingInfo, error) {
|
|
var recordings []RecordingInfo
|
|
|
|
err := filepath.WalkDir(g.recordingsPath, func(path string, d fs.DirEntry, err error) error {
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Skip directories
|
|
if d.IsDir() {
|
|
return nil
|
|
}
|
|
|
|
// Only process mp3 files
|
|
if !strings.HasSuffix(strings.ToLower(d.Name()), ".mp3") {
|
|
return nil
|
|
}
|
|
|
|
// Extract timestamp from filename
|
|
matches := recordingPattern.FindStringSubmatch(d.Name())
|
|
if len(matches) < 3 {
|
|
// Skip files not matching our pattern
|
|
slog.Debug("Skipping non-conforming filename", "filename", d.Name())
|
|
return nil
|
|
}
|
|
|
|
// Parse the timestamp (now in the 3rd capture group [2])
|
|
timestamp, err := time.Parse("20060102_150405", matches[2])
|
|
if err != nil {
|
|
slog.Warn("Failed to parse timestamp from filename", "filename", d.Name(), "error", err)
|
|
return nil
|
|
}
|
|
|
|
info, err := d.Info()
|
|
if err != nil {
|
|
slog.Warn("Failed to get file info", "filename", d.Name(), "error", err)
|
|
return nil
|
|
}
|
|
|
|
// Calculate an estimated duration based on file size
|
|
// Assuming ~128kbps MP3 bitrate: 16KB per second
|
|
estimatedDuration := time.Duration(info.Size()/16000) * time.Second
|
|
|
|
// Generate a stable hash for the recording
|
|
filename := filepath.Base(path)
|
|
fileHash := hash.GenerateGUID(g.streamName, timestamp, filename)
|
|
|
|
recordings = append(recordings, RecordingInfo{
|
|
Filename: filename,
|
|
Hash: fileHash,
|
|
FileSize: info.Size(),
|
|
Duration: estimatedDuration,
|
|
RecordedAt: timestamp,
|
|
})
|
|
|
|
return nil
|
|
})
|
|
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to walk recordings directory: %w", err)
|
|
}
|
|
|
|
// Sort recordings by timestamp (newest first)
|
|
sort.Slice(recordings, func(i, j int) bool {
|
|
return recordings[i].RecordedAt.After(recordings[j].RecordedAt)
|
|
})
|
|
|
|
// Limit number of items if specified
|
|
if maxItems > 0 && maxItems < len(recordings) {
|
|
recordings = recordings[:maxItems]
|
|
}
|
|
|
|
return recordings, nil
|
|
}
|