- Add delete button (SVG X, hover-reveal) and confirmation modal to comic cards - Add DELETE /api/comics/delete endpoint with path traversal protection - Fix container downloads: delegate Cloudflare-blocked requests to FlareSolverr (headless Chrome sidecar) instead of retrying with Go HTTP client, whose Linux TCP fingerprint is flagged by Cloudflare even with network_mode: host - Add FlareSolverr service to docker-compose; inject FLARESOLVERR_URL env var - Add diagnostic logging to BatcaveBizMarkup request flow - Trim URL whitespace before storing in download job - Guard Archive() against empty filelist; fix runJob error-check ordering
403 lines
9.3 KiB
Go
403 lines
9.3 KiB
Go
package web
|
|
|
|
import (
|
|
"archive/zip"
|
|
"embed"
|
|
"encoding/json"
|
|
"fmt"
|
|
"io"
|
|
"io/fs"
|
|
"net/http"
|
|
"net/url"
|
|
"os"
|
|
"path/filepath"
|
|
"sort"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/PuerkitoBio/goquery"
|
|
"yoink/comic"
|
|
)
|
|
|
|
//go:embed static
|
|
var staticFiles embed.FS
|
|
|
|
type JobStatus string
|
|
|
|
const (
|
|
StatusPending JobStatus = "pending"
|
|
StatusRunning JobStatus = "running"
|
|
StatusComplete JobStatus = "complete"
|
|
StatusError JobStatus = "error"
|
|
)
|
|
|
|
type Job struct {
|
|
ID string `json:"id"`
|
|
URL string `json:"url"`
|
|
Title string `json:"title"`
|
|
Status JobStatus `json:"status"`
|
|
Error string `json:"error,omitempty"`
|
|
CreatedAt time.Time `json:"created_at"`
|
|
}
|
|
|
|
type ComicEntry struct {
|
|
Title string `json:"title"`
|
|
CoverURL string `json:"cover_url"`
|
|
FileURL string `json:"file_url"`
|
|
DownloadedAt time.Time `json:"downloaded_at"`
|
|
}
|
|
|
|
type Server struct {
|
|
libraryPath string
|
|
jobs map[string]*Job
|
|
mu sync.RWMutex
|
|
}
|
|
|
|
func NewServer(libraryPath string) *Server {
|
|
return &Server{
|
|
libraryPath: libraryPath,
|
|
jobs: make(map[string]*Job),
|
|
}
|
|
}
|
|
|
|
func (s *Server) Handler() http.Handler {
|
|
mux := http.NewServeMux()
|
|
|
|
// Embedded static assets
|
|
staticFS, _ := fs.Sub(staticFiles, "static")
|
|
mux.Handle("/static/", http.StripPrefix("/static/", http.FileServer(http.FS(staticFS))))
|
|
|
|
// Library files: covers (inline) and cbz downloads (attachment)
|
|
mux.Handle("/covers/", http.StripPrefix("/covers/", http.FileServer(http.Dir(s.libraryPath))))
|
|
mux.Handle("/files/", http.StripPrefix("/files/", s.downloadHandler()))
|
|
|
|
// API
|
|
mux.HandleFunc("/api/download", s.handleDownload)
|
|
mux.HandleFunc("/api/upload", s.handleUpload)
|
|
mux.HandleFunc("/api/comics", s.handleComics)
|
|
mux.HandleFunc("/api/comics/delete", s.handleDeleteComic)
|
|
mux.HandleFunc("/api/jobs", s.handleJobs)
|
|
mux.HandleFunc("/health", func(w http.ResponseWriter, r *http.Request) {
|
|
w.WriteHeader(http.StatusOK)
|
|
})
|
|
|
|
// SPA root
|
|
mux.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) {
|
|
if r.URL.Path != "/" {
|
|
http.NotFound(w, r)
|
|
return
|
|
}
|
|
data, _ := staticFiles.ReadFile("static/index.html")
|
|
w.Header().Set("Content-Type", "text/html; charset=utf-8")
|
|
w.Write(data)
|
|
})
|
|
|
|
return mux
|
|
}
|
|
|
|
// downloadHandler wraps the library file server to force Content-Disposition: attachment.
|
|
func (s *Server) downloadHandler() http.Handler {
|
|
fs := http.FileServer(http.Dir(s.libraryPath))
|
|
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Disposition", "attachment")
|
|
fs.ServeHTTP(w, r)
|
|
})
|
|
}
|
|
|
|
func (s *Server) handleDownload(w http.ResponseWriter, r *http.Request) {
|
|
if r.Method != http.MethodPost {
|
|
http.Error(w, "method not allowed", http.StatusMethodNotAllowed)
|
|
return
|
|
}
|
|
|
|
var req struct {
|
|
URL string `json:"url"`
|
|
}
|
|
if err := json.NewDecoder(r.Body).Decode(&req); err != nil || strings.TrimSpace(req.URL) == "" {
|
|
http.Error(w, "invalid request", http.StatusBadRequest)
|
|
return
|
|
}
|
|
req.URL = strings.TrimSpace(req.URL)
|
|
|
|
job := &Job{
|
|
ID: fmt.Sprintf("%d", time.Now().UnixNano()),
|
|
URL: req.URL,
|
|
Status: StatusPending,
|
|
CreatedAt: time.Now(),
|
|
}
|
|
|
|
s.mu.Lock()
|
|
s.jobs[job.ID] = job
|
|
s.mu.Unlock()
|
|
|
|
go s.runJob(job)
|
|
|
|
w.Header().Set("Content-Type", "application/json")
|
|
json.NewEncoder(w).Encode(job)
|
|
}
|
|
|
|
func (s *Server) runJob(job *Job) {
|
|
s.mu.Lock()
|
|
job.Status = StatusRunning
|
|
s.mu.Unlock()
|
|
|
|
markupCh := make(chan *goquery.Document)
|
|
imageCh := make(chan []string)
|
|
|
|
c := comic.NewComic(job.URL, s.libraryPath, imageCh, markupCh)
|
|
|
|
s.mu.Lock()
|
|
job.Title = c.Title
|
|
s.mu.Unlock()
|
|
|
|
if len(c.Filelist) == 0 {
|
|
s.mu.Lock()
|
|
job.Status = StatusError
|
|
job.Error = "no images found"
|
|
s.mu.Unlock()
|
|
return
|
|
}
|
|
|
|
errs := c.Download(len(c.Filelist))
|
|
|
|
if err := c.Archive(); err != nil {
|
|
c.Cleanup()
|
|
s.mu.Lock()
|
|
job.Status = StatusError
|
|
job.Error = err.Error()
|
|
s.mu.Unlock()
|
|
return
|
|
}
|
|
|
|
c.Cleanup()
|
|
|
|
if len(errs) > 0 {
|
|
s.mu.Lock()
|
|
job.Status = StatusError
|
|
job.Error = errs[0].Error()
|
|
s.mu.Unlock()
|
|
return
|
|
}
|
|
|
|
s.mu.Lock()
|
|
job.Status = StatusComplete
|
|
s.mu.Unlock()
|
|
}
|
|
|
|
func (s *Server) handleComics(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Type", "application/json")
|
|
|
|
entries := []ComicEntry{}
|
|
|
|
dirs, err := os.ReadDir(s.libraryPath)
|
|
if err != nil {
|
|
json.NewEncoder(w).Encode(entries)
|
|
return
|
|
}
|
|
|
|
for _, dir := range dirs {
|
|
if !dir.IsDir() {
|
|
continue
|
|
}
|
|
|
|
title := dir.Name()
|
|
dirPath := filepath.Join(s.libraryPath, title)
|
|
|
|
var coverURL, fileURL string
|
|
var downloadedAt time.Time
|
|
|
|
files, _ := os.ReadDir(dirPath)
|
|
for _, f := range files {
|
|
name := f.Name()
|
|
|
|
if strings.HasSuffix(name, ".cbz") {
|
|
fileURL = "/files/" + url.PathEscape(title) + "/" + url.PathEscape(name)
|
|
if info, err := f.Info(); err == nil {
|
|
downloadedAt = info.ModTime()
|
|
}
|
|
}
|
|
|
|
// Cover kept by Cleanup: "<Title> 001.jpg"
|
|
stripped := strings.TrimSpace(strings.TrimPrefix(name, title))
|
|
if strings.HasPrefix(strings.ToLower(stripped), "001") {
|
|
coverURL = "/covers/" + url.PathEscape(title) + "/" + url.PathEscape(name)
|
|
}
|
|
}
|
|
|
|
if fileURL != "" {
|
|
entries = append(entries, ComicEntry{
|
|
Title: title,
|
|
CoverURL: coverURL,
|
|
FileURL: fileURL,
|
|
DownloadedAt: downloadedAt,
|
|
})
|
|
}
|
|
}
|
|
|
|
// Default: newest first
|
|
sort.Slice(entries, func(i, j int) bool {
|
|
return entries[i].DownloadedAt.After(entries[j].DownloadedAt)
|
|
})
|
|
|
|
json.NewEncoder(w).Encode(entries)
|
|
}
|
|
|
|
func (s *Server) handleJobs(w http.ResponseWriter, r *http.Request) {
|
|
s.mu.RLock()
|
|
jobs := make([]*Job, 0, len(s.jobs))
|
|
for _, j := range s.jobs {
|
|
jobs = append(jobs, j)
|
|
}
|
|
s.mu.RUnlock()
|
|
|
|
w.Header().Set("Content-Type", "application/json")
|
|
json.NewEncoder(w).Encode(jobs)
|
|
}
|
|
|
|
func (s *Server) handleUpload(w http.ResponseWriter, r *http.Request) {
|
|
if r.Method != http.MethodPost {
|
|
http.Error(w, "method not allowed", http.StatusMethodNotAllowed)
|
|
return
|
|
}
|
|
|
|
// 500 MB limit
|
|
if err := r.ParseMultipartForm(500 << 20); err != nil {
|
|
http.Error(w, "request too large", http.StatusRequestEntityTooLarge)
|
|
return
|
|
}
|
|
|
|
title := strings.TrimSpace(r.FormValue("title"))
|
|
if title == "" {
|
|
http.Error(w, "title required", http.StatusBadRequest)
|
|
return
|
|
}
|
|
// Sanitize: no path separators or shell-special characters
|
|
title = filepath.Base(title)
|
|
title = strings.Map(func(r rune) rune {
|
|
if strings.ContainsRune(`/\:*?"<>|`, r) {
|
|
return '_'
|
|
}
|
|
return r
|
|
}, title)
|
|
|
|
fileHeaders := r.MultipartForm.File["images"]
|
|
if len(fileHeaders) == 0 {
|
|
http.Error(w, "no images provided", http.StatusBadRequest)
|
|
return
|
|
}
|
|
|
|
// Sort by original filename so page order is preserved
|
|
sort.Slice(fileHeaders, func(i, j int) bool {
|
|
return fileHeaders[i].Filename < fileHeaders[j].Filename
|
|
})
|
|
|
|
dir := filepath.Join(s.libraryPath, title)
|
|
if err := os.MkdirAll(dir, 0o755); err != nil {
|
|
http.Error(w, "failed to create directory", http.StatusInternalServerError)
|
|
return
|
|
}
|
|
|
|
cbzPath := filepath.Join(dir, title+".cbz")
|
|
cbzFile, err := os.Create(cbzPath)
|
|
if err != nil {
|
|
http.Error(w, "failed to create archive", http.StatusInternalServerError)
|
|
return
|
|
}
|
|
defer cbzFile.Close()
|
|
|
|
zw := zip.NewWriter(cbzFile)
|
|
defer zw.Close()
|
|
|
|
imageExts := map[string]bool{".jpg": true, ".jpeg": true, ".png": true, ".webp": true}
|
|
idx := 1
|
|
|
|
for _, fh := range fileHeaders {
|
|
ext := strings.ToLower(filepath.Ext(fh.Filename))
|
|
if !imageExts[ext] {
|
|
continue
|
|
}
|
|
if ext == ".jpeg" {
|
|
ext = ".jpg"
|
|
}
|
|
|
|
entryName := fmt.Sprintf("%03d%s", idx, ext)
|
|
|
|
src, err := fh.Open()
|
|
if err != nil {
|
|
continue
|
|
}
|
|
|
|
// Save first image as cover: "<Title> 001.jpg"
|
|
if idx == 1 {
|
|
coverPath := filepath.Join(dir, title+" "+entryName)
|
|
if cf, err := os.Create(coverPath); err == nil {
|
|
io.Copy(cf, src)
|
|
cf.Close()
|
|
src.Close()
|
|
src, err = fh.Open()
|
|
if err != nil {
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
|
|
ze, err := zw.Create(entryName)
|
|
if err != nil {
|
|
src.Close()
|
|
continue
|
|
}
|
|
io.Copy(ze, src)
|
|
src.Close()
|
|
idx++
|
|
}
|
|
|
|
if idx == 1 {
|
|
// Nothing was written — no valid images
|
|
os.RemoveAll(dir)
|
|
http.Error(w, "no valid images in upload", http.StatusBadRequest)
|
|
return
|
|
}
|
|
|
|
w.Header().Set("Content-Type", "application/json")
|
|
json.NewEncoder(w).Encode(map[string]string{"title": title, "status": "complete"})
|
|
}
|
|
|
|
func (s *Server) handleDeleteComic(w http.ResponseWriter, r *http.Request) {
|
|
if r.Method != http.MethodDelete {
|
|
http.Error(w, "method not allowed", http.StatusMethodNotAllowed)
|
|
return
|
|
}
|
|
|
|
var req struct {
|
|
Title string `json:"title"`
|
|
}
|
|
if err := json.NewDecoder(r.Body).Decode(&req); err != nil || strings.TrimSpace(req.Title) == "" {
|
|
http.Error(w, "invalid request", http.StatusBadRequest)
|
|
return
|
|
}
|
|
|
|
// Sanitize: prevent path traversal
|
|
title := filepath.Base(strings.TrimSpace(req.Title))
|
|
comicDir := filepath.Join(s.libraryPath, title)
|
|
|
|
// Ensure the resolved path is still under the library
|
|
if !strings.HasPrefix(comicDir, filepath.Clean(s.libraryPath)+string(filepath.Separator)) {
|
|
http.Error(w, "invalid title", http.StatusBadRequest)
|
|
return
|
|
}
|
|
|
|
if err := os.RemoveAll(comicDir); err != nil {
|
|
http.Error(w, "failed to delete comic", http.StatusInternalServerError)
|
|
return
|
|
}
|
|
|
|
w.WriteHeader(http.StatusNoContent)
|
|
}
|
|
|
|
func Listen(addr string, libraryPath string) error {
|
|
srv := NewServer(libraryPath)
|
|
fmt.Printf("Yoink web server listening on %s\n", addr)
|
|
return http.ListenAndServe(addr, srv.Handler())
|
|
}
|