aboutsummaryrefslogtreecommitdiff
path: root/internal/reader/processor/processor.go
diff options
context:
space:
mode:
authorGravatar Frédéric Guillot <f@miniflux.net> 2023-09-24 16:32:09 -0700
committerGravatar Frédéric Guillot <f@miniflux.net> 2023-09-24 22:37:33 -0700
commitc0e954f19d707fef8ef8271636ec661634a4c4c7 (patch)
tree5aa052a048f470e233a454e5ad9071eed1fa37c0 /internal/reader/processor/processor.go
parent54cb8fa0286e4a2f1a81c32b5a89722d93b30bf7 (diff)
downloadv2-c0e954f19d707fef8ef8271636ec661634a4c4c7.tar.gz
v2-c0e954f19d707fef8ef8271636ec661634a4c4c7.tar.zst
v2-c0e954f19d707fef8ef8271636ec661634a4c4c7.zip
Implement structured logging using log/slog package
Diffstat (limited to 'internal/reader/processor/processor.go')
-rw-r--r--internal/reader/processor/processor.go78
1 files changed, 68 insertions, 10 deletions
diff --git a/internal/reader/processor/processor.go b/internal/reader/processor/processor.go
index d56d4289..c6514bf6 100644
--- a/internal/reader/processor/processor.go
+++ b/internal/reader/processor/processor.go
@@ -6,6 +6,7 @@ package processor
import (
"errors"
"fmt"
+ "log/slog"
"math"
"regexp"
"strconv"
@@ -15,7 +16,6 @@ import (
"miniflux.app/v2/internal/config"
"miniflux.app/v2/internal/http/client"
- "miniflux.app/v2/internal/logger"
"miniflux.app/v2/internal/metric"
"miniflux.app/v2/internal/model"
"miniflux.app/v2/internal/reader/browser"
@@ -43,7 +43,13 @@ func ProcessFeedEntries(store *storage.Storage, feed *model.Feed, user *model.Us
for i := len(feed.Entries) - 1; i >= 0; i-- {
entry := feed.Entries[i]
- logger.Debug("[Processor] Processing entry %q from feed %q", entry.URL, feed.FeedURL)
+ slog.Debug("Processing entry",
+ slog.Int64("user_id", user.ID),
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ )
if isBlockedEntry(feed, entry) || !isAllowedEntry(feed, entry) {
continue
@@ -52,7 +58,13 @@ func ProcessFeedEntries(store *storage.Storage, feed *model.Feed, user *model.Us
url := getUrlFromEntry(feed, entry)
entryIsNew := !store.EntryURLExists(feed.ID, entry.URL)
if feed.Crawler && (entryIsNew || forceRefresh) {
- logger.Debug("[Processor] Crawling entry %q from feed %q", url, feed.FeedURL)
+ slog.Debug("Scraping entry",
+ slog.Int64("user_id", user.ID),
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ )
startTime := time.Now()
content, scraperErr := scraper.Fetch(
@@ -73,7 +85,14 @@ func ProcessFeedEntries(store *storage.Storage, feed *model.Feed, user *model.Us
}
if scraperErr != nil {
- logger.Error(`[Processor] Unable to crawl this entry: %q => %v`, entry.URL, scraperErr)
+ slog.Warn("Unable to scrape entry",
+ slog.Int64("user_id", user.ID),
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.Any("error", scraperErr),
+ )
} else if content != "" {
// We replace the entry content only if the scraper doesn't return any error.
entry.Content = content
@@ -96,7 +115,13 @@ func isBlockedEntry(feed *model.Feed, entry *model.Entry) bool {
if feed.BlocklistRules != "" {
match, _ := regexp.MatchString(feed.BlocklistRules, entry.Title)
if match {
- logger.Debug("[Processor] Blocking entry %q from feed %q based on rule %q", entry.Title, feed.FeedURL, feed.BlocklistRules)
+ slog.Debug("Blocking entry based on rule",
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.String("rule", feed.BlocklistRules),
+ )
return true
}
}
@@ -107,7 +132,13 @@ func isAllowedEntry(feed *model.Feed, entry *model.Entry) bool {
if feed.KeeplistRules != "" {
match, _ := regexp.MatchString(feed.KeeplistRules, entry.Title)
if match {
- logger.Debug("[Processor] Allow entry %q from feed %q based on rule %q", entry.Title, feed.FeedURL, feed.KeeplistRules)
+ slog.Debug("Allow entry based on rule",
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.String("rule", feed.KeeplistRules),
+ )
return true
}
return false
@@ -160,9 +191,22 @@ func getUrlFromEntry(feed *model.Feed, entry *model.Entry) string {
if len(parts) >= 3 {
re := regexp.MustCompile(parts[1])
url = re.ReplaceAllString(entry.URL, parts[2])
- logger.Debug(`[Processor] Rewriting entry URL %s to %s`, entry.URL, url)
+ slog.Debug("Rewriting entry URL",
+ slog.Int64("entry_id", entry.ID),
+ slog.String("original_entry_url", entry.URL),
+ slog.String("rewritten_entry_url", url),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ )
} else {
- logger.Debug("[Processor] Cannot find search and replace terms for replace rule %s", feed.UrlRewriteRules)
+ slog.Debug("Cannot find search and replace terms for replace rule",
+ slog.Int64("entry_id", entry.ID),
+ slog.String("original_entry_url", entry.URL),
+ slog.String("rewritten_entry_url", url),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.String("url_rewrite_rules", feed.UrlRewriteRules),
+ )
}
}
return url
@@ -173,7 +217,14 @@ func updateEntryReadingTime(store *storage.Storage, feed *model.Feed, entry *mod
if entryIsNew {
watchTime, err := fetchYouTubeWatchTime(entry.URL)
if err != nil {
- logger.Error("[Processor] Unable to fetch YouTube watch time: %q => %v", entry.URL, err)
+ slog.Warn("Unable to fetch YouTube watch time",
+ slog.Int64("user_id", user.ID),
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.Any("error", err),
+ )
}
entry.ReadingTime = watchTime
} else {
@@ -185,7 +236,14 @@ func updateEntryReadingTime(store *storage.Storage, feed *model.Feed, entry *mod
if entryIsNew {
watchTime, err := fetchOdyseeWatchTime(entry.URL)
if err != nil {
- logger.Error("[Processor] Unable to fetch Odysee watch time: %q => %v", entry.URL, err)
+ slog.Warn("Unable to fetch Odysee watch time",
+ slog.Int64("user_id", user.ID),
+ slog.Int64("entry_id", entry.ID),
+ slog.String("entry_url", entry.URL),
+ slog.Int64("feed_id", feed.ID),
+ slog.String("feed_url", feed.FeedURL),
+ slog.Any("error", err),
+ )
}
entry.ReadingTime = watchTime
} else {