aboutsummaryrefslogtreecommitdiffhomepage
path: root/reader
diff options
context:
space:
mode:
authorGravatar Frédéric Guillot <fred@miniflux.net>2017-12-15 18:55:57 -0800
committerGravatar Frédéric Guillot <fred@miniflux.net>2017-12-15 18:55:57 -0800
commit1d8193b892bf63db6b91c66b9d048873bf1bc4cf (patch)
tree4ca54691f7fab88f73f657f005e26d75ae5f0058 /reader
parentc6d9eb361454b70406cf6573b284d5e99ba3a2d2 (diff)
Add logger
Diffstat (limited to 'reader')
-rw-r--r--reader/atom/atom.go4
-rw-r--r--reader/feed/handler.go16
-rw-r--r--reader/icon/finder.go4
-rw-r--r--reader/json/json.go4
-rw-r--r--reader/opml/handler.go10
-rw-r--r--reader/opml/serializer.go5
-rw-r--r--reader/processor/processor.go5
-rw-r--r--reader/readability/readability.go7
-rw-r--r--reader/rss/rss.go4
-rw-r--r--reader/scraper/scraper.go6
-rw-r--r--reader/subscription/finder.go4
11 files changed, 34 insertions, 35 deletions
diff --git a/reader/atom/atom.go b/reader/atom/atom.go
index 315961e..61844a5 100644
--- a/reader/atom/atom.go
+++ b/reader/atom/atom.go
@@ -6,12 +6,12 @@ package atom
import (
"encoding/xml"
- "log"
"strconv"
"strings"
"time"
"github.com/miniflux/miniflux/helper"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/reader/date"
"github.com/miniflux/miniflux/url"
@@ -130,7 +130,7 @@ func getDate(a *atomEntry) time.Time {
if a.Updated != "" {
result, err := date.Parse(a.Updated)
if err != nil {
- log.Println(err)
+ logger.Error("atom: %v", err)
return time.Now()
}
diff --git a/reader/feed/handler.go b/reader/feed/handler.go
index 5926bbc..e804e29 100644
--- a/reader/feed/handler.go
+++ b/reader/feed/handler.go
@@ -6,12 +6,12 @@ package feed
import (
"fmt"
- "log"
"time"
"github.com/miniflux/miniflux/errors"
"github.com/miniflux/miniflux/helper"
"github.com/miniflux/miniflux/http"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/reader/icon"
"github.com/miniflux/miniflux/reader/processor"
@@ -80,13 +80,13 @@ func (h *Handler) CreateFeed(userID, categoryID int64, url string, crawler bool)
return nil, err
}
- log.Println("[Handler:CreateFeed] Feed saved with ID:", subscription.ID)
+ logger.Debug("[Handler:CreateFeed] Feed saved with ID: %d", subscription.ID)
icon, err := icon.FindIcon(subscription.SiteURL)
if err != nil {
- log.Println(err)
+ logger.Error("[Handler:CreateFeed] %v", err)
} else if icon == nil {
- log.Printf("No icon found for feedID=%d\n", subscription.ID)
+ logger.Info("No icon found for feedID=%d", subscription.ID)
} else {
h.store.CreateFeedIcon(subscription, icon)
}
@@ -128,7 +128,7 @@ func (h *Handler) RefreshFeed(userID, feedID int64) error {
}
if response.IsModified(originalFeed.EtagHeader, originalFeed.LastModifiedHeader) {
- log.Printf("[Handler:RefreshFeed] Feed #%d has been modified\n", feedID)
+ logger.Debug("[Handler:RefreshFeed] Feed #%d has been modified", feedID)
body, err := response.NormalizeBodyEncoding()
if err != nil {
return errors.NewLocalizedError(errEncoding, err)
@@ -156,16 +156,16 @@ func (h *Handler) RefreshFeed(userID, feedID int64) error {
}
if !h.store.HasIcon(originalFeed.ID) {
- log.Println("[Handler:RefreshFeed] Looking for feed icon")
+ logger.Debug("[Handler:RefreshFeed] Looking for feed icon")
icon, err := icon.FindIcon(originalFeed.SiteURL)
if err != nil {
- log.Println("[Handler:RefreshFeed]", err)
+ logger.Error("[Handler:RefreshFeed] %v", err)
} else {
h.store.CreateFeedIcon(originalFeed, icon)
}
}
} else {
- log.Printf("[Handler:RefreshFeed] Feed #%d not modified\n", feedID)
+ logger.Debug("[Handler:RefreshFeed] Feed #%d not modified", feedID)
}
originalFeed.ParsingErrorCount = 0
diff --git a/reader/icon/finder.go b/reader/icon/finder.go
index 2a5beb5..71f4845 100644
--- a/reader/icon/finder.go
+++ b/reader/icon/finder.go
@@ -8,10 +8,10 @@ import (
"fmt"
"io"
"io/ioutil"
- "log"
"github.com/miniflux/miniflux/helper"
"github.com/miniflux/miniflux/http"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/url"
@@ -36,7 +36,7 @@ func FindIcon(websiteURL string) (*model.Icon, error) {
return nil, err
}
- log.Println("[FindIcon] Fetching icon =>", iconURL)
+ logger.Debug("[FindIcon] Fetching icon => %s", iconURL)
icon, err := downloadIcon(iconURL)
if err != nil {
return nil, err
diff --git a/reader/json/json.go b/reader/json/json.go
index ad92068..900a920 100644
--- a/reader/json/json.go
+++ b/reader/json/json.go
@@ -5,11 +5,11 @@
package json
import (
- "log"
"strings"
"time"
"github.com/miniflux/miniflux/helper"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/reader/date"
"github.com/miniflux/miniflux/reader/sanitizer"
@@ -87,7 +87,7 @@ func (j *jsonItem) GetDate() time.Time {
if value != "" {
d, err := date.Parse(value)
if err != nil {
- log.Println(err)
+ logger.Error("json: %v", err)
return time.Now()
}
diff --git a/reader/opml/handler.go b/reader/opml/handler.go
index 192e04a..ec05670 100644
--- a/reader/opml/handler.go
+++ b/reader/opml/handler.go
@@ -8,8 +8,8 @@ import (
"errors"
"fmt"
"io"
- "log"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/storage"
)
@@ -23,7 +23,7 @@ type Handler struct {
func (h *Handler) Export(userID int64) (string, error) {
feeds, err := h.store.Feeds(userID)
if err != nil {
- log.Println(err)
+ logger.Error("[OPML:Export] %v", err)
return "", errors.New("unable to fetch feeds")
}
@@ -54,13 +54,13 @@ func (h *Handler) Import(userID int64, data io.Reader) (err error) {
if subscription.CategoryName == "" {
category, err = h.store.FirstCategory(userID)
if err != nil {
- log.Println(err)
+ logger.Error("[OPML:Import] %v", err)
return errors.New("unable to find first category")
}
} else {
category, err = h.store.CategoryByTitle(userID, subscription.CategoryName)
if err != nil {
- log.Println(err)
+ logger.Error("[OPML:Import] %v", err)
return errors.New("unable to search category by title")
}
@@ -72,7 +72,7 @@ func (h *Handler) Import(userID int64, data io.Reader) (err error) {
err := h.store.CreateCategory(category)
if err != nil {
- log.Println(err)
+ logger.Error("[OPML:Import] %v", err)
return fmt.Errorf(`unable to create this category: "%s"`, subscription.CategoryName)
}
}
diff --git a/reader/opml/serializer.go b/reader/opml/serializer.go
index 5ba494e..68cafa9 100644
--- a/reader/opml/serializer.go
+++ b/reader/opml/serializer.go
@@ -8,7 +8,8 @@ import (
"bufio"
"bytes"
"encoding/xml"
- "log"
+
+ "github.com/miniflux/miniflux/logger"
)
// Serialize returns a SubcriptionList in OPML format.
@@ -37,7 +38,7 @@ func Serialize(subscriptions SubcriptionList) string {
encoder := xml.NewEncoder(writer)
encoder.Indent(" ", " ")
if err := encoder.Encode(feeds); err != nil {
- log.Println(err)
+ logger.Error("[OPML:Serialize] %v", err)
return ""
}
diff --git a/reader/processor/processor.go b/reader/processor/processor.go
index 3cdefb0..33aa8ed 100644
--- a/reader/processor/processor.go
+++ b/reader/processor/processor.go
@@ -5,8 +5,7 @@
package processor
import (
- "log"
-
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/reader/rewrite"
"github.com/miniflux/miniflux/reader/sanitizer"
@@ -42,7 +41,7 @@ func (f *FeedProcessor) Process() {
if f.crawler {
content, err := scraper.Fetch(entry.URL, f.scraperRules)
if err != nil {
- log.Println("[FeedProcessor]", err)
+ logger.Error("[FeedProcessor] %v", err)
} else {
entry.Content = content
}
diff --git a/reader/readability/readability.go b/reader/readability/readability.go
index 37b4813..5eb4cc3 100644
--- a/reader/readability/readability.go
+++ b/reader/readability/readability.go
@@ -8,12 +8,12 @@ import (
"bytes"
"fmt"
"io"
- "log"
"math"
"regexp"
"strings"
"github.com/PuerkitoBio/goquery"
+ "github.com/miniflux/miniflux/logger"
"golang.org/x/net/html"
)
@@ -83,10 +83,10 @@ func ExtractContent(page io.Reader) (string, error) {
removeUnlikelyCandidates(document)
candidates := getCandidates(document)
- log.Println("Candidates:", candidates)
+ logger.Debug("[Readability] Candidates: %v", candidates)
topCandidate := getTopCandidate(document, candidates)
- log.Println("TopCandidate:", topCandidate)
+ logger.Debug("[Readability] TopCandidate: %v", topCandidate)
output := getArticle(topCandidate, candidates)
return output, nil
@@ -142,7 +142,6 @@ func removeUnlikelyCandidates(document *goquery.Document) {
str := class + id
if blacklistCandidatesRegexp.MatchString(str) || (unlikelyCandidatesRegexp.MatchString(str) && !okMaybeItsACandidateRegexp.MatchString(str)) {
- // log.Printf("Removing unlikely candidate - %s\n", str)
removeNodes(s)
}
})
diff --git a/reader/rss/rss.go b/reader/rss/rss.go
index fcfccbd..2a7dc3e 100644
--- a/reader/rss/rss.go
+++ b/reader/rss/rss.go
@@ -6,13 +6,13 @@ package rss
import (
"encoding/xml"
- "log"
"path"
"strconv"
"strings"
"time"
"github.com/miniflux/miniflux/helper"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/model"
"github.com/miniflux/miniflux/reader/date"
"github.com/miniflux/miniflux/url"
@@ -130,7 +130,7 @@ func (r *rssItem) GetDate() time.Time {
if value != "" {
result, err := date.Parse(value)
if err != nil {
- log.Println(err)
+ logger.Error("rss: %v", err)
return time.Now()
}
diff --git a/reader/scraper/scraper.go b/reader/scraper/scraper.go
index 3d1fb41..d2e0d4d 100644
--- a/reader/scraper/scraper.go
+++ b/reader/scraper/scraper.go
@@ -7,11 +7,11 @@ package scraper
import (
"errors"
"io"
- "log"
"strings"
"github.com/PuerkitoBio/goquery"
"github.com/miniflux/miniflux/http"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/reader/readability"
"github.com/miniflux/miniflux/url"
)
@@ -42,10 +42,10 @@ func Fetch(websiteURL, rules string) (string, error) {
var content string
if rules != "" {
- log.Printf(`[Scraper] Using rules "%s" for "%s"`, rules, websiteURL)
+ logger.Debug(`[Scraper] Using rules "%s" for "%s"`, rules, websiteURL)
content, err = scrapContent(page, rules)
} else {
- log.Printf(`[Scraper] Using readability for "%s"`, websiteURL)
+ logger.Debug(`[Scraper] Using readability for "%s"`, websiteURL)
content, err = readability.ExtractContent(page)
}
diff --git a/reader/subscription/finder.go b/reader/subscription/finder.go
index 2935b93..a550b13 100644
--- a/reader/subscription/finder.go
+++ b/reader/subscription/finder.go
@@ -8,12 +8,12 @@ import (
"bytes"
"fmt"
"io"
- "log"
"time"
"github.com/miniflux/miniflux/errors"
"github.com/miniflux/miniflux/helper"
"github.com/miniflux/miniflux/http"
+ "github.com/miniflux/miniflux/logger"
"github.com/miniflux/miniflux/reader/feed"
"github.com/miniflux/miniflux/url"
@@ -87,7 +87,7 @@ func parseDocument(websiteURL string, data io.Reader) (Subscriptions, error) {
}
if subscription.URL != "" {
- log.Println("[FindSubscriptions]", subscription)
+ logger.Debug("[FindSubscriptions] %s", subscription)
subscriptions = append(subscriptions, subscription)
}
})