From ef097f02fe76572d2b1887c28da3f2bd83a993a0 Mon Sep 17 00:00:00 2001 From: Frédéric Guillot Date: Tue, 12 Dec 2017 19:19:36 -0800 Subject: Add the possibility to enable crawler for feeds --- reader/feed/handler.go | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) (limited to 'reader/feed') diff --git a/reader/feed/handler.go b/reader/feed/handler.go index 7a98613..6b94627 100644 --- a/reader/feed/handler.go +++ b/reader/feed/handler.go @@ -33,7 +33,7 @@ type Handler struct { } // CreateFeed fetch, parse and store a new feed. -func (h *Handler) CreateFeed(userID, categoryID int64, url string) (*model.Feed, error) { +func (h *Handler) CreateFeed(userID, categoryID int64, url string, crawler bool) (*model.Feed, error) { defer helper.ExecutionTime(time.Now(), fmt.Sprintf("[Handler:CreateFeed] feedUrl=%s", url)) if !h.store.CategoryExists(userID, categoryID) { @@ -65,6 +65,7 @@ func (h *Handler) CreateFeed(userID, categoryID int64, url string) (*model.Feed, } feedProcessor := processor.NewFeedProcessor(subscription) + feedProcessor.WithCrawler(crawler) feedProcessor.Process() subscription.Category = &model.Category{ID: categoryID} @@ -72,6 +73,7 @@ func (h *Handler) CreateFeed(userID, categoryID int64, url string) (*model.Feed, subscription.LastModifiedHeader = response.LastModified subscription.FeedURL = response.EffectiveURL subscription.UserID = userID + subscription.Crawler = crawler err = h.store.CreateFeed(subscription) if err != nil { @@ -143,6 +145,7 @@ func (h *Handler) RefreshFeed(userID, feedID int64) error { feedProcessor := processor.NewFeedProcessor(subscription) feedProcessor.WithScraperRules(originalFeed.ScraperRules) feedProcessor.WithRewriteRules(originalFeed.RewriteRules) + feedProcessor.WithCrawler(originalFeed.Crawler) feedProcessor.Process() originalFeed.EtagHeader = response.ETag -- cgit v1.2.3