Removed code duplication in IndexerBase
This commit is contained in:
parent
a877312b1a
commit
ba12465bb0
|
@ -85,10 +85,9 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
|
|
||||||
var result = new List<EpisodeParseResult>();
|
var result = new List<EpisodeParseResult>();
|
||||||
|
|
||||||
foreach (var url in Urls)
|
|
||||||
{
|
result = Fetch(Urls);
|
||||||
result.AddRange(Fetch(url));
|
|
||||||
}
|
|
||||||
|
|
||||||
_logger.Info("Finished processing feeds from " + Name);
|
_logger.Info("Finished processing feeds from " + Name);
|
||||||
return result;
|
return result;
|
||||||
|
@ -98,14 +97,8 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
{
|
{
|
||||||
_logger.Debug("Searching {0} for {1}-Season {2}", Name, seriesTitle, seasonNumber);
|
_logger.Debug("Searching {0} for {1}-Season {2}", Name, seriesTitle, seasonNumber);
|
||||||
|
|
||||||
var result = new List<EpisodeParseResult>();
|
|
||||||
|
|
||||||
var searchUrls = GetSeasonSearchUrls(GetQueryTitle(seriesTitle), seasonNumber);
|
var searchUrls = GetSeasonSearchUrls(GetQueryTitle(seriesTitle), seasonNumber);
|
||||||
|
var result = Fetch(searchUrls);
|
||||||
foreach (var url in searchUrls)
|
|
||||||
{
|
|
||||||
result.AddRange(Fetch(url));
|
|
||||||
}
|
|
||||||
|
|
||||||
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
||||||
|
|
||||||
|
@ -117,14 +110,10 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
{
|
{
|
||||||
_logger.Debug("Searching {0} for {1}-Season {2}, Prefix: {3}", Name, seriesTitle, seasonNumber, episodePrefix);
|
_logger.Debug("Searching {0} for {1}-Season {2}, Prefix: {3}", Name, seriesTitle, seasonNumber, episodePrefix);
|
||||||
|
|
||||||
var result = new List<EpisodeParseResult>();
|
|
||||||
|
|
||||||
var searchUrls = GetPartialSeasonSearchUrls(GetQueryTitle(seriesTitle), seasonNumber, episodePrefix);
|
var searchUrls = GetPartialSeasonSearchUrls(GetQueryTitle(seriesTitle), seasonNumber, episodePrefix);
|
||||||
|
|
||||||
foreach (var url in searchUrls)
|
var result = Fetch(searchUrls);
|
||||||
{
|
|
||||||
result.AddRange(Fetch(url));
|
|
||||||
}
|
|
||||||
|
|
||||||
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
||||||
|
|
||||||
|
@ -136,14 +125,9 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
{
|
{
|
||||||
_logger.Debug("Searching {0} for {1}-S{2:00}E{3:00}", Name, seriesTitle, seasonNumber, episodeNumber);
|
_logger.Debug("Searching {0} for {1}-S{2:00}E{3:00}", Name, seriesTitle, seasonNumber, episodeNumber);
|
||||||
|
|
||||||
var result = new List<EpisodeParseResult>();
|
|
||||||
|
|
||||||
var searchUrls = GetEpisodeSearchUrls(GetQueryTitle(seriesTitle), seasonNumber, episodeNumber);
|
var searchUrls = GetEpisodeSearchUrls(GetQueryTitle(seriesTitle), seasonNumber, episodeNumber);
|
||||||
|
|
||||||
foreach (var url in searchUrls)
|
var result = Fetch(searchUrls);
|
||||||
{
|
|
||||||
result.AddRange(Fetch(url));
|
|
||||||
}
|
|
||||||
|
|
||||||
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
||||||
|
|
||||||
|
@ -156,14 +140,9 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
{
|
{
|
||||||
_logger.Debug("Searching {0} for {1}-{2}", Name, seriesTitle, airDate.ToShortDateString());
|
_logger.Debug("Searching {0} for {1}-{2}", Name, seriesTitle, airDate.ToShortDateString());
|
||||||
|
|
||||||
var result = new List<EpisodeParseResult>();
|
|
||||||
|
|
||||||
var searchUrls = GetDailyEpisodeSearchUrls(GetQueryTitle(seriesTitle), airDate);
|
var searchUrls = GetDailyEpisodeSearchUrls(GetQueryTitle(seriesTitle), airDate);
|
||||||
|
|
||||||
foreach (var url in searchUrls)
|
var result = Fetch(searchUrls);
|
||||||
{
|
|
||||||
result.AddRange(Fetch(url));
|
|
||||||
}
|
|
||||||
|
|
||||||
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
result = result.Where(e => e.CleanTitle == Parser.NormalizeTitle(seriesTitle)).ToList();
|
||||||
|
|
||||||
|
@ -172,40 +151,43 @@ namespace NzbDrone.Core.Providers.Indexer
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private IEnumerable<EpisodeParseResult> Fetch(string url)
|
private List<EpisodeParseResult> Fetch(IEnumerable<string> urls)
|
||||||
{
|
{
|
||||||
var result = new List<EpisodeParseResult>();
|
var result = new List<EpisodeParseResult>();
|
||||||
|
|
||||||
try
|
foreach (var url in urls)
|
||||||
{
|
{
|
||||||
_logger.Trace("Downloading RSS " + url);
|
try
|
||||||
|
|
||||||
var reader = new SyndicationFeedXmlReader(_httpProvider.DownloadStream(url, Credentials));
|
|
||||||
var feed = SyndicationFeed.Load(reader).Items;
|
|
||||||
|
|
||||||
foreach (var item in feed)
|
|
||||||
{
|
{
|
||||||
try
|
_logger.Trace("Downloading RSS " + url);
|
||||||
{
|
|
||||||
var parsedEpisode = ParseFeed(item);
|
|
||||||
if (parsedEpisode != null)
|
|
||||||
{
|
|
||||||
parsedEpisode.NzbUrl = NzbDownloadUrl(item);
|
|
||||||
parsedEpisode.Indexer = Name;
|
|
||||||
parsedEpisode.NzbTitle = item.Title.Text;
|
|
||||||
result.Add(parsedEpisode);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
catch (Exception itemEx)
|
|
||||||
{
|
|
||||||
_logger.ErrorException("An error occurred while processing feed item", itemEx);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
var reader = new SyndicationFeedXmlReader(_httpProvider.DownloadStream(url, Credentials));
|
||||||
|
var feed = SyndicationFeed.Load(reader).Items;
|
||||||
|
|
||||||
|
foreach (var item in feed)
|
||||||
|
{
|
||||||
|
try
|
||||||
|
{
|
||||||
|
var parsedEpisode = ParseFeed(item);
|
||||||
|
if (parsedEpisode != null)
|
||||||
|
{
|
||||||
|
parsedEpisode.NzbUrl = NzbDownloadUrl(item);
|
||||||
|
parsedEpisode.Indexer = Name;
|
||||||
|
parsedEpisode.NzbTitle = item.Title.Text;
|
||||||
|
result.Add(parsedEpisode);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
catch (Exception itemEx)
|
||||||
|
{
|
||||||
|
_logger.ErrorException("An error occurred while processing feed item", itemEx);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
catch (Exception feedEx)
|
||||||
|
{
|
||||||
|
_logger.ErrorException("An error occurred while processing feed", feedEx);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
catch (Exception feedEx)
|
|
||||||
{
|
|
||||||
_logger.ErrorException("An error occurred while processing feed", feedEx);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return result;
|
return result;
|
||||||
|
|
Loading…
Reference in New Issue