mirror of
https://github.com/g3rv4/GetMoarFediverse.git
synced 2025-02-16 09:19:32 +01:00
If we can't pull the robots information, don't break completely (#37)
This commit is contained in:
parent
32280c334b
commit
91ac0e2385
@ -41,7 +41,14 @@ await Parallel.ForEachAsync(Config.Instance.Sites,
|
|||||||
new ParallelOptions { MaxDegreeOfParallelism = Config.Instance.Sites.Length },
|
new ParallelOptions { MaxDegreeOfParallelism = Config.Instance.Sites.Length },
|
||||||
async (site, _) =>
|
async (site, _) =>
|
||||||
{
|
{
|
||||||
sitesRobotFile[site.Host] = await robotsFileParser.FromUriAsync(new Uri($"http://{site.Host}/robots.txt"));
|
try
|
||||||
|
{
|
||||||
|
sitesRobotFile[site.Host] = await robotsFileParser.FromUriAsync(new Uri($"http://{site.Host}/robots.txt"));
|
||||||
|
}
|
||||||
|
catch
|
||||||
|
{
|
||||||
|
Console.WriteLine($"Ignoring {site.Host} because had issues fetching its robots data (is the site down?)");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
);
|
);
|
||||||
|
|
||||||
@ -88,6 +95,11 @@ await Parallel.ForEachAsync(sitesTags, new ParallelOptions{MaxDegreeOfParallelis
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
Console.WriteLine($"Not scraping {url} because I couldn't fetch robots data.");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
HttpResponseMessage? response = null;
|
HttpResponseMessage? response = null;
|
||||||
try
|
try
|
||||||
|
Loading…
Reference in New Issue
Block a user