WebReaper 3.0.7
See the version list below for details.
dotnet add package WebReaper --version 3.0.7
NuGet\Install-Package WebReaper -Version 3.0.7
<PackageReference Include="WebReaper" Version="3.0.7" />
paket add WebReaper --version 3.0.7
#r "nuget: WebReaper, 3.0.7"
// Install WebReaper as a Cake Addin #addin nuget:?package=WebReaper&version=3.0.7 // Install WebReaper as a Cake Tool #tool nuget:?package=WebReaper&version=3.0.7
WebReaper
Please star this project if you find it useful!
Overview
Declarative high performance web scraper in C#. Easily crawl any web site and parse the data, save structed result to a file, DB, or pretty much to anywhere you want.
It provides a simple yet extensible API to make web scrapig a breeze.
Install
dotnet add package WebReaper
Requirements
.NET 6
📋 Example:
using WebReaper.Core.Builders;
_ = new EngineBuilder("reddit")
.Get("https://www.reddit.com/r/dotnet/")
.Follow("a.SQnoC3ObvgnGjWt90zD9Z._2INHSNB8V5eaWp4P0rY_mE")
.Parse(new()
{
new("title", "._eYtD2XCVieq6emjKBH3m"),
new("text", "._3xX726aBn29LDbsDtzr_6E._1Ap4F5maDtT1E1YuCiaO0r.D3IL3FD0RFy_mkKLPwL4")
})
.WriteToJsonFile("output.json")
.LogToConsole()
.Build()
.Run();
Console.ReadLine();
Features:
- ⚡ It's extremly fast due to parallelism and asynchrony
- 🗒 Declarative parsing with a structured scheme
- 💾 Saving data to any sinks such as JSON or CSV file, MongoDB, CosmosDB, Redis, etc.
- 🌎 Distributed crawling support: run your web scraper on ony cloud VMs, serverless functions, on-prem servers, etc.
- 🐙 Crowling and parsing Single Page Applications with Puppeteer
- 🖥 Proxy support
- 🌀 Automatic reties
Usage examples
- Data mining
- Gathering data for machine learning
- Online price change monitoring and price comparison
- News aggregation
- Product review scraping (to watch the competition)
- Gathering real estate listings
- Tracking online presence and reputation
- Web mashup and web data integration
- MAP compliance
- Lead generation
API overview
SPA parsing example
Parsing single page applications is super simple, just use the GetWithBrowser and/or FollowWithBrowser method. In this case Puppeteer will be used to load the pages.
_ = new EngineBuilder("reddit")
.GetWithBrowser("https://www.reddit.com/r/dotnet/")
.Follow("a.SQnoC3ObvgnGjWt90zD9Z._2INHSNB8V5eaWp4P0rY_mE")
.Parse(new()
{
new("title", "._eYtD2XCVieq6emjKBH3m"),
new("text", "._3xX726aBn29LDbsDtzr_6E._1Ap4F5maDtT1E1YuCiaO0r.D3IL3FD0RFy_mkKLPwL4")
})
.WriteToJsonFile("output.json")
.LogToConsole()
.Build()
.Run(1);
Additionaly, you can run any JavaScript on dynamic pages as they are loaded with headless browser. In order to do that you need to add some page actions:
using WebReaper.Core.Builders;
_ = new EngineBuilder("reddit")
.GetWithBrowser("https://www.reddit.com/r/dotnet/", actions => actions
.ScrollToEnd()
.Build())
.Follow("a.SQnoC3ObvgnGjWt90zD9Z._2INHSNB8V5eaWp4P0rY_mE")
.Parse(new()
{
new("title", "._eYtD2XCVieq6emjKBH3m"),
new("text", "._3xX726aBn29LDbsDtzr_6E._1Ap4F5maDtT1E1YuCiaO0r.D3IL3FD0RFy_mkKLPwL4")
})
.WriteToJsonFile("output.json")
.LogToConsole()
.Build()
.Run(1);
Console.ReadLine();
It can be helpful if the required content is loaded only after some user interactions such as clicks, scrolls, etc.
Authorization
If you need to pass authorization before parsing the web site, you can call Authorize method on Scraper that has to return CookieContainer with all cookies required for authorization. You are responsible for performing the login operation with your credentials, the Scraper only uses the cookies that you provide.
_ = new ScraperEngineBuilder("rutracker")
.WithLogger(logger)
.Get("https://rutracker.org/forum/index.php?c=33")
.Authorize(() =>
{
var container = new CookieContainer();
container.Add(new Cookie("AuthToken", "123");
return container;
})
Distributed web scraping with Serverless approach
In the Examples folder you can find the project called WebReaper.AzureFuncs. It demonstrates the use of WebReaper with Azure Functions. It consists of two serverless functions:
StartScrapting
First of all, this function uses ScraperConfigBuilder to build the scraper configuration e. g.:
Secondly, this function writes the first web scraping job with startUrl to the Azure Service Bus queue:
WebReaperSpider
This Azure function is triggered by messages sent to the Azure Service Bus queue. Messages represent web scraping job.
Firstly, this function builds the spider that is going to execute the job from the queue.
Secondly, it executes the job by loading the page, parsing content, saving to the database, etc.
Finally, it iterates through these new jobs and sends them the the Job queue.
Extensibility
Adding a new sink to persist your data
Out of the box there are 4 sinks you can send your parsed data to: ConsoleSink, CsvFileSink, JsonFileSink, CosmosSink (Azure Cosmos database).
You can easly add your own by implementing the IScraperSink interface:
public interface IScraperSink
{
public Task EmitAsync(JObject scrapedData);
}
Here is an example of the Console sink:
public class ConsoleSink : IScraperSink
{
public Task EmitAsync(JObject scrapedData)
{
Console.WriteLine($"{scrapedData.ToString()}");
return Task.CompletedTask;
}
}
The scrapedData parameter is JSON object that contains scraped data that you specified in your schema.
Adding your sink to the Scraper is simple, just call AddSink method on the Scraper:
_ = new ScraperEngineBuilder("rutracker")
.AddSink(new ConsoleSink());
.Get("https://rutracker.org/forum/index.php?c=33")
.Follow("#cf-33 .forumlink>a")
.Follow(".forumlink>a")
.Paginate("a.torTopic", ".pg")
.Parse(new() {
new("name", "#topic-title"),
});
For other ways to extend your functionality see the next section.
Intrefaces
Interface | Description |
---|---|
IScheduler | Reading and writing from the job queue. By default, the in-memory queue is used, but you can provider your implementation |
ICrawledLinkTracker | Tracker of visited links. A default implementation is an in-memory tracker. You can provide your own for Redis, MongoDB, etc. |
IPageLoader | Loader that takes URL and returns HTML of the page as a string |
IContentParser | Takes HTML and schema and returns JSON representation (JObject). |
ILinkParser | Takes HTML as a string and returns page links |
IScraperSink | Represents a data store for writing the results of web scraping. Takes the JObject as parameter |
ISpider | A spider that does the crawling, parsing, and saving of the data |
Main entities
- Job - a record that represends a job for the spider
- LinkPathSelector - represents a selector for links to be crawled
Repository structure
Project | Description |
---|---|
WebReaper | Library for web scraping |
WebReaper.ScraperWorkerService | Example of using WebReaper library in a Worker Service .NET project. |
WebReaper.DistributedScraperWorkerService | Example of using WebReaper library in a distributed way wih Azure Service Bus |
WebReaper.AzureFuncs | Example of using WebReaper library with serverless approach using Azure Functions |
WebReaper.ConsoleApplication | Example of using WebReaper library with in a console application |
Coming soon:
- Nuget package
- Azure functions for the distributed crawling
- Parsing lists
- Loading pages with headless browser and flexible SPA page manipulations (clicks, scrolls, etc)
- Proxy support
- Add flexible conditions for ignoring or allowing certain pages
- Breadth first traversal with priprity channels
- Save auth cookies to redis
- Rest API example for web scraping
- Sitemap crawling support
- Ports to NodeJS and Go
Features under consideration
- Imbedded http server for monitoring, logs and statistics
- Saving logs to Seq
- Add LogTo method with Console and File support
- Site API support
- CRON for scheduling
- Request auto throttling
- Add bloom filter for revisiting same urls
- Simplify WebReaperSpider class
- Subscribe to logs with lambda expression
See the LICENSE file for license rights and limitations (GNU GPLv3).
License
Product | Versions Compatible and additional computed target framework versions. |
---|---|
.NET | net6.0 is compatible. net6.0-android was computed. net6.0-ios was computed. net6.0-maccatalyst was computed. net6.0-macos was computed. net6.0-tvos was computed. net6.0-windows was computed. net7.0 was computed. net7.0-android was computed. net7.0-ios was computed. net7.0-maccatalyst was computed. net7.0-macos was computed. net7.0-tvos was computed. net7.0-windows was computed. net8.0 was computed. net8.0-android was computed. net8.0-browser was computed. net8.0-ios was computed. net8.0-maccatalyst was computed. net8.0-macos was computed. net8.0-tvos was computed. net8.0-windows was computed. |
-
net6.0
- Azure.Messaging.ServiceBus (>= 7.11.0)
- Fizzler.Systems.HtmlAgilityPack (>= 1.2.1)
- Microsoft.Azure.Cosmos (>= 3.31.1)
- Microsoft.Extensions.Http (>= 6.0.0)
- Microsoft.Extensions.Logging.Abstractions (>= 6.0.2)
- MongoDB.Driver (>= 2.18.0)
- Newtonsoft.Json (>= 13.0.1)
- Polly (>= 7.2.3)
- PuppeteerExtraSharp (>= 1.3.2)
- PuppeteerSharp (>= 7.1.0)
- StackExchange.Redis (>= 2.6.70)
- System.Text.Encoding.CodePages (>= 6.0.0)
NuGet packages
This package is not used by any NuGet packages.
GitHub repositories
This package is not used by any popular GitHub repositories.
Version | Downloads | Last updated |
---|---|---|
3.5.2 | 218 | 10/19/2024 |
3.5.1 | 2,827 | 8/15/2023 |
3.5.0 | 148 | 8/9/2023 |
3.4.0 | 248 | 4/17/2023 |
3.3.0 | 220 | 4/3/2023 |
3.2.0 | 198 | 4/2/2023 |
3.1.0 | 320 | 2/28/2023 |
3.0.8 | 435 | 11/12/2022 |
3.0.7 | 357 | 11/4/2022 |
3.0.6 | 320 | 11/3/2022 |
3.0.5 | 347 | 10/31/2022 |
3.0.4 | 368 | 10/29/2022 |
3.0.3 | 357 | 10/29/2022 |
3.0.2 | 372 | 10/24/2022 |
3.0.1 | 379 | 10/21/2022 |
3.0.0 | 382 | 10/7/2022 |