Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/mehmetozkaya/dotnetcrawler
DotnetCrawler is a straightforward, lightweight web crawling/scrapying library for Entity Framework Core output based on dotnet core. This library designed like other strong crawler libraries like WebMagic and Scrapy but for enabling extandable your custom requirements. Medium link : https://medium.com/@mehmetozkaya/creating-custom-web-crawler-with-dotnet-core-using-entity-framework-core-ec8d23f0ca7c
https://github.com/mehmetozkaya/dotnetcrawler
crawler crawling csharp ddd-architecture dotnetcore entity-framework-core htmlagilitypack scraping scrapy scrapy-crawler webcrawler webcrawler-htmlagilitypack webcrawling webscraper webscraping
Last synced: 9 days ago
JSON representation
DotnetCrawler is a straightforward, lightweight web crawling/scrapying library for Entity Framework Core output based on dotnet core. This library designed like other strong crawler libraries like WebMagic and Scrapy but for enabling extandable your custom requirements. Medium link : https://medium.com/@mehmetozkaya/creating-custom-web-crawler-with-dotnet-core-using-entity-framework-core-ec8d23f0ca7c
- Host: GitHub
- URL: https://github.com/mehmetozkaya/dotnetcrawler
- Owner: mehmetozkaya
- Created: 2019-02-19T11:47:43.000Z (over 5 years ago)
- Default Branch: master
- Last Pushed: 2022-12-20T17:31:45.000Z (almost 2 years ago)
- Last Synced: 2024-10-12T04:06:39.644Z (24 days ago)
- Topics: crawler, crawling, csharp, ddd-architecture, dotnetcore, entity-framework-core, htmlagilitypack, scraping, scrapy, scrapy-crawler, webcrawler, webcrawler-htmlagilitypack, webcrawling, webscraper, webscraping
- Language: C#
- Homepage:
- Size: 70.3 KB
- Stars: 175
- Watchers: 12
- Forks: 65
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
- Funding: .github/FUNDING.yml
Awesome Lists containing this project
README
# DotnetCrawler
DotnetCrawler is a straightforward, lightweight **web crawling/scrapying** library for **Entity Framework Core output** based on dotnet core.
This library designed like other strong crawler libraries like WebMagic and Scrapy but for enabling extandable your custom requirements.
You can find detail explanation of this library at [medium.](https://medium.com/@mehmetozkaya/creating-custom-web-crawler-with-dotnet-core-using-entity-framework-core-ec8d23f0ca7c)For detail explanation : https://medium.com/@mehmetozkaya/creating-custom-web-crawler-with-dotnet-core-using-entity-framework-core-ec8d23f0ca7c
## Getting Started
This project intented for providing EF.Core database insert but it has very scale architecture in order to implement your custom scenarios. So the main design of architecture is very common for web crawler/scrapying frameworks, you can see below image.
![c45161e7-4721-4e4e-a33f-9925617a294f](https://user-images.githubusercontent.com/1147445/55534310-b8749800-56bc-11e9-913a-7e23ef745062.Jpeg)
As per above image, in this library created project structures including DotnetCrawler.Request-Downloader-Processor-Pipeline projects.
### Usage
You can use this library using DotnetCrawler class with builder pattern in your console applications;
```csharp
var crawler = new DotnetCrawler()
.AddRequest(new DotnetCrawlerRequest { Url = "https://www.ebay.com/b/Apple-iPhone/9355/bn_319682", Regex = @".*itm/.+", TimeOut = 5000 })
.AddDownloader(new DotnetCrawlerDownloader { DownloderType = DotnetCrawlerDownloaderType.FromMemory })
.AddProcessor(new DotnetCrawlerProcessor { })
.AddPipeline(new DotnetCrawlerPipeline { });await crawler.Crawle();
```Catalog is a generic type of DotnetCrawler and also generated by EF.Core scaffolding command in .Data project. DotnetCrawler.Data project installed EF.Core nuget pagkages and run this command in Package Manager Console.
```csharp
Scaffold-DbContext "Server=(localdb)\mssqllocaldb;Database=Microsoft.eShopOnWeb.CatalogDb;Trusted_Connection=True;" Microsoft.EntityFrameworkCore.SqlServer -OutputDir Models
```
By this command DotnetCrawler.Data project created Model folder and all entities in eShopOnWeb Microsoft's example.
After that need to configure your entity class with custom crawler attributes;
```csharp
[DotnetCrawlerEntity(XPath = "//*[@id='LeftSummaryPanel']/div[1]")]
public partial class Catalog : IEntity
{
public int Id { get; set; }
[DotnetCrawlerField(Expression = "1", SelectorType = SelectorType.FixedValue)]
public int CatalogBrandId { get; set; }
[DotnetCrawlerField(Expression = "1", SelectorType = SelectorType.FixedValue)]
public int CatalogTypeId { get; set; }
public string Description { get; set; }
[DotnetCrawlerField(Expression = "//*[@id='itemTitle']/text()", SelectorType = SelectorType.XPath)]
public string Name { get; set; }
public string PictureUri { get; set; }
public decimal Price { get; set; }public virtual CatalogBrand CatalogBrand { get; set; }
public virtual CatalogType CatalogType { get; set; }
}
```
With this code, basically, crawler requested given url and try to find given attributes which defined xpath addresses for target web url.When run this code, first create your request which url want to consume in your Request object. In these request object you can also provide Regex expression in order to filter your web urls. The process runs as define Request - set Downloader - set Processor - run Pipeline.
### Example of eShopOnWeb Microsoft Project Usage
This library also include example project which name is DotnetCrawler.Sample. Basically, in this sample project, implementing Microsoft eShopOnWeb repository. You can find this repo [here](https://github.com/dotnet-architecture/eShopOnWeb).
So in this example repo, implemented e-commerce project, it has "Catalog" table when you generate with EF.Core code-first approach.
Passing "Catalog" table as a generic type in DotnetCrawler class.
```csharp
var crawler = new DotnetCrawler()
```
### DotnetWebCrawler.RequestDotnetWebCrawler's example use this Catalog table and load this table by crawling [eBay website](https://www.ebay.com/b/Apple-iPhone/9355/bn_319682). Also aplying Regex filter for crawling only phone web pages in eBay. (Regex = @".*itm/.+")
All these parameters defined into DotnetCrawlerRequest class;
```csharp
.AddRequest(new DotnetCrawlerRequest { Url = "https://www.ebay.com/b/Apple-iPhone/9355/bn_319682", Regex = @".*itm/.+", TimeOut = 5000 })
```
### DotnetWebCrawler.Downloader
DotnetWebCrawler's example use downloader which provide to use System.Web and download targeted url with various types. You can find different types of usage as below code block;```csharp
.AddDownloader(new DotnetCrawlerDownloader { DownloderType = DotnetCrawlerDownloaderType.FromFile, DownloadPath = @"C:\DotnetCrawlercrawler\" }).AddDownloader(new DotnetCrawlerDownloader { DownloderType = DotnetCrawlerDownloaderType.FromMemory })
.AddDownloader(new DotnetCrawlerDownloader { DownloderType = DotnetCrawlerDownloaderType.FromWeb })
```
Definitions of these types explains;```csharp
///
/// Type of the downloaders when crawler download source web
///
public enum DotnetCrawlerDownloaderType
{
///
/// Download to local file
///
FromFile,
///
/// Without downloading to local file, download temp and directly use
///
FromMemory,
///
/// Read direct from web
///
FromWeb
}
```
### DotnetWebCrawler.ProcessorDotnetWebCrawler's example use processor which provide to use HtmlAgilityPack and process html pages as per intented xpaths.
```csharp
.AddProcessor(new DotnetCrawlerProcessor { })
```
DotnetWebCrawler has default processor which name is DotnetCrawlerProcessor.cs. You can pass EF.Core entity if you provide attributes and derived from IEntity.cs.
In addition you can create your own processor in order to apply your custom scenarios.### DotnetWebCrawler.Pipeline
DotnetWebCrawler's example use pipeline which provide to use EF.Core and insert records into EF.Core database.
```csharp
.AddPipeline(new DotnetCrawlerPipeline { });
```
DotnetWebCrawler has default pipeline which name is DotnetCrawlerPipeline.cs. You can pass EF.Core entity if you provide attributes and derived from IEntity.cs
In addition you can create your own pipeline in order to apply your custom scenarios.## Requirements
Before run the program you must configure your EF.Core entities. First need to scaffolding and after that you should devired from IEntity and should configure XPath definitions with attributes on entity property fields.
First;
```csharp
Scaffold-DbContext "Server=(localdb)\mssqllocaldb;Database=Microsoft.eShopOnWeb.CatalogDb;Trusted_Connection=True;" Microsoft.EntityFrameworkCore.SqlServer -OutputDir Models
```
Second; Configure your table OnModelCreating method, change id generator strategy and change MaxLength to 200.
```csharp
protected override void OnModelCreating(ModelBuilder modelBuilder)
{
modelBuilder.Entity(entity =>
{
entity.HasIndex(e => e.CatalogBrandId);entity.HasIndex(e => e.CatalogTypeId);
//entity.Property(e => e.Id).ValueGeneratedNever();
entity.Property(e => e.Id).ForSqlServerUseSequenceHiLo("catalog_hilo").IsRequired();entity.Property(e => e.Name)
.IsRequired()
.HasMaxLength(200);
.....
```
At last, In example of eBay crawler usage Catalog table as below way;
```csharp
[DotnetCrawlerEntity(XPath = "//*[@id='LeftSummaryPanel']/div[1]")]
public partial class Catalog : IEntity
{
public int Id { get; set; }
[DotnetCrawlerField(Expression = "1", SelectorType = SelectorType.FixedValue)]
public int CatalogBrandId { get; set; }
[DotnetCrawlerField(Expression = "1", SelectorType = SelectorType.FixedValue)]
public int CatalogTypeId { get; set; }
public string Description { get; set; }
[DotnetCrawlerField(Expression = "//*[@id='itemTitle']/text()", SelectorType = SelectorType.XPath)]
public string Name { get; set; }
public string PictureUri { get; set; }
public decimal Price { get; set; }public virtual CatalogBrand CatalogBrand { get; set; }
public virtual CatalogType CatalogType { get; set; }
}
```## Background of DotnetCrawler
Development environments;
* Visual Studio 2017
* .Net Core 2.2 or later
* EF Core 2.2 or later## Next Releases
This program only imported EF.Core and using default downloader-processor-pipeline classes. And this program only solve spesific problem of customer requirements. So it will evolve a product and extent with new features as listed below;
* Extend with different database providers.
* Extend for different downloader-processor-pipeline implementations which requested with different aproaches.
* Use with hangfire, quartz schedular frameworks in order to schedule and use async functions.## Authors
* **Mehmet Ozkaya** - *Initial work* - [mehmetozkaya](https://github.com/mehmetozkaya)
See also the list of [contributors](https://github.com/your/project/contributors) who participated in this project.
Check also [github page.](https://mehmetozkaya.github.io/DotnetCrawler/)Medium : https://medium.com/@mehmetozkaya/creating-custom-web-crawler-with-dotnet-core-using-entity-framework-core-ec8d23f0ca7c
## License
This project is licensed under the MIT License - see the [LICENSE.md](LICENSE.md) file for details
## Buy Me a Coffee