Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/feng19/spider_man
SpiderMan,a base-on Broadway fast high-level web crawling & scraping framework for Elixir.
https://github.com/feng19/spider_man
crawler data-mining elixir erlang framework spider
Last synced: 2 days ago
JSON representation
SpiderMan,a base-on Broadway fast high-level web crawling & scraping framework for Elixir.
- Host: GitHub
- URL: https://github.com/feng19/spider_man
- Owner: feng19
- License: apache-2.0
- Created: 2021-04-01T16:22:13.000Z (almost 4 years ago)
- Default Branch: master
- Last Pushed: 2024-12-24T02:05:52.000Z (about 1 month ago)
- Last Synced: 2025-01-16T14:27:13.854Z (9 days ago)
- Topics: crawler, data-mining, elixir, erlang, framework, spider
- Language: Elixir
- Homepage:
- Size: 211 KB
- Stars: 24
- Watchers: 4
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- fucking-awesome-elixir - SpiderMan - A base-on Broadway fast high-level web crawling & scraping framework for Elixir. (HTTP)
- awesome-elixir - SpiderMan - A base-on Broadway fast high-level web crawling & scraping framework for Elixir. (HTTP)
README
# SpiderMan
[![Module Version](https://img.shields.io/hexpm/v/spider_man.svg)](https://hex.pm/packages/spider_man)
[![Hex Docs](https://img.shields.io/badge/hex-docs-lightgreen.svg)](https://hexdocs.pm/spider_man/)
[![Total Download](https://img.shields.io/hexpm/dt/spider_man.svg)](https://hex.pm/packages/spider_man)
[![License](https://img.shields.io/hexpm/l/spider_man.svg)](https://github.com/feng19/spider_man/blob/master/LICENSE)
[![Last Updated](https://img.shields.io/github/last-commit/feng19/spider_man.svg)](https://github.com/feng19/spider_man/commits/master)**SpiderMan,a fast high-level web crawling & scraping framework for Elixir.**
Inspired by [Crawly](https://github.com/elixir-crawly/crawly)(Elixir) and [Scrapy](https://github.com/scrapy/scrapy)(Python).
## Usage
This example show how to crawly data from
[elixir-jobs](https://elixirjobs.net/) website by use SpiderMan.### Use in livebook
[![Run in Livebook](https://livebook.dev/badge/v1/blue.svg)](https://livebook.dev/run?url=https%3A%2F%2Fhexdocs.pm%2Fspider_man%2Felixirjobs.livemd)
### Use in script or mix project
```elixir
Mix.install([
{:spider_man, "~> 0.6"},
{:floki, "~> 0.34"},
{:nimble_csv, "~> 1.2"}
])defmodule SpiderList.ElixirJobs do
@moduledoc false
use SpiderMan
require Logger
alias SpiderMan.Response@base_url "https://elixirjobs.net/"
def run do
SpiderMan.run_until_zero(__MODULE__, [], 3_000)
end@impl true
def settings do
requester_options = [
base_url: @base_url,
middlewares: [
{SpiderMan.Middleware.UserAgent,
[
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/90.0.4430.93 Safari/537.36",
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.114 Safari/537.36",
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.82 Safari/537.36",
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/88.0.4389.82 Safari/537.36",
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4389.82 Safari/537.36",
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.94 Safari/537.36"
]},
{Tesla.Middleware.Headers,
[
{"referer", @base_url},
{"accept",
"text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.9"},
{"accept-encoding", "gzip, deflate"},
{"accept-language", "zh-CN,zh;q=0.9,zh-TW;q=0.8,en;q=0.7"}
]},
Tesla.Middleware.DecompressResponse
]
][
downloader_options: [requester: {SpiderMan.Requester.Finch, requester_options}],
spider_options: [pipelines: []],
item_processor_options: [
storage: [
{SpiderMan.Storage.ETS, "./data/jobs.ets"},
{SpiderMan.Storage.CSV,
file: "./data/jobs.csv",
headers: [
:link,
:title,
:sub_title,
:date,
:workplace,
:type
]}
]
]
]
end@impl true
def init(state) do
build_request(@base_url)
|> set_flag(:first_page)
|> then(&SpiderMan.insert_request(__MODULE__, &1))state
end@impl true
def handle_response(%Response{env: env, flag: :first_page}, _context) do
total_page =
Regex.run(~r/Showing page 1 of (\d+)/, env.body, capture: :all_but_first)
|> hd()
|> String.to_integer()Logger.info("total: #{total_page}")
requests =
Enum.map(2..total_page, fn n ->
build_request("/?page=#{n}")
|> set_flag({:list_page, n})
end)handle_list_page(env.body, 1)
|> Map.put(:requests, requests)
enddef handle_response(%Response{env: env, flag: {:list_page, n}}, _context) do
handle_list_page(env.body, n)
enddefp handle_list_page(body, n) do
Logger.info("processing page #{n}")
{:ok, document} = Floki.parse_document(body)jobs =
Floki.find(document, ".offers-index")
|> hd()
|> Floki.children(include_text: false)
|> Enum.filter(&match?({"a", _, _}, &1))items =
Enum.map(jobs, fn job ->
title = Floki.find(job, ".title strong") |> Floki.text() |> String.trim()
sub_title = Floki.find(job, ".title small") |> Floki.text() |> String.trim()
link = Floki.attribute(job, "a", "href") |> hd()[_, date, _, workplace, _, type] =
Floki.find(job, ".control .tag")
|> Enum.map(&(&1 |> Floki.text() |> String.trim()))build_item(
link,
%{
link: @base_url <> String.slice(link, 1..-1),
title: title,
sub_title: sub_title,
date: date,
workplace: workplace,
type: type
}
)
end)%{items: items}
end
endSpiderList.ElixirJobs.run()
```copy this script and save to `elixir_jobs.exs` and then start by command:
```shell
elixir elixir_jobs.exs
```## Copyright and License
Copyright (c) 2023 feng19
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at [http://www.apache.org/licenses/LICENSE-2.0](http://www.apache.org/licenses/LICENSE-2.0)Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.