Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/grangier/python-goose
Html Content / Article Extractor, web scrapping lib in Python
https://github.com/grangier/python-goose
Last synced: about 2 months ago
JSON representation
Html Content / Article Extractor, web scrapping lib in Python
- Host: GitHub
- URL: https://github.com/grangier/python-goose
- Owner: grangier
- License: apache-2.0
- Created: 2012-01-08T20:52:44.000Z (almost 13 years ago)
- Default Branch: develop
- Last Pushed: 2021-12-26T04:28:59.000Z (almost 3 years ago)
- Last Synced: 2024-10-03T06:25:17.333Z (2 months ago)
- Language: HTML
- Homepage:
- Size: 3.73 MB
- Stars: 3,976
- Watchers: 202
- Forks: 787
- Open Issues: 107
-
Metadata Files:
- Readme: README.rst
- License: LICENSE.txt
Awesome Lists containing this project
- awesome-projects - python-goose - Html Content / Article Extractor, web scrapping lib in Python (HTML)
- starred-awesome - python-goose - Html Content / Article Extractor, web scrapping lib in Python (HTML)
README
Python-Goose - Article Extractor |Build Status|
===============================================Intro
-----Goose was originally an article extractor written in Java that has most
recently (Aug2011) been converted to a `scala project `_.This is a complete rewrite in Python. The aim of the software is to
take any news article or article-type web page and not only extract what
is the main body of the article but also all meta data and most probable
image candidate.Goose will try to extract the following information:
- Main text of an article
- Main image of article
- Any YouTube/Vimeo movies embedded in article
- Meta Description
- Meta tagsThe Python version was rewritten by:
- Xavier Grangier
Licensing
---------If you find Goose useful or have issues please drop me a line. I'd love
to hear how you're using it or what features should be improved.Goose is licensed by Gravity.com under the Apache 2.0 license; see the
LICENSE file for more details.Setup
-----::
mkvirtualenv --no-site-packages goose
git clone https://github.com/grangier/python-goose.git
cd python-goose
pip install -r requirements.txt
python setup.py installTake it for a spin
------------------::
>>> from goose import Goose
>>> url = 'http://edition.cnn.com/2012/02/22/world/europe/uk-occupy-london/index.html?hpt=ieu_c2'
>>> g = Goose()
>>> article = g.extract(url=url)
>>> article.title
u'Occupy London loses eviction fight'
>>> article.meta_description
"Occupy London protesters who have been camped outside the landmark St. Paul's Cathedral for the past four months lost their court bid to avoid eviction Wednesday in a decision made by London's Court of Appeal."
>>> article.cleaned_text[:150]
(CNN) -- Occupy London protesters who have been camped outside the landmark St. Paul's Cathedral for the past four months lost their court bid to avoi
>>> article.top_image.src
http://i2.cdn.turner.com/cnn/dam/assets/111017024308-occupy-london-st-paul-s-cathedral-story-top.jpgConfiguration
-------------There are two ways to pass configuration to goose. The first one is to
pass goose a Configuration() object. The second one is to pass a
configuration dict.For instance, if you want to change the userAgent used by Goose just
pass:::
>>> g = Goose({'browser_user_agent': 'Mozilla'})
Switching parsers : Goose can now be used with lxml html parser or lxml
soup parser. By default the html parser is used. If you want to use the
soup parser pass it in the configuration dict :::
>>> g = Goose({'browser_user_agent': 'Mozilla', 'parser_class':'soup'})
Goose is now language aware
---------------------------For example, scraping a Spanish content page with correct meta language
tags:::
>>> from goose import Goose
>>> url = 'http://sociedad.elpais.com/sociedad/2012/10/27/actualidad/1351332873_157836.html'
>>> g = Goose()
>>> article = g.extract(url=url)
>>> article.title
u'Las listas de espera se agravan'
>>> article.cleaned_text[:150]
u'Los recortes pasan factura a los pacientes. De diciembre de 2010 a junio de 2012 las listas de espera para operarse aumentaron un 125%. Hay m\xe1s ciudad'Some pages don't have correct meta language tags, you can force it using
configuration :::
>>> from goose import Goose
>>> url = 'http://www.elmundo.es/elmundo/2012/10/28/espana/1351388909.html'
>>> g = Goose({'use_meta_language': False, 'target_language':'es'})
>>> article = g.extract(url=url)
>>> article.cleaned_text[:150]
u'Importante golpe a la banda terrorista ETA en Francia. La Guardia Civil ha detenido en un hotel de Macon, a 70 kil\xf3metros de Lyon, a Izaskun Lesaka y 'Passing {'use\_meta\_language': False, 'target\_language':'es'} will
forcibly select Spanish.Video extraction
----------------::
>>> import goose
>>> url = 'http://www.liberation.fr/politiques/2013/08/12/journee-de-jeux-pour-ayrault-dans-les-jardins-de-matignon_924350'
>>> g = goose.Goose({'target_language':'fr'})
>>> article = g.extract(url=url)
>>> article.movies
[]
>>> article.movies[0].src
'http://sa.kewego.com/embed/vp/?language_code=fr&playerKey=1764a824c13c&configKey=dcc707ec373f&suffix=&sig=9bc77afb496s&autostart=false'
>>> article.movies[0].embed_code
''
>>> article.movies[0].embed_type
'iframe'
>>> article.movies[0].width
'476'
>>> article.movies[0].height
'357'Goose in Chinese
----------------Some users want to use Goose for Chinese content. Chinese word
segmentation is way more difficult to deal with than occidental
languages. Chinese needs a dedicated StopWord analyser that need to be
passed to the config object.::
>>> from goose import Goose
>>> from goose.text import StopWordsChinese
>>> url = 'http://www.bbc.co.uk/zhongwen/simp/chinese_news/2012/12/121210_hongkong_politics.shtml'
>>> g = Goose({'stopwords_class': StopWordsChinese})
>>> article = g.extract(url=url)
>>> print article.cleaned_text[:150]
香港行政长官梁振英在各方压力下就其大宅的违章建筑(僭建)问题到立法会接受质询,并向香港民众道歉。梁振英在星期二(12月10日)的答问大会开始之际在其演说中道歉,但强调他在违章建筑问题上没有隐瞒的意图和动机。
一些亲北京阵营议员欢迎梁振英道歉,且认为应能获得香港民众接受,但这些议员也质问梁振英有
Goose in Arabic
---------------In order to use Goose in Arabic you have to use the StopWordsArabic
class.::
>>> from goose import Goose
>>> from goose.text import StopWordsArabic
>>> url = 'http://arabic.cnn.com/2013/middle_east/8/3/syria.clashes/index.html'
>>> g = Goose({'stopwords_class': StopWordsArabic})
>>> article = g.extract(url=url)
>>> print article.cleaned_text[:150]
دمشق، سوريا (CNN) -- أكدت جهات سورية معارضة أن فصائل مسلحة معارضة لنظام الرئيس بشار الأسد وعلى صلة بـ"الجيش الحر" تمكنت من السيطرة على مستودعات للأسلGoose in Korean
----------------In order to use Goose in Korean you have to use the StopWordsKorean
class.::
>>> from goose import Goose
>>> from goose.text import StopWordsKorean
>>> url='http://news.donga.com/3/all/20131023/58406128/1'
>>> g = Goose({'stopwords_class':StopWordsKorean})
>>> article = g.extract(url=url)
>>> print article.cleaned_text[:150]
경기도 용인에 자리 잡은 민간 시험인증 전문기업 ㈜디지털이엠씨(www.digitalemc.com).
14년째 세계 각국의 통신·안전·전파 규격 시험과 인증 한 우물만 파고 있는 이 회사 박채규 대표가 만나기로 한 주인공이다.
그는 전기전자·무선통신·자동차 전장품 분야에Known issues
------------- There are some issues with unicode URLs.
- Cookie handling : Some websites need cookie handling. At the moment the only work around is to use the raw_html extraction. For instance:>>> import urllib2
>>> import goose
>>> url = "http://www.nytimes.com/2013/08/18/world/middleeast/pressure-by-us-failed-to-sway-egypts-leaders.html?hp"
>>> opener = urllib2.build_opener(urllib2.HTTPCookieProcessor())
>>> response = opener.open(url)
>>> raw_html = response.read()
>>> g = goose.Goose()
>>> a = g.extract(raw_html=raw_html)
>>> a.cleaned_text
u'CAIRO \u2014 For a moment, at least, American and European diplomats trying to defuse the volatile standoff in Egypt thought they had a breakthrough.\n\nAs t'TODO
----- Video html5 tag extraction
.. |Build Status| image:: https://travis-ci.org/grangier/python-goose.png?branch=develop :target: https://travis-ci.org/grangier/python-goose