https://github.com/lorien/domselect
Universal interface to work with DOM built by different HTML parsing engines.
https://github.com/lorien/domselect
css dom html html-parser html-parsing lexbor selectolax selector selectors selectors-api xpath
Last synced: 3 months ago
JSON representation
Universal interface to work with DOM built by different HTML parsing engines.
- Host: GitHub
- URL: https://github.com/lorien/domselect
- Owner: lorien
- License: mit
- Created: 2025-08-30T15:55:37.000Z (4 months ago)
- Default Branch: main
- Last Pushed: 2025-09-14T23:09:26.000Z (4 months ago)
- Last Synced: 2025-09-15T01:08:50.898Z (4 months ago)
- Topics: css, dom, html, html-parser, html-parsing, lexbor, selectolax, selector, selectors, selectors-api, xpath
- Language: Python
- Homepage:
- Size: 49.8 KB
- Stars: 2
- Watchers: 0
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Domselect
Domselect provides univeral interface to work with structure of HTML document built by one of supported HTML
processing engines. To work with HTML document you have to create so-called selector object from raw content of HTML document.
That selector will be bound to the root node of HTML structure. Then you can call different methods of these selector
to build other selectors bound to nested parts of HTML structure.
Selector object extracts low-level nodes from DOM constructed by HTML processing backend and wraps them
into high-level selector interface. If you need, you can always access low-level node stored in selector object.
### Selector Backends
Domselect library provides these selectors:
1. LexborSelector powered by [selectolax](https://github.com/rushter/selectolax)
and [lexbor](https://github.com/lexbor/lexbor) libraries. The type of raw node is `selectolax.lexbor.LexborNode`.
Query language is CSS. Lexbor parser is x3-x4 times faster than lxml parser.
2. LxmlCssSelector powered by [lxml](https://github.com/lxml/lxml) library. The type of raw node is `lxml.html.HtmlElement`.
Query language is CSS.
2. LxmlXpathSelector powered by [lxml](https://github.com/lxml/lxml) library. The type of raw node is `lxml.html.HtmlElement`.
Query language is XPATH.
### Selector Creating
To create lexbor selector from content of HTML document:
```python
from domselect import LexborSelector
sel = LexborSelector.from_content("
test")
```
Also you can create selector from raw node:
```python
from domselect import LexborSelector
from selectolax.lexbor import LexborHTMLParser
node = LexborHTMLParser("
test").css_first("div")
sel = LexborSelector(node)
```
Same goes for lxml backend. Here is an example of creating lxml selector from raw node:
```python
from lxml.html import fromstring
from domselect import LxmlCssSelector, LxmlXpathSelector
node = fromstring("
test")
sel = LxmlCssSelector(node)
# or
sel = LxmlXpathSelector(node)
```
### Node Traversal Methods
Each of these methods return other selectors of same type i.e. LexborSelector return
other LexborSelectors and LxmlCssSelector returns other LxmlCssSelectors.
Method `find(query: str)` returns list of selectors bound to raw nodes found by query.
Method `first(query: str)` returns `None` of selector bound to first raw node found by query.
There is similar `find_raw` and `first_raw` methods which works in same way but returns low-level raw nodes
i.e. they do not wrap found nodes into selector interface.
Method `parent()` returns selector bound to raw node which is parent to raw node of current selector.
Method `exists(query: str)` returns boolean flag indicates if any node has been found by query.
Method `first_contains(query: str, pattern: str[, default: None])` returns selector bound to first raw node
found by query and which contains text as `pattern` parameter. If node is not found then
`NodeNotFoundError` is raised. You can pass `default=None` optional parameter to return `None` in case
of node is not found.
### Node Properties Methods
Method `attr(name: str[, default: None|str])` returns content of node's attribute of given name.
If node does not have such attribute the `AttributeNotFoundError` is raised. If you pass optional
`default: None|str` parameter the method will return `None` or `str` if attribute does not exists.
Method `text([strip: bool])` returns text content of current node and all its sub-nodes. By default
returned text is stripped at beginning and ending from whitespaces, tabulations and line-breaks. You
can turn off striping by passing `strip=False` parameter.
Method `tag()` returns tag name of raw node to which current selector is bound.
### Traversal and Properties Methods
These methods combine two operations: search node by query and do something on found node. They are helful
if you want to get text or attribute from found node, but this node might not exist. Such methods allows you
to return reasonable default value in case node is not found. On contrary, if you use call chain like `first().text()`
then you'll not be able to return default value from `text()` call because `first()` will raise Exception if
node is not found.
Method `first_attr(query: str, name: str[, default: None|str])` returns content of attribute of given name of node
found by given query. If node does not have such attribute the `AttributeNotFoundError` is raised.
If node is not found by given query the `NodeNotFoundError` is raised. If you pass optional
`default: None|str` parameter the method will return `None` or `str` instead of rasing exceptions.
Method `first_text(query: str[, default: None|str, strip: bool])` returns text content of raw node (and all its
sub-nodes) found by given query. If node is not found the `NodeNotFoundError` is raised. Use optional `default: None|str`
parametere to return `None` or `str` instead of raising exceptions. You can control text stripping with `strip`
parameter (see description of `text()` method).
### Usage example
This code downloads telegram channel preview page and parse external links from it.
```python
from html import unescape
from urllib.request import urlopen
from domselect import LexborSelector
content = urlopen("https://t.me/s/centralbank_russia").read()
sel = LexborSelector.from_content(content)
for msg_node in sel.find(".tgme_widget_message_wrap"):
msg_date = msg_node.first_attr(
".tgme_widget_message_date time", "datetime"
)
for text_node in msg_node.find(".tgme_widget_message_text"):
print("Message by {}".format(msg_date))
for link_node in text_node.find("a[href]"):
url = link_node.attr("href")
if url.startswith("http"):
print(" - {}".format(unescape(url)))
```
If you prefer XPATH, here is same task implemented with LxmlXpathSelector:
```python
from html import unescape
from urllib.request import urlopen
from domselect import LxmlXpathSelector
content = urlopen("https://t.me/s/centralbank_russia").read()
sel = LxmlXpathSelector.from_content(content)
for msg_node in sel.find('//*[contains(@class, "tgme_widget_message_wrap")]'):
msg_date = msg_node.first_attr(
'.//*[contains(@class, "tgme_widget_message_date")]/time', "datetime"
)
for text_node in msg_node.find(
'.//*[contains(@class, "tgme_widget_message_text")]'
):
print("Message by {}".format(msg_date))
for link_node in text_node.find(".//a[@href]"):
url = link_node.attr("href")
if url.startswith("http"):
print(" - {}".format(unescape(url)))
```