1291 lines
46 KiB
Python
1291 lines
46 KiB
Python
|
# -*- coding: utf-8 -*-
|
||
|
"""Tests for Beautiful Soup's tree traversal methods.
|
||
|
|
||
|
The tree traversal methods are the main advantage of using Beautiful
|
||
|
Soup over just using a parser.
|
||
|
|
||
|
Different parsers will build different Beautiful Soup trees given the
|
||
|
same markup, but all Beautiful Soup trees can be traversed with the
|
||
|
methods tested here.
|
||
|
"""
|
||
|
|
||
|
from pdb import set_trace
|
||
|
import pytest
|
||
|
import re
|
||
|
import warnings
|
||
|
from bs4 import BeautifulSoup
|
||
|
from bs4.builder import (
|
||
|
builder_registry,
|
||
|
HTMLParserTreeBuilder,
|
||
|
)
|
||
|
from bs4.element import (
|
||
|
CData,
|
||
|
Comment,
|
||
|
Declaration,
|
||
|
Doctype,
|
||
|
Formatter,
|
||
|
NavigableString,
|
||
|
Script,
|
||
|
SoupStrainer,
|
||
|
Stylesheet,
|
||
|
Tag,
|
||
|
TemplateString,
|
||
|
)
|
||
|
from . import (
|
||
|
SoupTest,
|
||
|
skipIf,
|
||
|
)
|
||
|
|
||
|
class TestFind(SoupTest):
|
||
|
"""Basic tests of the find() method.
|
||
|
|
||
|
find() just calls find_all() with limit=1, so it's not tested all
|
||
|
that thouroughly here.
|
||
|
"""
|
||
|
|
||
|
def test_find_tag(self):
|
||
|
soup = self.soup("<a>1</a><b>2</b><a>3</a><b>4</b>")
|
||
|
assert soup.find("b").string == "2"
|
||
|
|
||
|
def test_unicode_text_find(self):
|
||
|
soup = self.soup('<h1>Räksmörgås</h1>')
|
||
|
assert soup.find(string='Räksmörgås') == 'Räksmörgås'
|
||
|
|
||
|
def test_unicode_attribute_find(self):
|
||
|
soup = self.soup('<h1 id="Räksmörgås">here it is</h1>')
|
||
|
str(soup)
|
||
|
assert "here it is" == soup.find(id='Räksmörgås').text
|
||
|
|
||
|
|
||
|
def test_find_everything(self):
|
||
|
"""Test an optimization that finds all tags."""
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
assert 2 == len(soup.find_all())
|
||
|
|
||
|
def test_find_everything_with_name(self):
|
||
|
"""Test an optimization that finds all tags with a given name."""
|
||
|
soup = self.soup("<a>foo</a><b>bar</b><a>baz</a>")
|
||
|
assert 2 == len(soup.find_all('a'))
|
||
|
|
||
|
class TestFindAll(SoupTest):
|
||
|
"""Basic tests of the find_all() method."""
|
||
|
|
||
|
def test_find_all_text_nodes(self):
|
||
|
"""You can search the tree for text nodes."""
|
||
|
soup = self.soup("<html>Foo<b>bar</b>\xbb</html>")
|
||
|
# Exact match.
|
||
|
assert soup.find_all(string="bar") == ["bar"]
|
||
|
|
||
|
|
||
|
# Match any of a number of strings.
|
||
|
assert soup.find_all(string=["Foo", "bar"]) == ["Foo", "bar"]
|
||
|
# Match a regular expression.
|
||
|
assert soup.find_all(string=re.compile('.*')) == ["Foo", "bar", '\xbb']
|
||
|
# Match anything.
|
||
|
assert soup.find_all(string=True) == ["Foo", "bar", '\xbb']
|
||
|
|
||
|
def test_find_all_limit(self):
|
||
|
"""You can limit the number of items returned by find_all."""
|
||
|
soup = self.soup("<a>1</a><a>2</a><a>3</a><a>4</a><a>5</a>")
|
||
|
self.assert_selects(soup.find_all('a', limit=3), ["1", "2", "3"])
|
||
|
self.assert_selects(soup.find_all('a', limit=1), ["1"])
|
||
|
self.assert_selects(
|
||
|
soup.find_all('a', limit=10), ["1", "2", "3", "4", "5"])
|
||
|
|
||
|
# A limit of 0 means no limit.
|
||
|
self.assert_selects(
|
||
|
soup.find_all('a', limit=0), ["1", "2", "3", "4", "5"])
|
||
|
|
||
|
def test_calling_a_tag_is_calling_findall(self):
|
||
|
soup = self.soup("<a>1</a><b>2<a id='foo'>3</a></b>")
|
||
|
self.assert_selects(soup('a', limit=1), ["1"])
|
||
|
self.assert_selects(soup.b(id="foo"), ["3"])
|
||
|
|
||
|
def test_find_all_with_self_referential_data_structure_does_not_cause_infinite_recursion(self):
|
||
|
soup = self.soup("<a></a>")
|
||
|
# Create a self-referential list.
|
||
|
l = []
|
||
|
l.append(l)
|
||
|
|
||
|
# Without special code in _normalize_search_value, this would cause infinite
|
||
|
# recursion.
|
||
|
assert [] == soup.find_all(l)
|
||
|
|
||
|
def test_find_all_resultset(self):
|
||
|
"""All find_all calls return a ResultSet"""
|
||
|
soup = self.soup("<a></a>")
|
||
|
result = soup.find_all("a")
|
||
|
assert hasattr(result, "source")
|
||
|
|
||
|
result = soup.find_all(True)
|
||
|
assert hasattr(result, "source")
|
||
|
|
||
|
result = soup.find_all(string="foo")
|
||
|
assert hasattr(result, "source")
|
||
|
|
||
|
|
||
|
class TestFindAllBasicNamespaces(SoupTest):
|
||
|
|
||
|
def test_find_by_namespaced_name(self):
|
||
|
soup = self.soup('<mathml:msqrt>4</mathml:msqrt><a svg:fill="red">')
|
||
|
assert "4" == soup.find("mathml:msqrt").string
|
||
|
assert "a" == soup.find(attrs= { "svg:fill" : "red" }).name
|
||
|
|
||
|
|
||
|
class TestFindAllByName(SoupTest):
|
||
|
"""Test ways of finding tags by tag name."""
|
||
|
|
||
|
def setup_method(self):
|
||
|
self.tree = self.soup("""<a>First tag.</a>
|
||
|
<b>Second tag.</b>
|
||
|
<c>Third <a>Nested tag.</a> tag.</c>""")
|
||
|
|
||
|
def test_find_all_by_tag_name(self):
|
||
|
# Find all the <a> tags.
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all('a'), ['First tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_by_name_and_text(self):
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all('a', string='First tag.'), ['First tag.'])
|
||
|
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all('a', string=True), ['First tag.', 'Nested tag.'])
|
||
|
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all('a', string=re.compile("tag")),
|
||
|
['First tag.', 'Nested tag.'])
|
||
|
|
||
|
|
||
|
def test_find_all_on_non_root_element(self):
|
||
|
# You can call find_all on any node, not just the root.
|
||
|
self.assert_selects(self.tree.c.find_all('a'), ['Nested tag.'])
|
||
|
|
||
|
def test_calling_element_invokes_find_all(self):
|
||
|
self.assert_selects(self.tree('a'), ['First tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_by_tag_strainer(self):
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all(SoupStrainer('a')),
|
||
|
['First tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_by_tag_names(self):
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all(['a', 'b']),
|
||
|
['First tag.', 'Second tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_by_tag_dict(self):
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all({'a' : True, 'b' : True}),
|
||
|
['First tag.', 'Second tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_by_tag_re(self):
|
||
|
self.assert_selects(
|
||
|
self.tree.find_all(re.compile('^[ab]$')),
|
||
|
['First tag.', 'Second tag.', 'Nested tag.'])
|
||
|
|
||
|
def test_find_all_with_tags_matching_method(self):
|
||
|
# You can define an oracle method that determines whether
|
||
|
# a tag matches the search.
|
||
|
def id_matches_name(tag):
|
||
|
return tag.name == tag.get('id')
|
||
|
|
||
|
tree = self.soup("""<a id="a">Match 1.</a>
|
||
|
<a id="1">Does not match.</a>
|
||
|
<b id="b">Match 2.</a>""")
|
||
|
|
||
|
self.assert_selects(
|
||
|
tree.find_all(id_matches_name), ["Match 1.", "Match 2."])
|
||
|
|
||
|
def test_find_with_multi_valued_attribute(self):
|
||
|
soup = self.soup(
|
||
|
"<div class='a b'>1</div><div class='a c'>2</div><div class='a d'>3</div>"
|
||
|
)
|
||
|
r1 = soup.find('div', 'a d');
|
||
|
r2 = soup.find('div', re.compile(r'a d'));
|
||
|
r3, r4 = soup.find_all('div', ['a b', 'a d']);
|
||
|
assert '3' == r1.string
|
||
|
assert '3' == r2.string
|
||
|
assert '1' == r3.string
|
||
|
assert '3' == r4.string
|
||
|
|
||
|
|
||
|
class TestFindAllByAttribute(SoupTest):
|
||
|
|
||
|
def test_find_all_by_attribute_name(self):
|
||
|
# You can pass in keyword arguments to find_all to search by
|
||
|
# attribute.
|
||
|
tree = self.soup("""
|
||
|
<a id="first">Matching a.</a>
|
||
|
<a id="second">
|
||
|
Non-matching <b id="first">Matching b.</b>a.
|
||
|
</a>""")
|
||
|
self.assert_selects(tree.find_all(id='first'),
|
||
|
["Matching a.", "Matching b."])
|
||
|
|
||
|
def test_find_all_by_utf8_attribute_value(self):
|
||
|
peace = "םולש".encode("utf8")
|
||
|
data = '<a title="םולש"></a>'.encode("utf8")
|
||
|
soup = self.soup(data)
|
||
|
assert [soup.a] == soup.find_all(title=peace)
|
||
|
assert [soup.a] == soup.find_all(title=peace.decode("utf8"))
|
||
|
assert [soup.a], soup.find_all(title=[peace, "something else"])
|
||
|
|
||
|
def test_find_all_by_attribute_dict(self):
|
||
|
# You can pass in a dictionary as the argument 'attrs'. This
|
||
|
# lets you search for attributes like 'name' (a fixed argument
|
||
|
# to find_all) and 'class' (a reserved word in Python.)
|
||
|
tree = self.soup("""
|
||
|
<a name="name1" class="class1">Name match.</a>
|
||
|
<a name="name2" class="class2">Class match.</a>
|
||
|
<a name="name3" class="class3">Non-match.</a>
|
||
|
<name1>A tag called 'name1'.</name1>
|
||
|
""")
|
||
|
|
||
|
# This doesn't do what you want.
|
||
|
self.assert_selects(tree.find_all(name='name1'),
|
||
|
["A tag called 'name1'."])
|
||
|
# This does what you want.
|
||
|
self.assert_selects(tree.find_all(attrs={'name' : 'name1'}),
|
||
|
["Name match."])
|
||
|
|
||
|
self.assert_selects(tree.find_all(attrs={'class' : 'class2'}),
|
||
|
["Class match."])
|
||
|
|
||
|
def test_find_all_by_class(self):
|
||
|
tree = self.soup("""
|
||
|
<a class="1">Class 1.</a>
|
||
|
<a class="2">Class 2.</a>
|
||
|
<b class="1">Class 1.</b>
|
||
|
<c class="3 4">Class 3 and 4.</c>
|
||
|
""")
|
||
|
|
||
|
# Passing in the class_ keyword argument will search against
|
||
|
# the 'class' attribute.
|
||
|
self.assert_selects(tree.find_all('a', class_='1'), ['Class 1.'])
|
||
|
self.assert_selects(tree.find_all('c', class_='3'), ['Class 3 and 4.'])
|
||
|
self.assert_selects(tree.find_all('c', class_='4'), ['Class 3 and 4.'])
|
||
|
|
||
|
# Passing in a string to 'attrs' will also search the CSS class.
|
||
|
self.assert_selects(tree.find_all('a', '1'), ['Class 1.'])
|
||
|
self.assert_selects(tree.find_all(attrs='1'), ['Class 1.', 'Class 1.'])
|
||
|
self.assert_selects(tree.find_all('c', '3'), ['Class 3 and 4.'])
|
||
|
self.assert_selects(tree.find_all('c', '4'), ['Class 3 and 4.'])
|
||
|
|
||
|
def test_find_by_class_when_multiple_classes_present(self):
|
||
|
tree = self.soup("<gar class='foo bar'>Found it</gar>")
|
||
|
|
||
|
f = tree.find_all("gar", class_=re.compile("o"))
|
||
|
self.assert_selects(f, ["Found it"])
|
||
|
|
||
|
f = tree.find_all("gar", class_=re.compile("a"))
|
||
|
self.assert_selects(f, ["Found it"])
|
||
|
|
||
|
# If the search fails to match the individual strings "foo" and "bar",
|
||
|
# it will be tried against the combined string "foo bar".
|
||
|
f = tree.find_all("gar", class_=re.compile("o b"))
|
||
|
self.assert_selects(f, ["Found it"])
|
||
|
|
||
|
def test_find_all_with_non_dictionary_for_attrs_finds_by_class(self):
|
||
|
soup = self.soup("<a class='bar'>Found it</a>")
|
||
|
|
||
|
self.assert_selects(soup.find_all("a", re.compile("ba")), ["Found it"])
|
||
|
|
||
|
def big_attribute_value(value):
|
||
|
return len(value) > 3
|
||
|
|
||
|
self.assert_selects(soup.find_all("a", big_attribute_value), [])
|
||
|
|
||
|
def small_attribute_value(value):
|
||
|
return len(value) <= 3
|
||
|
|
||
|
self.assert_selects(
|
||
|
soup.find_all("a", small_attribute_value), ["Found it"])
|
||
|
|
||
|
def test_find_all_with_string_for_attrs_finds_multiple_classes(self):
|
||
|
soup = self.soup('<a class="foo bar"></a><a class="foo"></a>')
|
||
|
a, a2 = soup.find_all("a")
|
||
|
assert [a, a2], soup.find_all("a", "foo")
|
||
|
assert [a], soup.find_all("a", "bar")
|
||
|
|
||
|
# If you specify the class as a string that contains a
|
||
|
# space, only that specific value will be found.
|
||
|
assert [a] == soup.find_all("a", class_="foo bar")
|
||
|
assert [a] == soup.find_all("a", "foo bar")
|
||
|
assert [] == soup.find_all("a", "bar foo")
|
||
|
|
||
|
def test_find_all_by_attribute_soupstrainer(self):
|
||
|
tree = self.soup("""
|
||
|
<a id="first">Match.</a>
|
||
|
<a id="second">Non-match.</a>""")
|
||
|
|
||
|
strainer = SoupStrainer(attrs={'id' : 'first'})
|
||
|
self.assert_selects(tree.find_all(strainer), ['Match.'])
|
||
|
|
||
|
def test_find_all_with_missing_attribute(self):
|
||
|
# You can pass in None as the value of an attribute to find_all.
|
||
|
# This will match tags that do not have that attribute set.
|
||
|
tree = self.soup("""<a id="1">ID present.</a>
|
||
|
<a>No ID present.</a>
|
||
|
<a id="">ID is empty.</a>""")
|
||
|
self.assert_selects(tree.find_all('a', id=None), ["No ID present."])
|
||
|
|
||
|
def test_find_all_with_defined_attribute(self):
|
||
|
# You can pass in None as the value of an attribute to find_all.
|
||
|
# This will match tags that have that attribute set to any value.
|
||
|
tree = self.soup("""<a id="1">ID present.</a>
|
||
|
<a>No ID present.</a>
|
||
|
<a id="">ID is empty.</a>""")
|
||
|
self.assert_selects(
|
||
|
tree.find_all(id=True), ["ID present.", "ID is empty."])
|
||
|
|
||
|
def test_find_all_with_numeric_attribute(self):
|
||
|
# If you search for a number, it's treated as a string.
|
||
|
tree = self.soup("""<a id=1>Unquoted attribute.</a>
|
||
|
<a id="1">Quoted attribute.</a>""")
|
||
|
|
||
|
expected = ["Unquoted attribute.", "Quoted attribute."]
|
||
|
self.assert_selects(tree.find_all(id=1), expected)
|
||
|
self.assert_selects(tree.find_all(id="1"), expected)
|
||
|
|
||
|
def test_find_all_with_list_attribute_values(self):
|
||
|
# You can pass a list of attribute values instead of just one,
|
||
|
# and you'll get tags that match any of the values.
|
||
|
tree = self.soup("""<a id="1">1</a>
|
||
|
<a id="2">2</a>
|
||
|
<a id="3">3</a>
|
||
|
<a>No ID.</a>""")
|
||
|
self.assert_selects(tree.find_all(id=["1", "3", "4"]),
|
||
|
["1", "3"])
|
||
|
|
||
|
def test_find_all_with_regular_expression_attribute_value(self):
|
||
|
# You can pass a regular expression as an attribute value, and
|
||
|
# you'll get tags whose values for that attribute match the
|
||
|
# regular expression.
|
||
|
tree = self.soup("""<a id="a">One a.</a>
|
||
|
<a id="aa">Two as.</a>
|
||
|
<a id="ab">Mixed as and bs.</a>
|
||
|
<a id="b">One b.</a>
|
||
|
<a>No ID.</a>""")
|
||
|
|
||
|
self.assert_selects(tree.find_all(id=re.compile("^a+$")),
|
||
|
["One a.", "Two as."])
|
||
|
|
||
|
def test_find_by_name_and_containing_string(self):
|
||
|
soup = self.soup("<b>foo</b><b>bar</b><a>foo</a>")
|
||
|
a = soup.a
|
||
|
|
||
|
assert [a] == soup.find_all("a", string="foo")
|
||
|
assert [] == soup.find_all("a", string="bar")
|
||
|
|
||
|
def test_find_by_name_and_containing_string_when_string_is_buried(self):
|
||
|
soup = self.soup("<a>foo</a><a><b><c>foo</c></b></a>")
|
||
|
assert soup.find_all("a") == soup.find_all("a", string="foo")
|
||
|
|
||
|
def test_find_by_attribute_and_containing_string(self):
|
||
|
soup = self.soup('<b id="1">foo</b><a id="2">foo</a>')
|
||
|
a = soup.a
|
||
|
|
||
|
assert [a] == soup.find_all(id=2, string="foo")
|
||
|
assert [] == soup.find_all(id=1, string="bar")
|
||
|
|
||
|
|
||
|
class TestSmooth(SoupTest):
|
||
|
"""Test Tag.smooth."""
|
||
|
|
||
|
def test_smooth(self):
|
||
|
soup = self.soup("<div>a</div>")
|
||
|
div = soup.div
|
||
|
div.append("b")
|
||
|
div.append("c")
|
||
|
div.append(Comment("Comment 1"))
|
||
|
div.append(Comment("Comment 2"))
|
||
|
div.append("d")
|
||
|
builder = self.default_builder()
|
||
|
span = Tag(soup, builder, 'span')
|
||
|
span.append('1')
|
||
|
span.append('2')
|
||
|
div.append(span)
|
||
|
|
||
|
# At this point the tree has a bunch of adjacent
|
||
|
# NavigableStrings. This is normal, but it has no meaning in
|
||
|
# terms of HTML, so we may want to smooth things out for
|
||
|
# output.
|
||
|
|
||
|
# Since the <span> tag has two children, its .string is None.
|
||
|
assert None == div.span.string
|
||
|
|
||
|
assert 7 == len(div.contents)
|
||
|
div.smooth()
|
||
|
assert 5 == len(div.contents)
|
||
|
|
||
|
# The three strings at the beginning of div.contents have been
|
||
|
# merged into on string.
|
||
|
#
|
||
|
assert 'abc' == div.contents[0]
|
||
|
|
||
|
# The call is recursive -- the <span> tag was also smoothed.
|
||
|
assert '12' == div.span.string
|
||
|
|
||
|
# The two comments have _not_ been merged, even though
|
||
|
# comments are strings. Merging comments would change the
|
||
|
# meaning of the HTML.
|
||
|
assert 'Comment 1' == div.contents[1]
|
||
|
assert 'Comment 2' == div.contents[2]
|
||
|
|
||
|
|
||
|
class TestIndex(SoupTest):
|
||
|
"""Test Tag.index"""
|
||
|
def test_index(self):
|
||
|
tree = self.soup("""<div>
|
||
|
<a>Identical</a>
|
||
|
<b>Not identical</b>
|
||
|
<a>Identical</a>
|
||
|
|
||
|
<c><d>Identical with child</d></c>
|
||
|
<b>Also not identical</b>
|
||
|
<c><d>Identical with child</d></c>
|
||
|
</div>""")
|
||
|
div = tree.div
|
||
|
for i, element in enumerate(div.contents):
|
||
|
assert i == div.index(element)
|
||
|
with pytest.raises(ValueError):
|
||
|
tree.index(1)
|
||
|
|
||
|
|
||
|
class TestParentOperations(SoupTest):
|
||
|
"""Test navigation and searching through an element's parents."""
|
||
|
|
||
|
def setup_method(self):
|
||
|
self.tree = self.soup('''<ul id="empty"></ul>
|
||
|
<ul id="top">
|
||
|
<ul id="middle">
|
||
|
<ul id="bottom">
|
||
|
<b>Start here</b>
|
||
|
</ul>
|
||
|
</ul>''')
|
||
|
self.start = self.tree.b
|
||
|
|
||
|
|
||
|
def test_parent(self):
|
||
|
assert self.start.parent['id'] == 'bottom'
|
||
|
assert self.start.parent.parent['id'] == 'middle'
|
||
|
assert self.start.parent.parent.parent['id'] == 'top'
|
||
|
|
||
|
def test_parent_of_top_tag_is_soup_object(self):
|
||
|
top_tag = self.tree.contents[0]
|
||
|
assert top_tag.parent == self.tree
|
||
|
|
||
|
def test_soup_object_has_no_parent(self):
|
||
|
assert None == self.tree.parent
|
||
|
|
||
|
def test_find_parents(self):
|
||
|
self.assert_selects_ids(
|
||
|
self.start.find_parents('ul'), ['bottom', 'middle', 'top'])
|
||
|
self.assert_selects_ids(
|
||
|
self.start.find_parents('ul', id="middle"), ['middle'])
|
||
|
|
||
|
def test_find_parent(self):
|
||
|
assert self.start.find_parent('ul')['id'] == 'bottom'
|
||
|
assert self.start.find_parent('ul', id='top')['id'] == 'top'
|
||
|
|
||
|
def test_parent_of_text_element(self):
|
||
|
text = self.tree.find(string="Start here")
|
||
|
assert text.parent.name == 'b'
|
||
|
|
||
|
def test_text_element_find_parent(self):
|
||
|
text = self.tree.find(string="Start here")
|
||
|
assert text.find_parent('ul')['id'] == 'bottom'
|
||
|
|
||
|
def test_parent_generator(self):
|
||
|
parents = [parent['id'] for parent in self.start.parents
|
||
|
if parent is not None and 'id' in parent.attrs]
|
||
|
assert parents, ['bottom', 'middle' == 'top']
|
||
|
|
||
|
|
||
|
class ProximityTest(SoupTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
self.tree = self.soup(
|
||
|
'<html id="start"><head></head><body><b id="1">One</b><b id="2">Two</b><b id="3">Three</b></body></html>')
|
||
|
|
||
|
|
||
|
class TestNextOperations(ProximityTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
super(TestNextOperations, self).setup_method()
|
||
|
self.start = self.tree.b
|
||
|
|
||
|
def test_next(self):
|
||
|
assert self.start.next_element == "One"
|
||
|
assert self.start.next_element.next_element['id'] == "2"
|
||
|
|
||
|
def test_next_of_last_item_is_none(self):
|
||
|
last = self.tree.find(string="Three")
|
||
|
assert last.next_element == None
|
||
|
|
||
|
def test_next_of_root_is_none(self):
|
||
|
# The document root is outside the next/previous chain.
|
||
|
assert self.tree.next_element == None
|
||
|
|
||
|
def test_find_all_next(self):
|
||
|
self.assert_selects(self.start.find_all_next('b'), ["Two", "Three"])
|
||
|
self.start.find_all_next(id=3)
|
||
|
self.assert_selects(self.start.find_all_next(id=3), ["Three"])
|
||
|
|
||
|
def test_find_next(self):
|
||
|
assert self.start.find_next('b')['id'] == '2'
|
||
|
assert self.start.find_next(string="Three") == "Three"
|
||
|
|
||
|
def test_find_next_for_text_element(self):
|
||
|
text = self.tree.find(string="One")
|
||
|
assert text.find_next("b").string == "Two"
|
||
|
self.assert_selects(text.find_all_next("b"), ["Two", "Three"])
|
||
|
|
||
|
def test_next_generator(self):
|
||
|
start = self.tree.find(string="Two")
|
||
|
successors = [node for node in start.next_elements]
|
||
|
# There are two successors: the final <b> tag and its text contents.
|
||
|
tag, contents = successors
|
||
|
assert tag['id'] == '3'
|
||
|
assert contents == "Three"
|
||
|
|
||
|
class TestPreviousOperations(ProximityTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
super(TestPreviousOperations, self).setup_method()
|
||
|
self.end = self.tree.find(string="Three")
|
||
|
|
||
|
def test_previous(self):
|
||
|
assert self.end.previous_element['id'] == "3"
|
||
|
assert self.end.previous_element.previous_element == "Two"
|
||
|
|
||
|
def test_previous_of_first_item_is_none(self):
|
||
|
first = self.tree.find('html')
|
||
|
assert first.previous_element == None
|
||
|
|
||
|
def test_previous_of_root_is_none(self):
|
||
|
# The document root is outside the next/previous chain.
|
||
|
assert self.tree.previous_element == None
|
||
|
|
||
|
def test_find_all_previous(self):
|
||
|
# The <b> tag containing the "Three" node is the predecessor
|
||
|
# of the "Three" node itself, which is why "Three" shows up
|
||
|
# here.
|
||
|
self.assert_selects(
|
||
|
self.end.find_all_previous('b'), ["Three", "Two", "One"])
|
||
|
self.assert_selects(self.end.find_all_previous(id=1), ["One"])
|
||
|
|
||
|
def test_find_previous(self):
|
||
|
assert self.end.find_previous('b')['id'] == '3'
|
||
|
assert self.end.find_previous(string="One") == "One"
|
||
|
|
||
|
def test_find_previous_for_text_element(self):
|
||
|
text = self.tree.find(string="Three")
|
||
|
assert text.find_previous("b").string == "Three"
|
||
|
self.assert_selects(
|
||
|
text.find_all_previous("b"), ["Three", "Two", "One"])
|
||
|
|
||
|
def test_previous_generator(self):
|
||
|
start = self.tree.find(string="One")
|
||
|
predecessors = [node for node in start.previous_elements]
|
||
|
|
||
|
# There are four predecessors: the <b> tag containing "One"
|
||
|
# the <body> tag, the <head> tag, and the <html> tag.
|
||
|
b, body, head, html = predecessors
|
||
|
assert b['id'] == '1'
|
||
|
assert body.name == "body"
|
||
|
assert head.name == "head"
|
||
|
assert html.name == "html"
|
||
|
|
||
|
|
||
|
class SiblingTest(SoupTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
markup = '''<html>
|
||
|
<span id="1">
|
||
|
<span id="1.1"></span>
|
||
|
</span>
|
||
|
<span id="2">
|
||
|
<span id="2.1"></span>
|
||
|
</span>
|
||
|
<span id="3">
|
||
|
<span id="3.1"></span>
|
||
|
</span>
|
||
|
<span id="4"></span>
|
||
|
</html>'''
|
||
|
# All that whitespace looks good but makes the tests more
|
||
|
# difficult. Get rid of it.
|
||
|
markup = re.compile(r"\n\s*").sub("", markup)
|
||
|
self.tree = self.soup(markup)
|
||
|
|
||
|
|
||
|
class TestNextSibling(SiblingTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
super(TestNextSibling, self).setup_method()
|
||
|
self.start = self.tree.find(id="1")
|
||
|
|
||
|
def test_next_sibling_of_root_is_none(self):
|
||
|
assert self.tree.next_sibling == None
|
||
|
|
||
|
def test_next_sibling(self):
|
||
|
assert self.start.next_sibling['id'] == '2'
|
||
|
assert self.start.next_sibling.next_sibling['id'] == '3'
|
||
|
|
||
|
# Note the difference between next_sibling and next_element.
|
||
|
assert self.start.next_element['id'] == '1.1'
|
||
|
|
||
|
def test_next_sibling_may_not_exist(self):
|
||
|
assert self.tree.html.next_sibling == None
|
||
|
|
||
|
nested_span = self.tree.find(id="1.1")
|
||
|
assert nested_span.next_sibling == None
|
||
|
|
||
|
last_span = self.tree.find(id="4")
|
||
|
assert last_span.next_sibling == None
|
||
|
|
||
|
def test_find_next_sibling(self):
|
||
|
assert self.start.find_next_sibling('span')['id'] == '2'
|
||
|
|
||
|
def test_next_siblings(self):
|
||
|
self.assert_selects_ids(self.start.find_next_siblings("span"),
|
||
|
['2', '3', '4'])
|
||
|
|
||
|
self.assert_selects_ids(self.start.find_next_siblings(id='3'), ['3'])
|
||
|
|
||
|
def test_next_sibling_for_text_element(self):
|
||
|
soup = self.soup("Foo<b>bar</b>baz")
|
||
|
start = soup.find(string="Foo")
|
||
|
assert start.next_sibling.name == 'b'
|
||
|
assert start.next_sibling.next_sibling == 'baz'
|
||
|
|
||
|
self.assert_selects(start.find_next_siblings('b'), ['bar'])
|
||
|
assert start.find_next_sibling(string="baz") == "baz"
|
||
|
assert start.find_next_sibling(string="nonesuch") == None
|
||
|
|
||
|
|
||
|
class TestPreviousSibling(SiblingTest):
|
||
|
|
||
|
def setup_method(self):
|
||
|
super(TestPreviousSibling, self).setup_method()
|
||
|
self.end = self.tree.find(id="4")
|
||
|
|
||
|
def test_previous_sibling_of_root_is_none(self):
|
||
|
assert self.tree.previous_sibling == None
|
||
|
|
||
|
def test_previous_sibling(self):
|
||
|
assert self.end.previous_sibling['id'] == '3'
|
||
|
assert self.end.previous_sibling.previous_sibling['id'] == '2'
|
||
|
|
||
|
# Note the difference between previous_sibling and previous_element.
|
||
|
assert self.end.previous_element['id'] == '3.1'
|
||
|
|
||
|
def test_previous_sibling_may_not_exist(self):
|
||
|
assert self.tree.html.previous_sibling == None
|
||
|
|
||
|
nested_span = self.tree.find(id="1.1")
|
||
|
assert nested_span.previous_sibling == None
|
||
|
|
||
|
first_span = self.tree.find(id="1")
|
||
|
assert first_span.previous_sibling == None
|
||
|
|
||
|
def test_find_previous_sibling(self):
|
||
|
assert self.end.find_previous_sibling('span')['id'] == '3'
|
||
|
|
||
|
def test_previous_siblings(self):
|
||
|
self.assert_selects_ids(self.end.find_previous_siblings("span"),
|
||
|
['3', '2', '1'])
|
||
|
|
||
|
self.assert_selects_ids(self.end.find_previous_siblings(id='1'), ['1'])
|
||
|
|
||
|
def test_previous_sibling_for_text_element(self):
|
||
|
soup = self.soup("Foo<b>bar</b>baz")
|
||
|
start = soup.find(string="baz")
|
||
|
assert start.previous_sibling.name == 'b'
|
||
|
assert start.previous_sibling.previous_sibling == 'Foo'
|
||
|
|
||
|
self.assert_selects(start.find_previous_siblings('b'), ['bar'])
|
||
|
assert start.find_previous_sibling(string="Foo") == "Foo"
|
||
|
assert start.find_previous_sibling(string="nonesuch") == None
|
||
|
|
||
|
|
||
|
class TestTreeModification(SoupTest):
|
||
|
|
||
|
def test_attribute_modification(self):
|
||
|
soup = self.soup('<a id="1"></a>')
|
||
|
soup.a['id'] = 2
|
||
|
assert soup.decode() == self.document_for('<a id="2"></a>')
|
||
|
del(soup.a['id'])
|
||
|
assert soup.decode() == self.document_for('<a></a>')
|
||
|
soup.a['id2'] = 'foo'
|
||
|
assert soup.decode() == self.document_for('<a id2="foo"></a>')
|
||
|
|
||
|
def test_new_tag_creation(self):
|
||
|
builder = builder_registry.lookup('html')()
|
||
|
soup = self.soup("<body></body>", builder=builder)
|
||
|
a = Tag(soup, builder, 'a')
|
||
|
ol = Tag(soup, builder, 'ol')
|
||
|
a['href'] = 'http://foo.com/'
|
||
|
soup.body.insert(0, a)
|
||
|
soup.body.insert(1, ol)
|
||
|
assert soup.body.encode() == b'<body><a href="http://foo.com/"></a><ol></ol></body>'
|
||
|
|
||
|
def test_append_to_contents_moves_tag(self):
|
||
|
doc = """<p id="1">Don't leave me <b>here</b>.</p>
|
||
|
<p id="2">Don\'t leave!</p>"""
|
||
|
soup = self.soup(doc)
|
||
|
second_para = soup.find(id='2')
|
||
|
bold = soup.b
|
||
|
|
||
|
# Move the <b> tag to the end of the second paragraph.
|
||
|
soup.find(id='2').append(soup.b)
|
||
|
|
||
|
# The <b> tag is now a child of the second paragraph.
|
||
|
assert bold.parent == second_para
|
||
|
|
||
|
assert soup.decode() == self.document_for(
|
||
|
'<p id="1">Don\'t leave me .</p>\n'
|
||
|
'<p id="2">Don\'t leave!<b>here</b></p>'
|
||
|
)
|
||
|
|
||
|
def test_replace_with_returns_thing_that_was_replaced(self):
|
||
|
text = "<a></a><b><c></c></b>"
|
||
|
soup = self.soup(text)
|
||
|
a = soup.a
|
||
|
new_a = a.replace_with(soup.c)
|
||
|
assert a == new_a
|
||
|
|
||
|
def test_unwrap_returns_thing_that_was_replaced(self):
|
||
|
text = "<a><b></b><c></c></a>"
|
||
|
soup = self.soup(text)
|
||
|
a = soup.a
|
||
|
new_a = a.unwrap()
|
||
|
assert a == new_a
|
||
|
|
||
|
def test_replace_with_and_unwrap_give_useful_exception_when_tag_has_no_parent(self):
|
||
|
soup = self.soup("<a><b>Foo</b></a><c>Bar</c>")
|
||
|
a = soup.a
|
||
|
a.extract()
|
||
|
assert None == a.parent
|
||
|
with pytest.raises(ValueError):
|
||
|
a.unwrap()
|
||
|
with pytest.raises(ValueError):
|
||
|
a.replace_with(soup.c)
|
||
|
|
||
|
def test_replace_tag_with_itself(self):
|
||
|
text = "<a><b></b><c>Foo<d></d></c></a><a><e></e></a>"
|
||
|
soup = self.soup(text)
|
||
|
c = soup.c
|
||
|
soup.c.replace_with(c)
|
||
|
assert soup.decode() == self.document_for(text)
|
||
|
|
||
|
def test_replace_tag_with_its_parent_raises_exception(self):
|
||
|
text = "<a><b></b></a>"
|
||
|
soup = self.soup(text)
|
||
|
with pytest.raises(ValueError):
|
||
|
soup.b.replace_with(soup.a)
|
||
|
|
||
|
def test_insert_tag_into_itself_raises_exception(self):
|
||
|
text = "<a><b></b></a>"
|
||
|
soup = self.soup(text)
|
||
|
with pytest.raises(ValueError):
|
||
|
soup.a.insert(0, soup.a)
|
||
|
|
||
|
def test_insert_beautifulsoup_object_inserts_children(self):
|
||
|
"""Inserting one BeautifulSoup object into another actually inserts all
|
||
|
of its children -- you'll never combine BeautifulSoup objects.
|
||
|
"""
|
||
|
soup = self.soup("<p>And now, a word:</p><p>And we're back.</p>")
|
||
|
|
||
|
text = "<p>p2</p><p>p3</p>"
|
||
|
to_insert = self.soup(text)
|
||
|
soup.insert(1, to_insert)
|
||
|
|
||
|
for i in soup.descendants:
|
||
|
assert not isinstance(i, BeautifulSoup)
|
||
|
|
||
|
p1, p2, p3, p4 = list(soup.children)
|
||
|
assert "And now, a word:" == p1.string
|
||
|
assert "p2" == p2.string
|
||
|
assert "p3" == p3.string
|
||
|
assert "And we're back." == p4.string
|
||
|
|
||
|
|
||
|
def test_replace_with_maintains_next_element_throughout(self):
|
||
|
soup = self.soup('<p><a>one</a><b>three</b></p>')
|
||
|
a = soup.a
|
||
|
b = a.contents[0]
|
||
|
# Make it so the <a> tag has two text children.
|
||
|
a.insert(1, "two")
|
||
|
|
||
|
# Now replace each one with the empty string.
|
||
|
left, right = a.contents
|
||
|
left.replaceWith('')
|
||
|
right.replaceWith('')
|
||
|
|
||
|
# The <b> tag is still connected to the tree.
|
||
|
assert "three" == soup.b.string
|
||
|
|
||
|
def test_replace_final_node(self):
|
||
|
soup = self.soup("<b>Argh!</b>")
|
||
|
soup.find(string="Argh!").replace_with("Hooray!")
|
||
|
new_text = soup.find(string="Hooray!")
|
||
|
b = soup.b
|
||
|
assert new_text.previous_element == b
|
||
|
assert new_text.parent == b
|
||
|
assert new_text.previous_element.next_element == new_text
|
||
|
assert new_text.next_element == None
|
||
|
|
||
|
def test_consecutive_text_nodes(self):
|
||
|
# A builder should never create two consecutive text nodes,
|
||
|
# but if you insert one next to another, Beautiful Soup will
|
||
|
# handle it correctly.
|
||
|
soup = self.soup("<a><b>Argh!</b><c></c></a>")
|
||
|
soup.b.insert(1, "Hooray!")
|
||
|
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<a><b>Argh!Hooray!</b><c></c></a>"
|
||
|
)
|
||
|
|
||
|
new_text = soup.find(string="Hooray!")
|
||
|
assert new_text.previous_element == "Argh!"
|
||
|
assert new_text.previous_element.next_element == new_text
|
||
|
|
||
|
assert new_text.previous_sibling == "Argh!"
|
||
|
assert new_text.previous_sibling.next_sibling == new_text
|
||
|
|
||
|
assert new_text.next_sibling == None
|
||
|
assert new_text.next_element == soup.c
|
||
|
|
||
|
def test_insert_string(self):
|
||
|
soup = self.soup("<a></a>")
|
||
|
soup.a.insert(0, "bar")
|
||
|
soup.a.insert(0, "foo")
|
||
|
# The string were added to the tag.
|
||
|
assert ["foo", "bar"] == soup.a.contents
|
||
|
# And they were converted to NavigableStrings.
|
||
|
assert soup.a.contents[0].next_element == "bar"
|
||
|
|
||
|
def test_insert_tag(self):
|
||
|
builder = self.default_builder()
|
||
|
soup = self.soup(
|
||
|
"<a><b>Find</b><c>lady!</c><d></d></a>", builder=builder)
|
||
|
magic_tag = Tag(soup, builder, 'magictag')
|
||
|
magic_tag.insert(0, "the")
|
||
|
soup.a.insert(1, magic_tag)
|
||
|
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<a><b>Find</b><magictag>the</magictag><c>lady!</c><d></d></a>"
|
||
|
)
|
||
|
|
||
|
# Make sure all the relationships are hooked up correctly.
|
||
|
b_tag = soup.b
|
||
|
assert b_tag.next_sibling == magic_tag
|
||
|
assert magic_tag.previous_sibling == b_tag
|
||
|
|
||
|
find = b_tag.find(string="Find")
|
||
|
assert find.next_element == magic_tag
|
||
|
assert magic_tag.previous_element == find
|
||
|
|
||
|
c_tag = soup.c
|
||
|
assert magic_tag.next_sibling == c_tag
|
||
|
assert c_tag.previous_sibling == magic_tag
|
||
|
|
||
|
the = magic_tag.find(string="the")
|
||
|
assert the.parent == magic_tag
|
||
|
assert the.next_element == c_tag
|
||
|
assert c_tag.previous_element == the
|
||
|
|
||
|
def test_append_child_thats_already_at_the_end(self):
|
||
|
data = "<a><b></b></a>"
|
||
|
soup = self.soup(data)
|
||
|
soup.a.append(soup.b)
|
||
|
assert data == soup.decode()
|
||
|
|
||
|
def test_extend(self):
|
||
|
data = "<a><b><c><d><e><f><g></g></f></e></d></c></b></a>"
|
||
|
soup = self.soup(data)
|
||
|
l = [soup.g, soup.f, soup.e, soup.d, soup.c, soup.b]
|
||
|
soup.a.extend(l)
|
||
|
assert "<a><g></g><f></f><e></e><d></d><c></c><b></b></a>" == soup.decode()
|
||
|
|
||
|
def test_extend_with_another_tags_contents(self):
|
||
|
data = '<body><div id="d1"><a>1</a><a>2</a><a>3</a><a>4</a></div><div id="d2"></div></body>'
|
||
|
soup = self.soup(data)
|
||
|
d1 = soup.find('div', id='d1')
|
||
|
d2 = soup.find('div', id='d2')
|
||
|
d2.extend(d1)
|
||
|
assert '<div id="d1"></div>' == d1.decode()
|
||
|
assert '<div id="d2"><a>1</a><a>2</a><a>3</a><a>4</a></div>' == d2.decode()
|
||
|
|
||
|
def test_move_tag_to_beginning_of_parent(self):
|
||
|
data = "<a><b></b><c></c><d></d></a>"
|
||
|
soup = self.soup(data)
|
||
|
soup.a.insert(0, soup.d)
|
||
|
assert "<a><d></d><b></b><c></c></a>" == soup.decode()
|
||
|
|
||
|
def test_insert_works_on_empty_element_tag(self):
|
||
|
# This is a little strange, since most HTML parsers don't allow
|
||
|
# markup like this to come through. But in general, we don't
|
||
|
# know what the parser would or wouldn't have allowed, so
|
||
|
# I'm letting this succeed for now.
|
||
|
soup = self.soup("<br/>")
|
||
|
soup.br.insert(1, "Contents")
|
||
|
assert str(soup.br) == "<br>Contents</br>"
|
||
|
|
||
|
def test_insert_before(self):
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
soup.b.insert_before("BAZ")
|
||
|
soup.a.insert_before("QUUX")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"QUUX<a>foo</a>BAZ<b>bar</b>"
|
||
|
)
|
||
|
|
||
|
soup.a.insert_before(soup.b)
|
||
|
assert soup.decode() == self.document_for("QUUX<b>bar</b><a>foo</a>BAZ")
|
||
|
|
||
|
# Can't insert an element before itself.
|
||
|
b = soup.b
|
||
|
with pytest.raises(ValueError):
|
||
|
b.insert_before(b)
|
||
|
|
||
|
# Can't insert before if an element has no parent.
|
||
|
b.extract()
|
||
|
with pytest.raises(ValueError):
|
||
|
b.insert_before("nope")
|
||
|
|
||
|
# Can insert an identical element
|
||
|
soup = self.soup("<a>")
|
||
|
soup.a.insert_before(soup.new_tag("a"))
|
||
|
|
||
|
# TODO: OK but what happens?
|
||
|
|
||
|
def test_insert_multiple_before(self):
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
soup.b.insert_before("BAZ", " ", "QUUX")
|
||
|
soup.a.insert_before("QUUX", " ", "BAZ")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"QUUX BAZ<a>foo</a>BAZ QUUX<b>bar</b>"
|
||
|
)
|
||
|
|
||
|
soup.a.insert_before(soup.b, "FOO")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"QUUX BAZ<b>bar</b>FOO<a>foo</a>BAZ QUUX"
|
||
|
)
|
||
|
|
||
|
def test_insert_after(self):
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
soup.b.insert_after("BAZ")
|
||
|
soup.a.insert_after("QUUX")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<a>foo</a>QUUX<b>bar</b>BAZ"
|
||
|
)
|
||
|
soup.b.insert_after(soup.a)
|
||
|
assert soup.decode() == self.document_for("QUUX<b>bar</b><a>foo</a>BAZ")
|
||
|
|
||
|
# Can't insert an element after itself.
|
||
|
b = soup.b
|
||
|
with pytest.raises(ValueError):
|
||
|
b.insert_after(b)
|
||
|
|
||
|
# Can't insert after if an element has no parent.
|
||
|
b.extract()
|
||
|
with pytest.raises(ValueError):
|
||
|
b.insert_after("nope")
|
||
|
|
||
|
# Can insert an identical element
|
||
|
soup = self.soup("<a>")
|
||
|
soup.a.insert_before(soup.new_tag("a"))
|
||
|
|
||
|
# TODO: OK but what does it look like?
|
||
|
|
||
|
def test_insert_multiple_after(self):
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
soup.b.insert_after("BAZ", " ", "QUUX")
|
||
|
soup.a.insert_after("QUUX", " ", "BAZ")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<a>foo</a>QUUX BAZ<b>bar</b>BAZ QUUX"
|
||
|
)
|
||
|
soup.b.insert_after(soup.a, "FOO ")
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"QUUX BAZ<b>bar</b><a>foo</a>FOO BAZ QUUX"
|
||
|
)
|
||
|
|
||
|
def test_insert_after_raises_exception_if_after_has_no_meaning(self):
|
||
|
soup = self.soup("")
|
||
|
tag = soup.new_tag("a")
|
||
|
string = soup.new_string("")
|
||
|
with pytest.raises(ValueError):
|
||
|
string.insert_after(tag)
|
||
|
with pytest.raises(NotImplementedError):
|
||
|
soup.insert_after(tag)
|
||
|
with pytest.raises(ValueError):
|
||
|
tag.insert_after(tag)
|
||
|
|
||
|
def test_insert_before_raises_notimplementederror_if_before_has_no_meaning(self):
|
||
|
soup = self.soup("")
|
||
|
tag = soup.new_tag("a")
|
||
|
string = soup.new_string("")
|
||
|
with pytest.raises(ValueError):
|
||
|
string.insert_before(tag)
|
||
|
with pytest.raises(NotImplementedError):
|
||
|
soup.insert_before(tag)
|
||
|
with pytest.raises(ValueError):
|
||
|
tag.insert_before(tag)
|
||
|
|
||
|
def test_replace_with(self):
|
||
|
soup = self.soup(
|
||
|
"<p>There's <b>no</b> business like <b>show</b> business</p>")
|
||
|
no, show = soup.find_all('b')
|
||
|
show.replace_with(no)
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<p>There's business like <b>no</b> business</p>"
|
||
|
)
|
||
|
|
||
|
assert show.parent == None
|
||
|
assert no.parent == soup.p
|
||
|
assert no.next_element == "no"
|
||
|
assert no.next_sibling == " business"
|
||
|
|
||
|
def test_replace_with_errors(self):
|
||
|
# Can't replace a tag that's not part of a tree.
|
||
|
a_tag = Tag(name="a")
|
||
|
with pytest.raises(ValueError):
|
||
|
a_tag.replace_with("won't work")
|
||
|
|
||
|
# Can't replace a tag with its parent.
|
||
|
a_tag = self.soup("<a><b></b></a>").a
|
||
|
with pytest.raises(ValueError):
|
||
|
a_tag.b.replace_with(a_tag)
|
||
|
|
||
|
# Or with a list that includes its parent.
|
||
|
with pytest.raises(ValueError):
|
||
|
a_tag.b.replace_with("string1", a_tag, "string2")
|
||
|
|
||
|
def test_replace_with_multiple(self):
|
||
|
data = "<a><b></b><c></c></a>"
|
||
|
soup = self.soup(data)
|
||
|
d_tag = soup.new_tag("d")
|
||
|
d_tag.string = "Text In D Tag"
|
||
|
e_tag = soup.new_tag("e")
|
||
|
f_tag = soup.new_tag("f")
|
||
|
a_string = "Random Text"
|
||
|
soup.c.replace_with(d_tag, e_tag, a_string, f_tag)
|
||
|
assert soup.decode() == "<a><b></b><d>Text In D Tag</d><e></e>Random Text<f></f></a>"
|
||
|
assert soup.b.next_element == d_tag
|
||
|
assert d_tag.string.next_element==e_tag
|
||
|
assert e_tag.next_element.string == a_string
|
||
|
assert e_tag.next_element.next_element == f_tag
|
||
|
|
||
|
def test_replace_first_child(self):
|
||
|
data = "<a><b></b><c></c></a>"
|
||
|
soup = self.soup(data)
|
||
|
soup.b.replace_with(soup.c)
|
||
|
assert "<a><c></c></a>" == soup.decode()
|
||
|
|
||
|
def test_replace_last_child(self):
|
||
|
data = "<a><b></b><c></c></a>"
|
||
|
soup = self.soup(data)
|
||
|
soup.c.replace_with(soup.b)
|
||
|
assert "<a><b></b></a>" == soup.decode()
|
||
|
|
||
|
def test_nested_tag_replace_with(self):
|
||
|
soup = self.soup(
|
||
|
"""<a>We<b>reserve<c>the</c><d>right</d></b></a><e>to<f>refuse</f><g>service</g></e>""")
|
||
|
|
||
|
# Replace the entire <b> tag and its contents ("reserve the
|
||
|
# right") with the <f> tag ("refuse").
|
||
|
remove_tag = soup.b
|
||
|
move_tag = soup.f
|
||
|
remove_tag.replace_with(move_tag)
|
||
|
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<a>We<f>refuse</f></a><e>to<g>service</g></e>"
|
||
|
)
|
||
|
|
||
|
# The <b> tag is now an orphan.
|
||
|
assert remove_tag.parent == None
|
||
|
assert remove_tag.find(string="right").next_element == None
|
||
|
assert remove_tag.previous_element == None
|
||
|
assert remove_tag.next_sibling == None
|
||
|
assert remove_tag.previous_sibling == None
|
||
|
|
||
|
# The <f> tag is now connected to the <a> tag.
|
||
|
assert move_tag.parent == soup.a
|
||
|
assert move_tag.previous_element == "We"
|
||
|
assert move_tag.next_element.next_element == soup.e
|
||
|
assert move_tag.next_sibling == None
|
||
|
|
||
|
# The gap where the <f> tag used to be has been mended, and
|
||
|
# the word "to" is now connected to the <g> tag.
|
||
|
to_text = soup.find(string="to")
|
||
|
g_tag = soup.g
|
||
|
assert to_text.next_element == g_tag
|
||
|
assert to_text.next_sibling == g_tag
|
||
|
assert g_tag.previous_element == to_text
|
||
|
assert g_tag.previous_sibling == to_text
|
||
|
|
||
|
def test_unwrap(self):
|
||
|
tree = self.soup("""
|
||
|
<p>Unneeded <em>formatting</em> is unneeded</p>
|
||
|
""")
|
||
|
tree.em.unwrap()
|
||
|
assert tree.em == None
|
||
|
assert tree.p.text == "Unneeded formatting is unneeded"
|
||
|
|
||
|
def test_wrap(self):
|
||
|
soup = self.soup("I wish I was bold.")
|
||
|
value = soup.string.wrap(soup.new_tag("b"))
|
||
|
assert value.decode() == "<b>I wish I was bold.</b>"
|
||
|
assert soup.decode() == self.document_for("<b>I wish I was bold.</b>")
|
||
|
|
||
|
def test_wrap_extracts_tag_from_elsewhere(self):
|
||
|
soup = self.soup("<b></b>I wish I was bold.")
|
||
|
soup.b.next_sibling.wrap(soup.b)
|
||
|
assert soup.decode() == self.document_for("<b>I wish I was bold.</b>")
|
||
|
|
||
|
def test_wrap_puts_new_contents_at_the_end(self):
|
||
|
soup = self.soup("<b>I like being bold.</b>I wish I was bold.")
|
||
|
soup.b.next_sibling.wrap(soup.b)
|
||
|
assert 2 == len(soup.b.contents)
|
||
|
assert soup.decode() == self.document_for(
|
||
|
"<b>I like being bold.I wish I was bold.</b>"
|
||
|
)
|
||
|
|
||
|
def test_extract(self):
|
||
|
soup = self.soup(
|
||
|
'<html><body>Some content. <div id="nav">Nav crap</div> More content.</body></html>')
|
||
|
|
||
|
assert len(soup.body.contents) == 3
|
||
|
extracted = soup.find(id="nav").extract()
|
||
|
|
||
|
assert soup.decode() == "<html><body>Some content. More content.</body></html>"
|
||
|
assert extracted.decode() == '<div id="nav">Nav crap</div>'
|
||
|
|
||
|
# The extracted tag is now an orphan.
|
||
|
assert len(soup.body.contents) == 2
|
||
|
assert extracted.parent == None
|
||
|
assert extracted.previous_element == None
|
||
|
assert extracted.next_element.next_element == None
|
||
|
|
||
|
# The gap where the extracted tag used to be has been mended.
|
||
|
content_1 = soup.find(string="Some content. ")
|
||
|
content_2 = soup.find(string=" More content.")
|
||
|
assert content_1.next_element == content_2
|
||
|
assert content_1.next_sibling == content_2
|
||
|
assert content_2.previous_element == content_1
|
||
|
assert content_2.previous_sibling == content_1
|
||
|
|
||
|
def test_extract_distinguishes_between_identical_strings(self):
|
||
|
soup = self.soup("<a>foo</a><b>bar</b>")
|
||
|
foo_1 = soup.a.string
|
||
|
bar_1 = soup.b.string
|
||
|
foo_2 = soup.new_string("foo")
|
||
|
bar_2 = soup.new_string("bar")
|
||
|
soup.a.append(foo_2)
|
||
|
soup.b.append(bar_2)
|
||
|
|
||
|
# Now there are two identical strings in the <a> tag, and two
|
||
|
# in the <b> tag. Let's remove the first "foo" and the second
|
||
|
# "bar".
|
||
|
foo_1.extract()
|
||
|
bar_2.extract()
|
||
|
assert foo_2 == soup.a.string
|
||
|
assert bar_2 == soup.b.string
|
||
|
|
||
|
def test_extract_multiples_of_same_tag(self):
|
||
|
soup = self.soup("""
|
||
|
<html>
|
||
|
<head>
|
||
|
<script>foo</script>
|
||
|
</head>
|
||
|
<body>
|
||
|
<script>bar</script>
|
||
|
<a></a>
|
||
|
</body>
|
||
|
<script>baz</script>
|
||
|
</html>""")
|
||
|
[soup.script.extract() for i in soup.find_all("script")]
|
||
|
assert "<body>\n\n<a></a>\n</body>" == str(soup.body)
|
||
|
|
||
|
|
||
|
def test_extract_works_when_element_is_surrounded_by_identical_strings(self):
|
||
|
soup = self.soup(
|
||
|
'<html>\n'
|
||
|
'<body>hi</body>\n'
|
||
|
'</html>')
|
||
|
soup.find('body').extract()
|
||
|
assert None == soup.find('body')
|
||
|
|
||
|
|
||
|
def test_clear(self):
|
||
|
"""Tag.clear()"""
|
||
|
soup = self.soup("<p><a>String <em>Italicized</em></a> and another</p>")
|
||
|
# clear using extract()
|
||
|
a = soup.a
|
||
|
soup.p.clear()
|
||
|
assert len(soup.p.contents) == 0
|
||
|
assert hasattr(a, "contents")
|
||
|
|
||
|
# clear using decompose()
|
||
|
em = a.em
|
||
|
a.clear(decompose=True)
|
||
|
assert 0 == len(em.contents)
|
||
|
|
||
|
|
||
|
def test_decompose(self):
|
||
|
# Test PageElement.decompose() and PageElement.decomposed
|
||
|
soup = self.soup("<p><a>String <em>Italicized</em></a></p><p>Another para</p>")
|
||
|
p1, p2 = soup.find_all('p')
|
||
|
a = p1.a
|
||
|
text = p1.em.string
|
||
|
for i in [p1, p2, a, text]:
|
||
|
assert False == i.decomposed
|
||
|
|
||
|
# This sets p1 and everything beneath it to decomposed.
|
||
|
p1.decompose()
|
||
|
for i in [p1, a, text]:
|
||
|
assert True == i.decomposed
|
||
|
# p2 is unaffected.
|
||
|
assert False == p2.decomposed
|
||
|
|
||
|
def test_string_set(self):
|
||
|
"""Tag.string = 'string'"""
|
||
|
soup = self.soup("<a></a> <b><c></c></b>")
|
||
|
soup.a.string = "foo"
|
||
|
assert soup.a.contents == ["foo"]
|
||
|
soup.b.string = "bar"
|
||
|
assert soup.b.contents == ["bar"]
|
||
|
|
||
|
def test_string_set_does_not_affect_original_string(self):
|
||
|
soup = self.soup("<a><b>foo</b><c>bar</c>")
|
||
|
soup.b.string = soup.c.string
|
||
|
assert soup.a.encode() == b"<a><b>bar</b><c>bar</c></a>"
|
||
|
|
||
|
def test_set_string_preserves_class_of_string(self):
|
||
|
soup = self.soup("<a></a>")
|
||
|
cdata = CData("foo")
|
||
|
soup.a.string = cdata
|
||
|
assert isinstance(soup.a.string, CData)
|
||
|
|
||
|
|
||
|
class TestDeprecatedArguments(SoupTest):
|
||
|
|
||
|
def test_find_type_method_string(self):
|
||
|
soup = self.soup("<a>some</a><b>markup</b>")
|
||
|
with warnings.catch_warnings(record=True) as w:
|
||
|
[result] = soup.find_all(text='markup')
|
||
|
assert result == 'markup'
|
||
|
assert result.parent.name == 'b'
|
||
|
msg = str(w[0].message)
|
||
|
assert msg == "The 'text' argument to find()-type methods is deprecated. Use 'string' instead."
|
||
|
|
||
|
def test_soupstrainer_constructor_string(self):
|
||
|
with warnings.catch_warnings(record=True) as w:
|
||
|
strainer = SoupStrainer(text="text")
|
||
|
assert strainer.text == 'text'
|
||
|
msg = str(w[0].message)
|
||
|
assert msg == "The 'text' argument to the SoupStrainer constructor is deprecated. Use 'string' instead."
|
||
|
|