/external/chromium-trace/catapult/third_party/beautifulsoup4/bs4/tests/ |
D | test_tree.py | 66 soup = self.soup("<a>1</a><b>2</b><a>3</a><b>4</b>") 67 self.assertEqual(soup.find("b").string, "2") 70 soup = self.soup(u'<h1>Räksmörgås</h1>') 71 self.assertEqual(soup.find(text=u'Räksmörgås'), u'Räksmörgås') 75 soup = self.soup("<a>foo</a><b>bar</b>") 76 self.assertEqual(2, len(soup.find_all())) 80 soup = self.soup("<a>foo</a><b>bar</b><a>baz</a>") 81 self.assertEqual(2, len(soup.find_all('a'))) 88 soup = self.soup("<html>Foo<b>bar</b>\xbb</html>") 90 self.assertEqual(soup.find_all(text="bar"), [u"bar"]) [all …]
|
D | test_html5lib.py | 32 soup = self.soup(markup, parse_only=strainer) 34 soup.decode(), self.document_for(markup)) 70 soup = self.soup(markup) 72 self.assertEqual(b"<p>foo</p>", soup.p.encode()) 76 soup = self.soup(markup) 77 …Equal(u"<body><p><em>foo</em></p><em>\n</em><p><em>bar<a></a></em></p></body>", soup.body.decode()) 78 self.assertEqual(2, len(soup.find_all('p'))) 83 soup = self.soup(markup) 84 …ual(u"<body><p><em>foo</em></p><em>\n</em><p><em>bar<a></a></em></p>\n</body>", soup.body.decode()) 85 self.assertEqual(2, len(soup.find_all('p')))
|
D | test_soup.py | 43 soup = self.soup(data) 44 self.assertEqual(u"éé", soup.h1.string) 48 soup = self.soup(data) 49 self.assertEqual(u"foo\0bar", soup.h1.string) 56 soup = self.soup("<a><b></b></a>", parseOnlyThese=SoupStrainer("b")) 60 self.assertEqual(b"<b></b>", soup.encode()) 65 soup = self.soup(utf8, fromEncoding="utf8") 69 self.assertEqual("utf8", soup.original_encoding) 73 TypeError, self.soup, "<a>", no_such_argument=True) 82 soup = self.soup(filename) [all …]
|
D | test_lxml.py | 56 soup = self.soup("<!DOCTYPE>") 57 doctype = soup.contents[0] 64 soup = BeautifulStoneSoup("<b />") 65 self.assertEqual(u"<b/>", unicode(soup.b)) 76 soup = self.soup(markup) 78 soup.encode("utf-8").replace(b"\n", b''),
|
/external/chromium-trace/catapult/third_party/beautifulsoup4/bs4/ |
D | testing.py | 26 def soup(self, markup, **kwargs): member in SoupTest 59 doctype_str, soup = self._document_with_doctype(doctype_fragment) 62 doctype = soup.contents[0] 65 self.assertEqual(str(soup)[:len(doctype_str)], doctype_str) 69 self.assertEqual(soup.p.contents[0], 'foo') 75 soup = self.soup(markup) 76 return doctype, soup 85 soup = self.soup("<!DOCTYPE>") 86 doctype = soup.contents[0] 112 soup = self.soup(markup) [all …]
|
D | diagnose.py | 58 soup = BeautifulSoup(data, parser) 65 print soup.prettify() 167 soup = BeautifulSoup(data, parser)
|
D | __init__.py | 156 self.builder.soup = self 204 self.builder.soup = None 405 soup = BeautifulSoup(sys.stdin) variable 406 print soup.prettify()
|
/external/chromium-trace/catapult/third_party/beautifulsoup4/bs4/builder/ |
D | _html5lib.py | 34 if self.soup.parse_only is not None: 49 self.soup, namespaceHTMLElements) 59 def __init__(self, soup, namespaceHTMLElements): argument 60 self.soup = soup 64 self.soup.reset() 65 return Element(self.soup, self.soup, None) 73 self.soup.object_was_parsed(doctype) 76 tag = self.soup.new_tag(name, namespace) 77 return Element(tag, self.soup, namespace) 80 return TextNode(Comment(data), self.soup) [all …]
|
D | _htmlparser.py | 56 self.soup.handle_starttag(name, None, None, attr_dict) 59 self.soup.handle_endtag(name) 62 self.soup.handle_data(data) 90 self.soup.endData() 91 self.soup.handle_data(data) 92 self.soup.endData(Comment) 95 self.soup.endData() 101 self.soup.handle_data(data) 102 self.soup.endData(Doctype) 110 self.soup.endData() [all …]
|
D | _lxml.py | 61 self.soup = None 112 self.parser = self.parser_for(self.soup.original_encoding) 164 self.soup.handle_starttag(name, namespace, nsprefix, attrs) 176 self.soup.endData() 177 completed_tag = self.soup.tagStack[-1] 185 self.soup.handle_endtag(name, nsprefix) 195 self.soup.handle_data(content) 198 self.soup.endData() 200 self.soup.object_was_parsed(doctype) 204 self.soup.endData() [all …]
|
D | __init__.py | 96 self.soup = None 187 self.soup.handle_starttag(name, attrs) 191 self.soup.handle_endtag(name) 212 self.soup.handle_data(content)
|
/external/chromium-trace/catapult/third_party/py_vulcanize/py_vulcanize/ |
D | parse_html_deps.py | 38 def __init__(self, soup): argument 39 if not soup: 41 self._soup = soup 79 soup = bs4.BeautifulSoup() 80 soup.reset() 84 soup.append(n) 88 soup.append(n) 89 return soup 129 soup = _CreateSoupWithoutHeadOrBody(unicode(self._soup)) 132 for x in soup.contents: [all …]
|
/external/chromium-trace/catapult/catapult_build/ |
D | html_checks.py | 36 soup = parse_html.BeautifulSoup('\n'.join(affected_file.NewContents())) 38 check(path, soup, results, output_api) 41 def CheckDoctype(path, soup, results, output_api): argument 42 if _HasHtml5Declaration(soup): 48 def _HasHtml5Declaration(soup): argument 49 for item in soup.contents: 55 def CheckImportOrder(path, soup, results, output_api): argument 57 for link in soup.find_all('link'):
|
D | js_checks.py | 203 soup = parse_html.BeautifulSoup(contents) 204 script_elements = soup.find_all('script', src=None)
|
/external/chromium-trace/catapult/third_party/beautifulsoup4/ |
D | README.txt | 4 >>> soup = BeautifulSoup("<p>Some<b>bad<i>HTML") 5 >>> print soup.prettify() 19 >>> soup.find(text="bad") 22 >>> soup.i 25 >>> soup = BeautifulSoup("<tag1>Some<tag2/>bad<tag3>XML", "xml") 26 >>> print soup.prettify() 60 http://readthedocs.org/docs/beautiful-soup-4/
|
D | NEWS.txt | 457 new <p> tag will look like "<p />" if the soup object was created to 458 parse XML, but it will look like "<p></p>" if the soup object was 689 constructed soup object should always be valid as part of an XML 714 arguments would replace it, but they don't. You can't call soup('a', 768 soup(args={"id" : "5"}) with soup(id="5"). You can still use args if 784 string or a filehandle into the soup constructor, not with feed after 785 the soup has been created. There is still a feed method, but it's the 959 soup.bodyTag.pTag.tableTag('th'): 967 This lets you do tree traversals like soup.htmlTag.headTag.titleTag
|
/external/chromium-trace/catapult/third_party/beautifulsoup4/doc/source/ |
D | index.rst | 68 soup = BeautifulSoup(html_doc) 70 print(soup.prettify()) 106 soup.title 109 soup.title.name 112 soup.title.string 115 soup.title.parent.name 118 soup.p 121 soup.p['class'] 124 soup.a 127 soup.find_all('a') [all …]
|
/external/chromium-trace/catapult/third_party/beautifulsoup4/scripts/ |
D | demonstrate_parser_differences.py | 44 soup = BeautifulSoup(self.markup, parser) 47 output = soup.div 49 output = soup
|
/external/autotest/contrib/ |
D | watch_expensive_rpcs.py | 26 soup = BeautifulSoup(page) 28 for row in soup.table.findAll('tr'):
|
/external/autotest/site_utils/rpm_control_system/ |
D | dli.py | 69 soup=BeautifulSoup.BeautifulSoup(url) 71 powertable=soup.findAll('table')[5]
|
D | BeautifulSoup.py | 2016 soup = BeautifulSoup(sys.stdin) variable 2017 print soup.prettify()
|
/external/chromium-trace/catapult/perf_insights/perf_insights_build/ |
D | pi_report_to_html.py | 56 soup = bs4.BeautifulSoup(html_contents) 57 elements = soup.findAll('polymer-element')
|
/external/chromium-trace/catapult/third_party/webtest/webtest/ |
D | response.py | 426 soup = BeautifulSoup(self.testbody, self.parser_features) 427 return soup
|
/external/icu/icu4j/demos/src/com/ibm/icu/dev/demo/translit/resources/ |
D | Transliterator_Kanji_English.txt | 4159 羮>'[soup]'; 4164 羹>'[soup]';
|
/external/libpcap/ |
D | CHANGES | 179 Treat either EPERM or EACCES as "no soup for you".
|