10BC0 TST: win32 paths cannot be turned into URLs by prefixing them with "file://" by mindw · Pull Request #4580 · pandas-dev/pandas · GitHub
[go: up one dir, main page]

Skip to content
Closed
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

10BC0 Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Next Next commit
win32 paths cannot be turned into URLs by prefixing them with "file://"
  • Loading branch information
mindw committed Aug 16, 2013
commit d64a0476c0d39ce173da97ffd4417484a413bf29
12 changes: 9 additions & 3 deletions pandas/io/tests/test_html.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,12 @@
from pandas.util.testing import makeCustomDataframe as mkdf


import urlparse, urllib
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can you put this in pandas/io/common.py ?


def path2url(path):
return urlparse.urljoin(
'file:', urllib.pathname2url(path))
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this you can take out


def _have_module(module_name):
try:
import_module(module_name)
Expand Down Expand Up @@ -292,7 +298,7 @@ def test_bad_url_protocol(self):
@slow
def test_file_url(self):
url = self.banklist_data
dfs = self.run_read_html('file://' + url, 'First',
dfs = self.run_read_html(path2url(url), 'First',
attrs={'id': 'table'})
self.assertIsInstance(dfs, list)
for df in dfs:
Expand Down Expand Up @@ -338,7 +344,7 @@ def test_multiindex_header_index_skiprows(self):
@slow
def test_regex_idempotency(self):
url = self.banklist_data
dfs = self.run_read_html('file://' + url,
dfs = self.run_read_html(path2url(url),
match=re.compile(re.compile('Florida')),
attrs={'id': 'table'})
self.assertIsInstance(dfs, list)
Expand Down Expand Up @@ -464,7 +470,7 @@ def test_invalid_flavor():

def get_elements_from_url(url, element='table', base_url="file://"):
_skip_if_none_of(('bs4', 'html5lib'))
url = "".join([base_url, url])
url = path2url(url) if base_url == "file://" else "".join([base_url, url])
from bs4 import BeautifulSoup
with urlopen(url) as f:
soup = BeautifulSoup(f, features='html5lib')
Expand Down
0