Keep py3.X compatibility for urllib
Use six.moves.urllib.parse instead of urllib Change-Id: I9c563e0e54f52b0b26919fd0bd2d31dc8873d8fd Closes-bug: #1280105
This commit is contained in:
parent
0893834b4d
commit
09981aaf0a
@ -38,7 +38,7 @@ RETENTION = '90d'
|
||||
import json
|
||||
import sys
|
||||
import time
|
||||
import urllib
|
||||
import six.moves.urllib.parse as urlparse
|
||||
import urllib2
|
||||
|
||||
|
||||
@ -50,7 +50,7 @@ def influxdb_get(uri, query, db=None):
|
||||
getparams['db'] = db
|
||||
|
||||
try:
|
||||
params = urllib.urlencode(getparams)
|
||||
params = urlparse.urlencode(getparams)
|
||||
uri = "{}&{}".format(uri,params)
|
||||
req = urllib2.urlopen(uri)
|
||||
json_value = json.loads(req.read())
|
||||
|
@ -15,8 +15,7 @@
|
||||
|
||||
import datetime
|
||||
import json
|
||||
import urllib
|
||||
import urlparse
|
||||
import six.moves.urllib.parse as urlparse
|
||||
|
||||
import falcon
|
||||
from oslo_log import log
|
||||
@ -323,7 +322,7 @@ def paginate(resource, uri, limit):
|
||||
|
||||
next_link = build_base_uri(parsed_uri)
|
||||
|
||||
new_query_params = [u'offset' + '=' + urllib.quote(
|
||||
new_query_params = [u'offset' + '=' + urlparse.quote(
|
||||
new_offset.encode('utf8'), safe='')]
|
||||
|
||||
_get_old_query_params_except_offset(new_query_params, parsed_uri)
|
||||
@ -365,7 +364,7 @@ def paginate_measurement(measurement, uri, limit):
|
||||
|
||||
next_link = build_base_uri(parsed_uri)
|
||||
|
||||
new_query_params = [u'offset' + '=' + urllib.quote(
|
||||
new_query_params = [u'offset' + '=' + urlparse.quote(
|
||||
new_offset.encode('utf8'), safe='')]
|
||||
|
||||
_get_old_query_params_except_offset(new_query_params, parsed_uri)
|
||||
@ -403,10 +402,10 @@ def _get_old_query_params(parsed_uri):
|
||||
for query_param in parsed_uri.query.split('&'):
|
||||
query_param_name, query_param_val = query_param.split('=')
|
||||
|
||||
old_query_params.append(urllib.quote(
|
||||
old_query_params.append(urlparse.quote(
|
||||
query_param_name.encode('utf8'), safe='')
|
||||
+ "="
|
||||
+ urllib.quote(query_param_val.encode('utf8'), safe=''))
|
||||
+ urlparse.quote(query_param_val.encode('utf8'), safe=''))
|
||||
|
||||
return old_query_params
|
||||
|
||||
@ -417,9 +416,9 @@ def _get_old_query_params_except_offset(new_query_params, parsed_uri):
|
||||
for query_param in parsed_uri.query.split('&'):
|
||||
query_param_name, query_param_val = query_param.split('=')
|
||||
if query_param_name.lower() != 'offset':
|
||||
new_query_params.append(urllib.quote(
|
||||
new_query_params.append(urlparse.quote(
|
||||
query_param_name.encode(
|
||||
'utf8'), safe='') + "=" + urllib.quote(
|
||||
'utf8'), safe='') + "=" + urlparse.quote(
|
||||
query_param_val.encode(
|
||||
'utf8'), safe=''))
|
||||
|
||||
@ -444,7 +443,7 @@ def paginate_statistics(statistic, uri, limit):
|
||||
|
||||
next_link = build_base_uri(parsed_uri)
|
||||
|
||||
new_query_params = [u'offset' + '=' + urllib.quote(
|
||||
new_query_params = [u'offset' + '=' + urlparse.quote(
|
||||
new_offset.encode('utf8'), safe='')]
|
||||
|
||||
_get_old_query_params_except_offset(new_query_params, parsed_uri)
|
||||
|
Loading…
Reference in New Issue
Block a user