2013-07-30 01:19:04 +08:00
|
|
|
from __future__ import unicode_literals
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
import copy
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import sys
|
|
|
|
from io import BytesIO
|
2014-05-26 04:52:47 +08:00
|
|
|
from itertools import chain
|
2012-10-22 02:12:59 +08:00
|
|
|
from pprint import pformat
|
|
|
|
|
|
|
|
from django.conf import settings
|
|
|
|
from django.core import signing
|
2013-05-16 07:14:28 +08:00
|
|
|
from django.core.exceptions import DisallowedHost, ImproperlyConfigured
|
2012-10-22 02:12:59 +08:00
|
|
|
from django.core.files import uploadhandler
|
2013-09-30 23:55:14 +08:00
|
|
|
from django.http.multipartparser import MultiPartParser, MultiPartParserError
|
2012-10-22 02:12:59 +08:00
|
|
|
from django.utils import six
|
2015-01-28 20:35:27 +08:00
|
|
|
from django.utils.datastructures import ImmutableList, MultiValueDict
|
2014-10-31 23:43:34 +08:00
|
|
|
from django.utils.encoding import (
|
2015-01-28 20:35:27 +08:00
|
|
|
escape_uri_path, force_bytes, force_str, force_text, iri_to_uri,
|
|
|
|
)
|
|
|
|
from django.utils.six.moves.urllib.parse import (
|
|
|
|
parse_qsl, quote, urlencode, urljoin, urlsplit,
|
2014-10-31 23:43:34 +08:00
|
|
|
)
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
RAISE_ERROR = object()
|
2012-11-28 05:19:37 +08:00
|
|
|
host_validation_re = re.compile(r"^([a-z0-9.-]+|\[[a-f0-9]*:[a-f0-9:]+\])(:\d+)?$")
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
|
|
|
|
class UnreadablePostError(IOError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2013-10-08 20:05:39 +08:00
|
|
|
class RawPostDataException(Exception):
|
|
|
|
"""
|
|
|
|
You cannot access raw_post_data from a request that has
|
|
|
|
multipart/* POST data if it has been accessed via POST,
|
|
|
|
FILES, etc..
|
|
|
|
"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2012-10-22 02:12:59 +08:00
|
|
|
class HttpRequest(object):
|
|
|
|
"""A basic HTTP request."""
|
|
|
|
|
|
|
|
# The encoding used in GET/POST dicts. None means use default setting.
|
|
|
|
_encoding = None
|
|
|
|
_upload_handlers = []
|
|
|
|
|
|
|
|
def __init__(self):
|
2013-06-26 15:36:25 +08:00
|
|
|
# WARNING: The `WSGIRequest` subclass doesn't call `super`.
|
|
|
|
# Any variable assignment made here should also happen in
|
|
|
|
# `WSGIRequest.__init__()`.
|
|
|
|
|
Fixed #22799 -- Made GET and POST on HttpRequest QueryDicts, and FILES a MultiValueDict.
Previously, GET, POST, and FILES on an HttpRequest were created in
the __init__ method as dictionaries. This was not something you would
usually notice causing trouble in production as you'd only see a
WSGIRequest, but in testing using the test client, calling .getlist
on GET, POST, or FILES for a request with no get/post data resulted in
an AttributeError.
Changed GET and POST on an HttpRequest object to be mutable
QueryDicts (mutable because the Django tests, and probably many
third party tests, were expecting it).
2014-06-08 07:47:43 +08:00
|
|
|
self.GET = QueryDict(mutable=True)
|
|
|
|
self.POST = QueryDict(mutable=True)
|
|
|
|
self.COOKIES = {}
|
|
|
|
self.META = {}
|
|
|
|
self.FILES = MultiValueDict()
|
|
|
|
|
2012-10-22 02:12:59 +08:00
|
|
|
self.path = ''
|
|
|
|
self.path_info = ''
|
|
|
|
self.method = None
|
2013-03-11 06:24:34 +08:00
|
|
|
self.resolver_match = None
|
2012-10-22 02:12:59 +08:00
|
|
|
self._post_parse_error = False
|
|
|
|
|
|
|
|
def __repr__(self):
|
2014-11-20 17:28:16 +08:00
|
|
|
if self.method is None or not self.get_full_path():
|
|
|
|
return force_str('<%s>' % self.__class__.__name__)
|
|
|
|
return force_str(
|
|
|
|
'<%s: %s %r>' % (self.__class__.__name__, self.method, force_str(self.get_full_path()))
|
|
|
|
)
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def get_host(self):
|
|
|
|
"""Returns the HTTP host using the environment or request headers."""
|
|
|
|
# We try three options, in order of decreasing preference.
|
|
|
|
if settings.USE_X_FORWARDED_HOST and (
|
2013-11-26 17:43:46 +08:00
|
|
|
'HTTP_X_FORWARDED_HOST' in self.META):
|
2012-10-22 02:12:59 +08:00
|
|
|
host = self.META['HTTP_X_FORWARDED_HOST']
|
|
|
|
elif 'HTTP_HOST' in self.META:
|
|
|
|
host = self.META['HTTP_HOST']
|
|
|
|
else:
|
|
|
|
# Reconstruct the host using the algorithm from PEP 333.
|
|
|
|
host = self.META['SERVER_NAME']
|
|
|
|
server_port = str(self.META['SERVER_PORT'])
|
|
|
|
if server_port != ('443' if self.is_secure() else '80'):
|
|
|
|
host = '%s:%s' % (host, server_port)
|
|
|
|
|
2013-03-28 00:37:08 +08:00
|
|
|
# There is no hostname validation when DEBUG=True
|
|
|
|
if settings.DEBUG:
|
|
|
|
return host
|
|
|
|
|
2013-04-04 04:27:20 +08:00
|
|
|
domain, port = split_domain_port(host)
|
2013-03-28 00:37:08 +08:00
|
|
|
if domain and validate_host(domain, settings.ALLOWED_HOSTS):
|
2013-02-10 01:17:01 +08:00
|
|
|
return host
|
|
|
|
else:
|
2013-04-04 04:27:20 +08:00
|
|
|
msg = "Invalid HTTP_HOST header: %r." % host
|
|
|
|
if domain:
|
2014-07-06 02:19:36 +08:00
|
|
|
msg += " You may need to add %r to ALLOWED_HOSTS." % domain
|
2013-03-28 00:37:08 +08:00
|
|
|
else:
|
2014-07-06 02:19:36 +08:00
|
|
|
msg += " The domain name provided is not valid according to RFC 1034/1035."
|
2013-05-16 07:14:28 +08:00
|
|
|
raise DisallowedHost(msg)
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def get_full_path(self):
|
|
|
|
# RFC 3986 requires query string arguments to be in the ASCII range.
|
|
|
|
# Rather than crash if this doesn't happen, we encode defensively.
|
2014-09-04 20:15:09 +08:00
|
|
|
return '%s%s' % (
|
2014-10-31 23:43:34 +08:00
|
|
|
escape_uri_path(self.path),
|
2014-09-04 20:15:09 +08:00
|
|
|
('?' + iri_to_uri(self.META.get('QUERY_STRING', ''))) if self.META.get('QUERY_STRING', '') else ''
|
|
|
|
)
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def get_signed_cookie(self, key, default=RAISE_ERROR, salt='', max_age=None):
|
|
|
|
"""
|
|
|
|
Attempts to return a signed cookie. If the signature fails or the
|
|
|
|
cookie has expired, raises an exception... unless you provide the
|
|
|
|
default argument in which case that value will be returned instead.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
cookie_value = self.COOKIES[key]
|
|
|
|
except KeyError:
|
|
|
|
if default is not RAISE_ERROR:
|
|
|
|
return default
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
try:
|
|
|
|
value = signing.get_cookie_signer(salt=key + salt).unsign(
|
|
|
|
cookie_value, max_age=max_age)
|
|
|
|
except signing.BadSignature:
|
|
|
|
if default is not RAISE_ERROR:
|
|
|
|
return default
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
return value
|
|
|
|
|
|
|
|
def build_absolute_uri(self, location=None):
|
|
|
|
"""
|
|
|
|
Builds an absolute URI from the location and the variables available in
|
2013-11-04 07:34:11 +08:00
|
|
|
this request. If no ``location`` is specified, the absolute URI is
|
|
|
|
built on ``request.get_full_path()``. Anyway, if the location is
|
|
|
|
absolute, it is simply converted to an RFC 3987 compliant URI and
|
|
|
|
returned and if location is relative or is scheme-relative (i.e.,
|
|
|
|
``//example.com/``), it is urljoined to a base URL constructed from the
|
|
|
|
request variables.
|
2012-10-22 02:12:59 +08:00
|
|
|
"""
|
2013-11-04 07:34:11 +08:00
|
|
|
if location is None:
|
|
|
|
# Make it an absolute url (but schemeless and domainless) for the
|
|
|
|
# edge case that the path starts with '//'.
|
|
|
|
location = '//%s' % self.get_full_path()
|
|
|
|
bits = urlsplit(location)
|
|
|
|
if not (bits.scheme and bits.netloc):
|
|
|
|
current_uri = '{scheme}://{host}{path}'.format(scheme=self.scheme,
|
|
|
|
host=self.get_host(),
|
|
|
|
path=self.path)
|
|
|
|
# Join the constructed URL with the provided location, which will
|
|
|
|
# allow the provided ``location`` to apply query strings to the
|
|
|
|
# base path as well as override the host, if it begins with //
|
2012-10-22 02:12:59 +08:00
|
|
|
location = urljoin(current_uri, location)
|
|
|
|
return iri_to_uri(location)
|
|
|
|
|
2013-10-09 02:30:29 +08:00
|
|
|
def _get_scheme(self):
|
|
|
|
return 'https' if os.environ.get("HTTPS") == "on" else 'http'
|
2012-10-22 02:12:59 +08:00
|
|
|
|
2013-10-09 02:30:29 +08:00
|
|
|
@property
|
|
|
|
def scheme(self):
|
2012-10-22 02:12:59 +08:00
|
|
|
# First, check the SECURE_PROXY_SSL_HEADER setting.
|
|
|
|
if settings.SECURE_PROXY_SSL_HEADER:
|
|
|
|
try:
|
|
|
|
header, value = settings.SECURE_PROXY_SSL_HEADER
|
|
|
|
except ValueError:
|
2014-09-04 20:15:09 +08:00
|
|
|
raise ImproperlyConfigured(
|
|
|
|
'The SECURE_PROXY_SSL_HEADER setting must be a tuple containing two values.'
|
|
|
|
)
|
2012-10-22 02:12:59 +08:00
|
|
|
if self.META.get(header, None) == value:
|
2013-10-09 02:30:29 +08:00
|
|
|
return 'https'
|
|
|
|
# Failing that, fall back to _get_scheme(), which is a hook for
|
2012-10-22 02:12:59 +08:00
|
|
|
# subclasses to implement.
|
2013-10-09 02:30:29 +08:00
|
|
|
return self._get_scheme()
|
|
|
|
|
|
|
|
def is_secure(self):
|
|
|
|
return self.scheme == 'https'
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def is_ajax(self):
|
|
|
|
return self.META.get('HTTP_X_REQUESTED_WITH') == 'XMLHttpRequest'
|
|
|
|
|
|
|
|
@property
|
|
|
|
def encoding(self):
|
|
|
|
return self._encoding
|
|
|
|
|
|
|
|
@encoding.setter
|
|
|
|
def encoding(self, val):
|
|
|
|
"""
|
|
|
|
Sets the encoding used for GET/POST accesses. If the GET or POST
|
|
|
|
dictionary has already been created, it is removed and recreated on the
|
|
|
|
next access (so that it is decoded correctly).
|
|
|
|
"""
|
|
|
|
self._encoding = val
|
|
|
|
if hasattr(self, '_get'):
|
|
|
|
del self._get
|
|
|
|
if hasattr(self, '_post'):
|
|
|
|
del self._post
|
|
|
|
|
|
|
|
def _initialize_handlers(self):
|
|
|
|
self._upload_handlers = [uploadhandler.load_handler(handler, self)
|
|
|
|
for handler in settings.FILE_UPLOAD_HANDLERS]
|
|
|
|
|
|
|
|
@property
|
|
|
|
def upload_handlers(self):
|
|
|
|
if not self._upload_handlers:
|
|
|
|
# If there are no upload handlers defined, initialize them from settings.
|
|
|
|
self._initialize_handlers()
|
|
|
|
return self._upload_handlers
|
|
|
|
|
|
|
|
@upload_handlers.setter
|
|
|
|
def upload_handlers(self, upload_handlers):
|
|
|
|
if hasattr(self, '_files'):
|
|
|
|
raise AttributeError("You cannot set the upload handlers after the upload has been processed.")
|
|
|
|
self._upload_handlers = upload_handlers
|
|
|
|
|
|
|
|
def parse_file_upload(self, META, post_data):
|
|
|
|
"""Returns a tuple of (POST QueryDict, FILES MultiValueDict)."""
|
|
|
|
self.upload_handlers = ImmutableList(
|
|
|
|
self.upload_handlers,
|
|
|
|
warning="You cannot alter upload handlers after the upload has been processed."
|
|
|
|
)
|
|
|
|
parser = MultiPartParser(META, post_data, self.upload_handlers, self.encoding)
|
|
|
|
return parser.parse()
|
|
|
|
|
|
|
|
@property
|
|
|
|
def body(self):
|
|
|
|
if not hasattr(self, '_body'):
|
|
|
|
if self._read_started:
|
2013-10-08 20:05:39 +08:00
|
|
|
raise RawPostDataException("You cannot access body after reading from request's data stream")
|
2012-10-22 02:12:59 +08:00
|
|
|
try:
|
|
|
|
self._body = self.read()
|
|
|
|
except IOError as e:
|
|
|
|
six.reraise(UnreadablePostError, UnreadablePostError(*e.args), sys.exc_info()[2])
|
|
|
|
self._stream = BytesIO(self._body)
|
|
|
|
return self._body
|
|
|
|
|
|
|
|
def _mark_post_parse_error(self):
|
|
|
|
self._post = QueryDict('')
|
|
|
|
self._files = MultiValueDict()
|
|
|
|
self._post_parse_error = True
|
|
|
|
|
|
|
|
def _load_post_and_files(self):
|
|
|
|
"""Populate self._post and self._files if the content-type is a form type"""
|
|
|
|
if self.method != 'POST':
|
|
|
|
self._post, self._files = QueryDict('', encoding=self._encoding), MultiValueDict()
|
|
|
|
return
|
|
|
|
if self._read_started and not hasattr(self, '_body'):
|
|
|
|
self._mark_post_parse_error()
|
|
|
|
return
|
|
|
|
|
|
|
|
if self.META.get('CONTENT_TYPE', '').startswith('multipart/form-data'):
|
|
|
|
if hasattr(self, '_body'):
|
|
|
|
# Use already read data
|
|
|
|
data = BytesIO(self._body)
|
|
|
|
else:
|
|
|
|
data = self
|
|
|
|
try:
|
|
|
|
self._post, self._files = self.parse_file_upload(self.META, data)
|
2013-09-30 23:55:14 +08:00
|
|
|
except MultiPartParserError:
|
2014-04-27 01:18:45 +08:00
|
|
|
# An error occurred while parsing POST data. Since when
|
2012-10-22 02:12:59 +08:00
|
|
|
# formatting the error the request handler might access
|
|
|
|
# self.POST, set self._post and self._file to prevent
|
|
|
|
# attempts to parse POST data again.
|
2014-04-27 01:18:45 +08:00
|
|
|
# Mark that an error occurred. This allows self.__repr__ to
|
2012-10-22 02:12:59 +08:00
|
|
|
# be explicit about it instead of simply representing an
|
|
|
|
# empty POST
|
2013-11-16 07:54:20 +08:00
|
|
|
self._mark_post_parse_error()
|
2012-10-22 02:12:59 +08:00
|
|
|
raise
|
|
|
|
elif self.META.get('CONTENT_TYPE', '').startswith('application/x-www-form-urlencoded'):
|
|
|
|
self._post, self._files = QueryDict(self.body, encoding=self._encoding), MultiValueDict()
|
|
|
|
else:
|
|
|
|
self._post, self._files = QueryDict('', encoding=self._encoding), MultiValueDict()
|
|
|
|
|
2014-05-26 04:52:47 +08:00
|
|
|
def close(self):
|
|
|
|
if hasattr(self, '_files'):
|
|
|
|
for f in chain.from_iterable(l[1] for l in self._files.lists()):
|
|
|
|
f.close()
|
|
|
|
|
2013-11-03 05:02:56 +08:00
|
|
|
# File-like and iterator interface.
|
|
|
|
#
|
|
|
|
# Expects self._stream to be set to an appropriate source of bytes by
|
|
|
|
# a corresponding request subclass (e.g. WSGIRequest).
|
|
|
|
# Also when request data has already been read by request.POST or
|
|
|
|
# request.body, self._stream points to a BytesIO instance
|
|
|
|
# containing that data.
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def read(self, *args, **kwargs):
|
|
|
|
self._read_started = True
|
2013-06-01 16:26:46 +08:00
|
|
|
try:
|
|
|
|
return self._stream.read(*args, **kwargs)
|
|
|
|
except IOError as e:
|
|
|
|
six.reraise(UnreadablePostError, UnreadablePostError(*e.args), sys.exc_info()[2])
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def readline(self, *args, **kwargs):
|
|
|
|
self._read_started = True
|
2013-06-01 16:26:46 +08:00
|
|
|
try:
|
|
|
|
return self._stream.readline(*args, **kwargs)
|
|
|
|
except IOError as e:
|
|
|
|
six.reraise(UnreadablePostError, UnreadablePostError(*e.args), sys.exc_info()[2])
|
2012-10-22 02:12:59 +08:00
|
|
|
|
|
|
|
def xreadlines(self):
|
|
|
|
while True:
|
|
|
|
buf = self.readline()
|
|
|
|
if not buf:
|
|
|
|
break
|
|
|
|
yield buf
|
|
|
|
|
|
|
|
__iter__ = xreadlines
|
|
|
|
|
|
|
|
def readlines(self):
|
|
|
|
return list(iter(self))
|
|
|
|
|
|
|
|
|
|
|
|
class QueryDict(MultiValueDict):
|
|
|
|
"""
|
2014-06-12 04:41:25 +08:00
|
|
|
A specialized MultiValueDict which represents a query string.
|
2012-10-22 02:12:59 +08:00
|
|
|
|
2014-06-12 04:41:25 +08:00
|
|
|
A QueryDict can be used to represent GET or POST data. It subclasses
|
|
|
|
MultiValueDict since keys in such data can be repeated, for instance
|
|
|
|
in the data from a form with a <select multiple> field.
|
|
|
|
|
|
|
|
By default QueryDicts are immutable, though the copy() method
|
|
|
|
will always return a mutable copy.
|
|
|
|
|
|
|
|
Both keys and values set on this class are converted from the given encoding
|
2012-10-22 02:12:59 +08:00
|
|
|
(DEFAULT_CHARSET by default) to unicode.
|
|
|
|
"""
|
2014-06-12 04:41:25 +08:00
|
|
|
|
2012-10-22 02:12:59 +08:00
|
|
|
# These are both reset in __init__, but is specified here at the class
|
|
|
|
# level so that unpickling will have valid values
|
|
|
|
_mutable = True
|
|
|
|
_encoding = None
|
|
|
|
|
2014-06-25 10:01:39 +08:00
|
|
|
def __init__(self, query_string=None, mutable=False, encoding=None):
|
2012-10-22 02:12:59 +08:00
|
|
|
super(QueryDict, self).__init__()
|
|
|
|
if not encoding:
|
|
|
|
encoding = settings.DEFAULT_CHARSET
|
|
|
|
self.encoding = encoding
|
|
|
|
if six.PY3:
|
2012-11-03 19:54:06 +08:00
|
|
|
if isinstance(query_string, bytes):
|
2014-07-13 01:37:59 +08:00
|
|
|
# query_string normally contains URL-encoded data, a subset of ASCII.
|
|
|
|
try:
|
|
|
|
query_string = query_string.decode(encoding)
|
|
|
|
except UnicodeDecodeError:
|
|
|
|
# ... but some user agents are misbehaving :-(
|
|
|
|
query_string = query_string.decode('iso-8859-1')
|
2012-10-22 02:12:59 +08:00
|
|
|
for key, value in parse_qsl(query_string or '',
|
|
|
|
keep_blank_values=True,
|
|
|
|
encoding=encoding):
|
|
|
|
self.appendlist(key, value)
|
|
|
|
else:
|
|
|
|
for key, value in parse_qsl(query_string or '',
|
|
|
|
keep_blank_values=True):
|
2014-07-13 01:37:59 +08:00
|
|
|
try:
|
|
|
|
value = value.decode(encoding)
|
|
|
|
except UnicodeDecodeError:
|
|
|
|
value = value.decode('iso-8859-1')
|
2012-10-22 02:12:59 +08:00
|
|
|
self.appendlist(force_text(key, encoding, errors='replace'),
|
2014-07-13 01:37:59 +08:00
|
|
|
value)
|
2012-10-22 02:12:59 +08:00
|
|
|
self._mutable = mutable
|
|
|
|
|
|
|
|
@property
|
|
|
|
def encoding(self):
|
|
|
|
if self._encoding is None:
|
|
|
|
self._encoding = settings.DEFAULT_CHARSET
|
|
|
|
return self._encoding
|
|
|
|
|
|
|
|
@encoding.setter
|
|
|
|
def encoding(self, value):
|
|
|
|
self._encoding = value
|
|
|
|
|
|
|
|
def _assert_mutable(self):
|
|
|
|
if not self._mutable:
|
|
|
|
raise AttributeError("This QueryDict instance is immutable")
|
|
|
|
|
|
|
|
def __setitem__(self, key, value):
|
|
|
|
self._assert_mutable()
|
|
|
|
key = bytes_to_text(key, self.encoding)
|
|
|
|
value = bytes_to_text(value, self.encoding)
|
|
|
|
super(QueryDict, self).__setitem__(key, value)
|
|
|
|
|
|
|
|
def __delitem__(self, key):
|
|
|
|
self._assert_mutable()
|
|
|
|
super(QueryDict, self).__delitem__(key)
|
|
|
|
|
|
|
|
def __copy__(self):
|
|
|
|
result = self.__class__('', mutable=True, encoding=self.encoding)
|
|
|
|
for key, value in six.iterlists(self):
|
|
|
|
result.setlist(key, value)
|
|
|
|
return result
|
|
|
|
|
|
|
|
def __deepcopy__(self, memo):
|
|
|
|
result = self.__class__('', mutable=True, encoding=self.encoding)
|
|
|
|
memo[id(self)] = result
|
|
|
|
for key, value in six.iterlists(self):
|
|
|
|
result.setlist(copy.deepcopy(key, memo), copy.deepcopy(value, memo))
|
|
|
|
return result
|
|
|
|
|
|
|
|
def setlist(self, key, list_):
|
|
|
|
self._assert_mutable()
|
|
|
|
key = bytes_to_text(key, self.encoding)
|
|
|
|
list_ = [bytes_to_text(elt, self.encoding) for elt in list_]
|
|
|
|
super(QueryDict, self).setlist(key, list_)
|
|
|
|
|
|
|
|
def setlistdefault(self, key, default_list=None):
|
|
|
|
self._assert_mutable()
|
|
|
|
return super(QueryDict, self).setlistdefault(key, default_list)
|
|
|
|
|
|
|
|
def appendlist(self, key, value):
|
|
|
|
self._assert_mutable()
|
|
|
|
key = bytes_to_text(key, self.encoding)
|
|
|
|
value = bytes_to_text(value, self.encoding)
|
|
|
|
super(QueryDict, self).appendlist(key, value)
|
|
|
|
|
|
|
|
def pop(self, key, *args):
|
|
|
|
self._assert_mutable()
|
|
|
|
return super(QueryDict, self).pop(key, *args)
|
|
|
|
|
|
|
|
def popitem(self):
|
|
|
|
self._assert_mutable()
|
|
|
|
return super(QueryDict, self).popitem()
|
|
|
|
|
|
|
|
def clear(self):
|
|
|
|
self._assert_mutable()
|
|
|
|
super(QueryDict, self).clear()
|
|
|
|
|
|
|
|
def setdefault(self, key, default=None):
|
|
|
|
self._assert_mutable()
|
|
|
|
key = bytes_to_text(key, self.encoding)
|
|
|
|
default = bytes_to_text(default, self.encoding)
|
|
|
|
return super(QueryDict, self).setdefault(key, default)
|
|
|
|
|
|
|
|
def copy(self):
|
|
|
|
"""Returns a mutable copy of this object."""
|
|
|
|
return self.__deepcopy__({})
|
|
|
|
|
|
|
|
def urlencode(self, safe=None):
|
|
|
|
"""
|
|
|
|
Returns an encoded string of all query string arguments.
|
|
|
|
|
|
|
|
:arg safe: Used to specify characters which do not require quoting, for
|
|
|
|
example::
|
|
|
|
|
|
|
|
>>> q = QueryDict('', mutable=True)
|
|
|
|
>>> q['next'] = '/a&b/'
|
|
|
|
>>> q.urlencode()
|
|
|
|
'next=%2Fa%26b%2F'
|
|
|
|
>>> q.urlencode(safe='/')
|
|
|
|
'next=/a%26b/'
|
|
|
|
|
|
|
|
"""
|
|
|
|
output = []
|
|
|
|
if safe:
|
|
|
|
safe = force_bytes(safe, self.encoding)
|
|
|
|
encode = lambda k, v: '%s=%s' % ((quote(k, safe), quote(v, safe)))
|
|
|
|
else:
|
|
|
|
encode = lambda k, v: urlencode({k: v})
|
|
|
|
for k, list_ in self.lists():
|
|
|
|
k = force_bytes(k, self.encoding)
|
2014-12-07 05:00:09 +08:00
|
|
|
output.extend(encode(k, force_bytes(v, self.encoding))
|
|
|
|
for v in list_)
|
2012-10-22 02:12:59 +08:00
|
|
|
return '&'.join(output)
|
|
|
|
|
|
|
|
|
|
|
|
def build_request_repr(request, path_override=None, GET_override=None,
|
|
|
|
POST_override=None, COOKIES_override=None,
|
|
|
|
META_override=None):
|
|
|
|
"""
|
|
|
|
Builds and returns the request's representation string. The request's
|
|
|
|
attributes may be overridden by pre-processed values.
|
|
|
|
"""
|
|
|
|
# Since this is called as part of error handling, we need to be very
|
|
|
|
# robust against potentially malformed input.
|
|
|
|
try:
|
|
|
|
get = (pformat(GET_override)
|
|
|
|
if GET_override is not None
|
|
|
|
else pformat(request.GET))
|
|
|
|
except Exception:
|
|
|
|
get = '<could not parse>'
|
|
|
|
if request._post_parse_error:
|
|
|
|
post = '<could not parse>'
|
|
|
|
else:
|
|
|
|
try:
|
|
|
|
post = (pformat(POST_override)
|
|
|
|
if POST_override is not None
|
|
|
|
else pformat(request.POST))
|
|
|
|
except Exception:
|
|
|
|
post = '<could not parse>'
|
|
|
|
try:
|
|
|
|
cookies = (pformat(COOKIES_override)
|
|
|
|
if COOKIES_override is not None
|
|
|
|
else pformat(request.COOKIES))
|
|
|
|
except Exception:
|
|
|
|
cookies = '<could not parse>'
|
|
|
|
try:
|
|
|
|
meta = (pformat(META_override)
|
|
|
|
if META_override is not None
|
|
|
|
else pformat(request.META))
|
|
|
|
except Exception:
|
|
|
|
meta = '<could not parse>'
|
|
|
|
path = path_override if path_override is not None else request.path
|
|
|
|
return force_str('<%s\npath:%s,\nGET:%s,\nPOST:%s,\nCOOKIES:%s,\nMETA:%s>' %
|
|
|
|
(request.__class__.__name__,
|
|
|
|
path,
|
|
|
|
six.text_type(get),
|
|
|
|
six.text_type(post),
|
|
|
|
six.text_type(cookies),
|
|
|
|
six.text_type(meta)))
|
|
|
|
|
|
|
|
|
|
|
|
# It's neither necessary nor appropriate to use
|
|
|
|
# django.utils.encoding.smart_text for parsing URLs and form inputs. Thus,
|
|
|
|
# this slightly more restricted function, used by QueryDict.
|
|
|
|
def bytes_to_text(s, encoding):
|
|
|
|
"""
|
|
|
|
Converts basestring objects to unicode, using the given encoding. Illegally
|
|
|
|
encoded input characters are replaced with Unicode "unknown" codepoint
|
|
|
|
(\ufffd).
|
|
|
|
|
|
|
|
Returns any non-basestring objects without change.
|
|
|
|
"""
|
|
|
|
if isinstance(s, bytes):
|
|
|
|
return six.text_type(s, encoding, 'replace')
|
|
|
|
else:
|
|
|
|
return s
|
2013-02-10 01:17:01 +08:00
|
|
|
|
|
|
|
|
2013-04-04 04:27:20 +08:00
|
|
|
def split_domain_port(host):
|
|
|
|
"""
|
|
|
|
Return a (domain, port) tuple from a given host.
|
|
|
|
|
|
|
|
Returned domain is lower-cased. If the host is invalid, the domain will be
|
|
|
|
empty.
|
|
|
|
"""
|
|
|
|
host = host.lower()
|
|
|
|
|
|
|
|
if not host_validation_re.match(host):
|
|
|
|
return '', ''
|
|
|
|
|
|
|
|
if host[-1] == ']':
|
|
|
|
# It's an IPv6 address without a port.
|
|
|
|
return host, ''
|
|
|
|
bits = host.rsplit(':', 1)
|
|
|
|
if len(bits) == 2:
|
|
|
|
return tuple(bits)
|
|
|
|
return bits[0], ''
|
|
|
|
|
|
|
|
|
2013-02-10 01:17:01 +08:00
|
|
|
def validate_host(host, allowed_hosts):
|
|
|
|
"""
|
2013-04-04 04:27:20 +08:00
|
|
|
Validate the given host for this site.
|
2013-02-10 01:17:01 +08:00
|
|
|
|
|
|
|
Check that the host looks valid and matches a host or host pattern in the
|
|
|
|
given list of ``allowed_hosts``. Any pattern beginning with a period
|
|
|
|
matches a domain and all its subdomains (e.g. ``.example.com`` matches
|
|
|
|
``example.com`` and any subdomain), ``*`` matches anything, and anything
|
|
|
|
else must match exactly.
|
|
|
|
|
2013-04-04 04:27:20 +08:00
|
|
|
Note: This function assumes that the given host is lower-cased and has
|
|
|
|
already had the port, if any, stripped off.
|
|
|
|
|
2013-02-10 01:17:01 +08:00
|
|
|
Return ``True`` for a valid host, ``False`` otherwise.
|
|
|
|
|
|
|
|
"""
|
2013-10-25 01:34:40 +08:00
|
|
|
host = host[:-1] if host.endswith('.') else host
|
|
|
|
|
2013-02-10 01:17:01 +08:00
|
|
|
for pattern in allowed_hosts:
|
|
|
|
pattern = pattern.lower()
|
|
|
|
match = (
|
|
|
|
pattern == '*' or
|
|
|
|
pattern.startswith('.') and (
|
2013-04-04 04:27:20 +08:00
|
|
|
host.endswith(pattern) or host == pattern[1:]
|
2013-10-18 17:02:43 +08:00
|
|
|
) or
|
2013-04-04 04:27:20 +08:00
|
|
|
pattern == host
|
2013-10-18 17:02:43 +08:00
|
|
|
)
|
2013-02-10 01:17:01 +08:00
|
|
|
if match:
|
|
|
|
return True
|
|
|
|
|
|
|
|
return False
|