2014-11-15 00:48:00 +08:00
|
|
|
from collections import namedtuple
|
2012-08-20 16:46:21 +08:00
|
|
|
import re
|
|
|
|
from .base import FIELD_TYPE
|
2013-08-11 03:00:12 +08:00
|
|
|
from django.utils.datastructures import OrderedSet
|
2014-09-21 03:34:23 +08:00
|
|
|
from django.db.backends import BaseDatabaseIntrospection, FieldInfo, TableInfo
|
2013-04-02 01:51:53 +08:00
|
|
|
from django.utils.encoding import force_text
|
2012-08-20 16:46:21 +08:00
|
|
|
|
2014-11-15 00:48:00 +08:00
|
|
|
FieldInfo = namedtuple('FieldInfo', FieldInfo._fields + ('extra',))
|
2006-05-02 09:31:56 +08:00
|
|
|
|
|
|
|
foreign_key_re = re.compile(r"\sCONSTRAINT `[^`]*` FOREIGN KEY \(`([^`]*)`\) REFERENCES `([^`]*)` \(`([^`]*)`\)")
|
|
|
|
|
2013-07-08 08:39:54 +08:00
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
class DatabaseIntrospection(BaseDatabaseIntrospection):
|
|
|
|
data_types_reverse = {
|
|
|
|
FIELD_TYPE.BLOB: 'TextField',
|
|
|
|
FIELD_TYPE.CHAR: 'CharField',
|
|
|
|
FIELD_TYPE.DECIMAL: 'DecimalField',
|
2008-08-23 12:14:23 +08:00
|
|
|
FIELD_TYPE.NEWDECIMAL: 'DecimalField',
|
2008-08-11 20:11:25 +08:00
|
|
|
FIELD_TYPE.DATE: 'DateField',
|
|
|
|
FIELD_TYPE.DATETIME: 'DateTimeField',
|
|
|
|
FIELD_TYPE.DOUBLE: 'FloatField',
|
|
|
|
FIELD_TYPE.FLOAT: 'FloatField',
|
|
|
|
FIELD_TYPE.INT24: 'IntegerField',
|
|
|
|
FIELD_TYPE.LONG: 'IntegerField',
|
2009-12-17 23:10:38 +08:00
|
|
|
FIELD_TYPE.LONGLONG: 'BigIntegerField',
|
2014-08-26 13:22:55 +08:00
|
|
|
FIELD_TYPE.SHORT: 'SmallIntegerField',
|
2008-08-11 20:11:25 +08:00
|
|
|
FIELD_TYPE.STRING: 'CharField',
|
2013-02-01 03:40:53 +08:00
|
|
|
FIELD_TYPE.TIME: 'TimeField',
|
2008-08-11 20:11:25 +08:00
|
|
|
FIELD_TYPE.TIMESTAMP: 'DateTimeField',
|
|
|
|
FIELD_TYPE.TINY: 'IntegerField',
|
|
|
|
FIELD_TYPE.TINY_BLOB: 'TextField',
|
|
|
|
FIELD_TYPE.MEDIUM_BLOB: 'TextField',
|
|
|
|
FIELD_TYPE.LONG_BLOB: 'TextField',
|
|
|
|
FIELD_TYPE.VAR_STRING: 'CharField',
|
|
|
|
}
|
2006-05-02 09:31:56 +08:00
|
|
|
|
2014-11-15 00:48:00 +08:00
|
|
|
def get_field_type(self, data_type, description):
|
|
|
|
field_type = super(DatabaseIntrospection, self).get_field_type(data_type, description)
|
|
|
|
if field_type == 'IntegerField' and 'auto_increment' in description.extra:
|
|
|
|
return 'AutoField'
|
|
|
|
return field_type
|
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
def get_table_list(self, cursor):
|
2014-09-21 03:34:23 +08:00
|
|
|
"""
|
|
|
|
Returns a list of table and view names in the current database.
|
|
|
|
"""
|
|
|
|
cursor.execute("SHOW FULL TABLES")
|
|
|
|
return [TableInfo(row[0], {'BASE TABLE': 't', 'VIEW': 'v'}.get(row[1]))
|
|
|
|
for row in cursor.fetchall()]
|
2006-05-02 09:31:56 +08:00
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
def get_table_description(self, cursor, table_name):
|
2012-08-31 01:28:13 +08:00
|
|
|
"""
|
|
|
|
Returns a description of the table, with the DB-API cursor.description interface."
|
|
|
|
"""
|
2014-11-15 00:48:00 +08:00
|
|
|
# information_schema database gives more accurate results for some figures:
|
|
|
|
# - varchar length returned by cursor.description is an internal length,
|
|
|
|
# not visible length (#5725)
|
|
|
|
# - precision and scale (for decimal fields) (#5014)
|
|
|
|
# - auto_increment is not available in cursor.description
|
|
|
|
InfoLine = namedtuple('InfoLine', 'col_name data_type max_len num_prec num_scale extra')
|
2013-04-02 00:17:00 +08:00
|
|
|
cursor.execute("""
|
2014-11-15 00:48:00 +08:00
|
|
|
SELECT column_name, data_type, character_maximum_length, numeric_precision, numeric_scale, extra
|
|
|
|
FROM information_schema.columns
|
|
|
|
WHERE table_name = %s AND table_schema = DATABASE()""", [table_name])
|
|
|
|
field_info = dict((line[0], InfoLine(*line)) for line in cursor.fetchall())
|
2013-04-02 00:17:00 +08:00
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
cursor.execute("SELECT * FROM %s LIMIT 1" % self.connection.ops.quote_name(table_name))
|
2014-11-15 00:48:00 +08:00
|
|
|
to_int = lambda i: int(i) if i is not None else i
|
|
|
|
fields = []
|
|
|
|
for line in cursor.description:
|
|
|
|
col_name = force_text(line[0])
|
|
|
|
fields.append(
|
|
|
|
FieldInfo(*((col_name,)
|
2013-04-02 01:51:53 +08:00
|
|
|
+ line[1:3]
|
2014-11-15 00:48:00 +08:00
|
|
|
+ (to_int(field_info[col_name].max_len) or line[3],
|
|
|
|
to_int(field_info[col_name].num_prec) or line[4],
|
|
|
|
to_int(field_info[col_name].num_scale) or line[5])
|
|
|
|
+ (line[6],)
|
|
|
|
+ (field_info[col_name].extra,)))
|
|
|
|
)
|
|
|
|
return fields
|
2006-05-02 09:31:56 +08:00
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
def _name_to_index(self, cursor, table_name):
|
|
|
|
"""
|
|
|
|
Returns a dictionary of {field_name: field_index} for the given table.
|
|
|
|
Indexes are 0-based.
|
|
|
|
"""
|
2013-08-30 07:20:00 +08:00
|
|
|
return dict((d[0], i) for i, d in enumerate(self.get_table_description(cursor, table_name)))
|
2008-08-11 20:11:25 +08:00
|
|
|
|
|
|
|
def get_relations(self, cursor, table_name):
|
|
|
|
"""
|
|
|
|
Returns a dictionary of {field_index: (field_index_other_table, other_table)}
|
|
|
|
representing all relationships to the given table. Indexes are 0-based.
|
|
|
|
"""
|
|
|
|
my_field_dict = self._name_to_index(cursor, table_name)
|
2011-08-07 08:43:26 +08:00
|
|
|
constraints = self.get_key_columns(cursor, table_name)
|
2008-08-11 20:11:25 +08:00
|
|
|
relations = {}
|
2011-08-07 08:43:26 +08:00
|
|
|
for my_fieldname, other_table, other_field in constraints:
|
|
|
|
other_field_index = self._name_to_index(cursor, other_table)[other_field]
|
|
|
|
my_field_index = my_field_dict[my_fieldname]
|
|
|
|
relations[my_field_index] = (other_field_index, other_table)
|
|
|
|
return relations
|
|
|
|
|
|
|
|
def get_key_columns(self, cursor, table_name):
|
|
|
|
"""
|
|
|
|
Returns a list of (column_name, referenced_table_name, referenced_column_name) for all
|
|
|
|
key columns in given table.
|
|
|
|
"""
|
|
|
|
key_columns = []
|
2012-04-21 11:04:10 +08:00
|
|
|
cursor.execute("""
|
|
|
|
SELECT column_name, referenced_table_name, referenced_column_name
|
|
|
|
FROM information_schema.key_column_usage
|
|
|
|
WHERE table_name = %s
|
|
|
|
AND table_schema = DATABASE()
|
|
|
|
AND referenced_table_name IS NOT NULL
|
|
|
|
AND referenced_column_name IS NOT NULL""", [table_name])
|
|
|
|
key_columns.extend(cursor.fetchall())
|
2011-08-07 08:43:26 +08:00
|
|
|
return key_columns
|
2006-05-02 09:31:56 +08:00
|
|
|
|
2008-08-11 20:11:25 +08:00
|
|
|
def get_indexes(self, cursor, table_name):
|
|
|
|
cursor.execute("SHOW INDEX FROM %s" % self.connection.ops.quote_name(table_name))
|
2012-04-30 19:05:30 +08:00
|
|
|
# Do a two-pass search for indexes: on first pass check which indexes
|
|
|
|
# are multicolumn, on second pass check which single-column indexes
|
|
|
|
# are present.
|
|
|
|
rows = list(cursor.fetchall())
|
|
|
|
multicol_indexes = set()
|
|
|
|
for row in rows:
|
|
|
|
if row[3] > 1:
|
|
|
|
multicol_indexes.add(row[2])
|
2008-08-11 20:11:25 +08:00
|
|
|
indexes = {}
|
2012-04-30 19:05:30 +08:00
|
|
|
for row in rows:
|
|
|
|
if row[2] in multicol_indexes:
|
|
|
|
continue
|
2012-09-05 00:53:31 +08:00
|
|
|
if row[4] not in indexes:
|
|
|
|
indexes[row[4]] = {'primary_key': False, 'unique': False}
|
2012-09-18 06:45:00 +08:00
|
|
|
# It's possible to have the unique and PK constraints in separate indexes.
|
2012-09-05 00:53:31 +08:00
|
|
|
if row[2] == 'PRIMARY':
|
|
|
|
indexes[row[4]]['primary_key'] = True
|
2013-10-17 20:27:34 +08:00
|
|
|
if not row[1]:
|
2012-09-05 00:53:31 +08:00
|
|
|
indexes[row[4]]['unique'] = True
|
2008-08-11 20:11:25 +08:00
|
|
|
return indexes
|
2006-05-02 09:31:56 +08:00
|
|
|
|
2014-09-24 02:13:59 +08:00
|
|
|
def get_storage_engine(self, cursor, table_name):
|
|
|
|
"""
|
|
|
|
Retrieves the storage engine for a given table.
|
|
|
|
"""
|
|
|
|
cursor.execute(
|
|
|
|
"SELECT engine "
|
|
|
|
"FROM information_schema.tables "
|
|
|
|
"WHERE table_name = %s", [table_name])
|
|
|
|
return cursor.fetchone()[0]
|
|
|
|
|
2012-08-18 19:29:31 +08:00
|
|
|
def get_constraints(self, cursor, table_name):
|
|
|
|
"""
|
2012-08-31 06:11:56 +08:00
|
|
|
Retrieves any constraints or keys (unique, pk, fk, check, index) across one or more columns.
|
2012-08-18 19:29:31 +08:00
|
|
|
"""
|
|
|
|
constraints = {}
|
2012-08-18 20:48:54 +08:00
|
|
|
# Get the actual constraint names and columns
|
|
|
|
name_query = """
|
|
|
|
SELECT kc.`constraint_name`, kc.`column_name`,
|
|
|
|
kc.`referenced_table_name`, kc.`referenced_column_name`
|
|
|
|
FROM information_schema.key_column_usage AS kc
|
|
|
|
WHERE
|
|
|
|
kc.table_schema = %s AND
|
|
|
|
kc.table_name = %s
|
|
|
|
"""
|
|
|
|
cursor.execute(name_query, [self.connection.settings_dict['NAME'], table_name])
|
|
|
|
for constraint, column, ref_table, ref_column in cursor.fetchall():
|
|
|
|
if constraint not in constraints:
|
|
|
|
constraints[constraint] = {
|
2013-08-11 03:00:12 +08:00
|
|
|
'columns': OrderedSet(),
|
2012-08-18 20:48:54 +08:00
|
|
|
'primary_key': False,
|
|
|
|
'unique': False,
|
2012-08-31 06:11:56 +08:00
|
|
|
'index': False,
|
|
|
|
'check': False,
|
2012-08-18 20:48:54 +08:00
|
|
|
'foreign_key': (ref_table, ref_column) if ref_column else None,
|
|
|
|
}
|
|
|
|
constraints[constraint]['columns'].add(column)
|
|
|
|
# Now get the constraint types
|
|
|
|
type_query = """
|
|
|
|
SELECT c.constraint_name, c.constraint_type
|
|
|
|
FROM information_schema.table_constraints AS c
|
|
|
|
WHERE
|
|
|
|
c.table_schema = %s AND
|
|
|
|
c.table_name = %s
|
|
|
|
"""
|
|
|
|
cursor.execute(type_query, [self.connection.settings_dict['NAME'], table_name])
|
|
|
|
for constraint, kind in cursor.fetchall():
|
|
|
|
if kind.lower() == "primary key":
|
|
|
|
constraints[constraint]['primary_key'] = True
|
|
|
|
constraints[constraint]['unique'] = True
|
|
|
|
elif kind.lower() == "unique":
|
|
|
|
constraints[constraint]['unique'] = True
|
2012-08-31 06:11:56 +08:00
|
|
|
# Now add in the indexes
|
|
|
|
cursor.execute("SHOW INDEX FROM %s" % self.connection.ops.quote_name(table_name))
|
|
|
|
for table, non_unique, index, colseq, column in [x[:5] for x in cursor.fetchall()]:
|
|
|
|
if index not in constraints:
|
|
|
|
constraints[index] = {
|
2013-08-11 03:00:12 +08:00
|
|
|
'columns': OrderedSet(),
|
2012-08-31 06:11:56 +08:00
|
|
|
'primary_key': False,
|
|
|
|
'unique': False,
|
|
|
|
'index': True,
|
|
|
|
'check': False,
|
|
|
|
'foreign_key': None,
|
|
|
|
}
|
|
|
|
constraints[index]['index'] = True
|
|
|
|
constraints[index]['columns'].add(column)
|
2013-07-03 01:02:20 +08:00
|
|
|
# Convert the sorted sets to lists
|
|
|
|
for constraint in constraints.values():
|
|
|
|
constraint['columns'] = list(constraint['columns'])
|
2012-08-18 19:29:31 +08:00
|
|
|
return constraints
|