2019-02-01 08:50:16 +08:00
|
|
|
import datetime
|
2013-07-01 20:22:27 +08:00
|
|
|
import unittest
|
2015-01-28 20:35:27 +08:00
|
|
|
from copy import copy
|
|
|
|
from decimal import Decimal
|
2020-11-28 14:28:03 +08:00
|
|
|
from pathlib import Path
|
2009-12-22 23:18:51 +08:00
|
|
|
|
2015-01-28 20:35:27 +08:00
|
|
|
from django.conf import settings
|
2017-05-04 23:03:23 +08:00
|
|
|
from django.contrib.gis.gdal import DataSource
|
|
|
|
from django.contrib.gis.utils.layermapping import (
|
|
|
|
InvalidDecimal,
|
|
|
|
InvalidString,
|
|
|
|
LayerMapError,
|
|
|
|
LayerMapping,
|
|
|
|
MissingForeignKey,
|
|
|
|
)
|
2014-11-30 01:09:40 +08:00
|
|
|
from django.db import connection
|
2017-05-04 22:14:35 +08:00
|
|
|
from django.test import TestCase, override_settings
|
2009-12-22 23:18:51 +08:00
|
|
|
|
2017-05-04 23:03:23 +08:00
|
|
|
from .models import (
|
2019-02-01 08:50:16 +08:00
|
|
|
City,
|
|
|
|
County,
|
|
|
|
CountyFeat,
|
|
|
|
DoesNotAllowNulls,
|
|
|
|
HasNulls,
|
|
|
|
ICity1,
|
|
|
|
ICity2,
|
|
|
|
Interstate,
|
|
|
|
Invalid,
|
|
|
|
State,
|
|
|
|
city_mapping,
|
|
|
|
co_mapping,
|
|
|
|
cofeat_mapping,
|
|
|
|
has_nulls_mapping,
|
|
|
|
inter_mapping,
|
2017-05-04 23:03:23 +08:00
|
|
|
)
|
2011-10-18 02:45:22 +08:00
|
|
|
|
2020-11-28 14:28:03 +08:00
|
|
|
shp_path = Path(__file__).resolve().parent.parent / "data"
|
|
|
|
city_shp = shp_path / "cities" / "cities.shp"
|
|
|
|
co_shp = shp_path / "counties" / "counties.shp"
|
|
|
|
inter_shp = shp_path / "interstates" / "interstates.shp"
|
|
|
|
invalid_shp = shp_path / "invalid" / "emptypoints.shp"
|
|
|
|
has_nulls_geojson = shp_path / "has_nulls" / "has_nulls.geojson"
|
2008-08-06 02:13:06 +08:00
|
|
|
|
2010-10-11 20:55:17 +08:00
|
|
|
# Dictionaries to hold what's expected in the county shapefile.
|
2013-10-22 21:31:43 +08:00
|
|
|
NAMES = ["Bexar", "Galveston", "Harris", "Honolulu", "Pueblo"]
|
2013-11-03 05:02:56 +08:00
|
|
|
NUMS = [1, 2, 1, 19, 1] # Number of polygons for each.
|
2008-08-06 02:13:06 +08:00
|
|
|
STATES = ["Texas", "Texas", "Texas", "Hawaii", "Colorado"]
|
|
|
|
|
2012-09-20 16:08:08 +08:00
|
|
|
|
2010-12-02 13:58:21 +08:00
|
|
|
class LayerMapTest(TestCase):
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_init(self):
|
2008-08-06 02:13:06 +08:00
|
|
|
"Testing LayerMapping initialization."
|
|
|
|
|
|
|
|
# Model field that does not exist.
|
|
|
|
bad1 = copy(city_mapping)
|
|
|
|
bad1["foobar"] = "FooField"
|
|
|
|
|
|
|
|
# Shapefile field that does not exist.
|
|
|
|
bad2 = copy(city_mapping)
|
|
|
|
bad2["name"] = "Nombre"
|
|
|
|
|
|
|
|
# Nonexistent geographic field type.
|
|
|
|
bad3 = copy(city_mapping)
|
|
|
|
bad3["point"] = "CURVE"
|
|
|
|
|
|
|
|
# Incrementing through the bad mapping dictionaries and
|
|
|
|
# ensuring that a LayerMapError is raised.
|
|
|
|
for bad_map in (bad1, bad2, bad3):
|
2012-05-24 15:52:06 +08:00
|
|
|
with self.assertRaises(LayerMapError):
|
2013-09-08 23:05:16 +08:00
|
|
|
LayerMapping(City, city_shp, bad_map)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# A LookupError should be thrown for bogus encodings.
|
2012-05-24 15:52:06 +08:00
|
|
|
with self.assertRaises(LookupError):
|
2013-09-08 23:05:16 +08:00
|
|
|
LayerMapping(City, city_shp, city_mapping, encoding="foobar")
|
2008-08-06 02:13:06 +08:00
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_simple_layermap(self):
|
2008-08-06 02:13:06 +08:00
|
|
|
"Test LayerMapping import of a simple point shapefile."
|
|
|
|
# Setting up for the LayerMapping.
|
|
|
|
lm = LayerMapping(City, city_shp, city_mapping)
|
|
|
|
lm.save()
|
|
|
|
|
|
|
|
# There should be three cities in the shape file.
|
|
|
|
self.assertEqual(3, City.objects.count())
|
|
|
|
|
|
|
|
# Opening up the shapefile, and verifying the values in each
|
|
|
|
# of the features made it to the model.
|
|
|
|
ds = DataSource(city_shp)
|
|
|
|
layer = ds[0]
|
|
|
|
for feat in layer:
|
|
|
|
city = City.objects.get(name=feat["Name"].value)
|
|
|
|
self.assertEqual(feat["Population"].value, city.population)
|
|
|
|
self.assertEqual(Decimal(str(feat["Density"])), city.density)
|
|
|
|
self.assertEqual(feat["Created"].value, city.dt)
|
|
|
|
|
|
|
|
# Comparing the geometries.
|
|
|
|
pnt1, pnt2 = feat.geom, city.point
|
2011-09-10 03:51:06 +08:00
|
|
|
self.assertAlmostEqual(pnt1.x, pnt2.x, 5)
|
|
|
|
self.assertAlmostEqual(pnt1.y, pnt2.y, 5)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
2020-11-28 14:28:03 +08:00
|
|
|
def test_data_source_str(self):
|
|
|
|
lm = LayerMapping(City, str(city_shp), city_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(City.objects.count(), 3)
|
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_layermap_strict(self):
|
2008-08-06 02:13:06 +08:00
|
|
|
"Testing the `strict` keyword, and import of a LineString shapefile."
|
|
|
|
# When the `strict` keyword is set an error encountered will force
|
|
|
|
# the importation to stop.
|
2012-05-24 15:52:06 +08:00
|
|
|
with self.assertRaises(InvalidDecimal):
|
2008-08-06 02:13:06 +08:00
|
|
|
lm = LayerMapping(Interstate, inter_shp, inter_mapping)
|
|
|
|
lm.save(silent=True, strict=True)
|
2012-05-24 15:52:06 +08:00
|
|
|
Interstate.objects.all().delete()
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# This LayerMapping should work b/c `strict` is not set.
|
|
|
|
lm = LayerMapping(Interstate, inter_shp, inter_mapping)
|
|
|
|
lm.save(silent=True)
|
|
|
|
|
|
|
|
# Two interstate should have imported correctly.
|
|
|
|
self.assertEqual(2, Interstate.objects.count())
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
# Verifying the values in the layer w/the model.
|
|
|
|
ds = DataSource(inter_shp)
|
|
|
|
|
|
|
|
# Only the first two features of this shapefile are valid.
|
|
|
|
valid_feats = ds[0][:2]
|
|
|
|
for feat in valid_feats:
|
|
|
|
istate = Interstate.objects.get(name=feat["Name"].value)
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
if feat.fid == 0:
|
|
|
|
self.assertEqual(Decimal(str(feat["Length"])), istate.length)
|
|
|
|
elif feat.fid == 1:
|
|
|
|
# Everything but the first two decimal digits were truncated,
|
|
|
|
# because the Interstate model's `length` field has decimal_places=2.
|
|
|
|
self.assertAlmostEqual(feat.get("Length"), float(istate.length), 2)
|
|
|
|
|
|
|
|
for p1, p2 in zip(feat.geom, istate.path):
|
|
|
|
self.assertAlmostEqual(p1[0], p2[0], 6)
|
|
|
|
self.assertAlmostEqual(p1[1], p2[1], 6)
|
|
|
|
|
|
|
|
def county_helper(self, county_feat=True):
|
|
|
|
"Helper function for ensuring the integrity of the mapped County models."
|
|
|
|
for name, n, st in zip(NAMES, NUMS, STATES):
|
|
|
|
# Should only be one record b/c of `unique` keyword.
|
|
|
|
c = County.objects.get(name=name)
|
|
|
|
self.assertEqual(n, len(c.mpoly))
|
2013-11-03 05:02:56 +08:00
|
|
|
self.assertEqual(st, c.state.name) # Checking ForeignKey mapping.
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
# Multiple records because `unique` was not set.
|
|
|
|
if county_feat:
|
|
|
|
qs = CountyFeat.objects.filter(name=name)
|
|
|
|
self.assertEqual(n, qs.count())
|
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_layermap_unique_multigeometry_fk(self):
|
2008-08-06 02:13:06 +08:00
|
|
|
"""
|
|
|
|
The `unique`, and `transform`, geometry collection conversion, and
|
|
|
|
ForeignKey mappings.
|
2022-02-04 15:08:27 +08:00
|
|
|
"""
|
2008-08-06 02:13:06 +08:00
|
|
|
# All the following should work.
|
2015-06-05 15:32:29 +08:00
|
|
|
|
|
|
|
# Telling LayerMapping that we want no transformations performed on the data.
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False)
|
|
|
|
|
|
|
|
# Specifying the source spatial reference system via the `source_srs` keyword.
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, source_srs=4269)
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, source_srs="NAD83")
|
|
|
|
|
|
|
|
# Unique may take tuple or string parameters.
|
|
|
|
for arg in ("name", ("name", "mpoly")):
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False, unique=arg)
|
|
|
|
|
|
|
|
# Now test for failures
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
# Testing invalid params for the `unique` keyword.
|
|
|
|
for e, arg in (
|
|
|
|
(TypeError, 5.0),
|
|
|
|
(ValueError, "foobar"),
|
|
|
|
(ValueError, ("name", "mpolygon")),
|
|
|
|
):
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(e):
|
|
|
|
LayerMapping(County, co_shp, co_mapping, transform=False, unique=arg)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# No source reference system defined in the shapefile, should raise an error.
|
2014-08-22 00:47:57 +08:00
|
|
|
if connection.features.supports_transform:
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(LayerMapError):
|
|
|
|
LayerMapping(County, co_shp, co_mapping)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# Passing in invalid ForeignKey mapping parameters -- must be a dictionary
|
|
|
|
# mapping for the model the ForeignKey points to.
|
2013-10-11 19:25:14 +08:00
|
|
|
bad_fk_map1 = copy(co_mapping)
|
|
|
|
bad_fk_map1["state"] = "name"
|
|
|
|
bad_fk_map2 = copy(co_mapping)
|
2013-10-27 09:27:42 +08:00
|
|
|
bad_fk_map2["state"] = {"nombre": "State"}
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(TypeError):
|
|
|
|
LayerMapping(County, co_shp, bad_fk_map1, transform=False)
|
|
|
|
with self.assertRaises(LayerMapError):
|
|
|
|
LayerMapping(County, co_shp, bad_fk_map2, transform=False)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# There exist no State models for the ForeignKey mapping to work -- should raise
|
|
|
|
# a MissingForeignKey exception (this error would be ignored if the `strict`
|
|
|
|
# keyword is not set).
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False, unique="name")
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(MissingForeignKey):
|
|
|
|
lm.save(silent=True, strict=True)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# Now creating the state models so the ForeignKey mapping may work.
|
2012-05-24 15:52:06 +08:00
|
|
|
State.objects.bulk_create(
|
|
|
|
[State(name="Colorado"), State(name="Hawaii"), State(name="Texas")]
|
|
|
|
)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# If a mapping is specified as a collection, all OGR fields that
|
|
|
|
# are not collections will be converted into them. For example,
|
|
|
|
# a Point column would be converted to MultiPoint. Other things being done
|
|
|
|
# w/the keyword args:
|
2010-10-11 20:55:17 +08:00
|
|
|
# `transform=False`: Specifies that no transform is to be done; this
|
2008-08-06 02:13:06 +08:00
|
|
|
# has the effect of ignoring the spatial reference check (because the
|
|
|
|
# county shapefile does not have implicit spatial reference info).
|
2010-10-11 20:55:17 +08:00
|
|
|
#
|
|
|
|
# `unique='name'`: Creates models on the condition that they have
|
2008-08-06 02:13:06 +08:00
|
|
|
# unique county names; geometries from each feature however will be
|
|
|
|
# appended to the geometry collection of the unique model. Thus,
|
|
|
|
# all of the various islands in Honolulu county will be in in one
|
|
|
|
# database record with a MULTIPOLYGON type.
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False, unique="name")
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
|
|
|
|
# A reference that doesn't use the unique keyword; a new database record will
|
|
|
|
# created for each polygon.
|
|
|
|
lm = LayerMapping(CountyFeat, co_shp, cofeat_mapping, transform=False)
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
|
|
|
|
# The county helper is called to ensure integrity of County models.
|
|
|
|
self.county_helper()
|
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_test_fid_range_step(self):
|
2008-08-06 02:13:06 +08:00
|
|
|
"Tests the `fid_range` keyword and the `step` keyword of .save()."
|
|
|
|
# Function for clearing out all the counties before testing.
|
2013-10-17 16:17:41 +08:00
|
|
|
def clear_counties():
|
|
|
|
County.objects.all().delete()
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
State.objects.bulk_create(
|
|
|
|
[State(name="Colorado"), State(name="Hawaii"), State(name="Texas")]
|
|
|
|
)
|
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
# Initializing the LayerMapping object to use in these tests.
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False, unique="name")
|
|
|
|
|
|
|
|
# Bad feature id ranges should raise a type error.
|
|
|
|
bad_ranges = (5.0, "foo", co_shp)
|
|
|
|
for bad in bad_ranges:
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(TypeError):
|
|
|
|
lm.save(fid_range=bad)
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# Step keyword should not be allowed w/`fid_range`.
|
2013-11-03 05:02:56 +08:00
|
|
|
fr = (3, 5) # layer[3:5]
|
2016-01-17 19:26:39 +08:00
|
|
|
with self.assertRaises(LayerMapError):
|
|
|
|
lm.save(fid_range=fr, step=10)
|
2008-08-06 02:13:06 +08:00
|
|
|
lm.save(fid_range=fr)
|
2010-10-11 20:55:17 +08:00
|
|
|
|
2008-08-06 02:13:06 +08:00
|
|
|
# Features IDs 3 & 4 are for Galveston County, Texas -- only
|
|
|
|
# one model is returned because the `unique` keyword was set.
|
|
|
|
qs = County.objects.all()
|
|
|
|
self.assertEqual(1, qs.count())
|
|
|
|
self.assertEqual("Galveston", qs[0].name)
|
|
|
|
|
|
|
|
# Features IDs 5 and beyond for Honolulu County, Hawaii, and
|
|
|
|
# FID 0 is for Pueblo County, Colorado.
|
|
|
|
clear_counties()
|
2013-11-03 05:02:56 +08:00
|
|
|
lm.save(fid_range=slice(5, None), silent=True, strict=True) # layer[5:]
|
|
|
|
lm.save(fid_range=slice(None, 1), silent=True, strict=True) # layer[:1]
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# Only Pueblo & Honolulu counties should be present because of
|
2010-10-11 20:55:17 +08:00
|
|
|
# the `unique` keyword. Have to set `order_by` on this QuerySet
|
2008-12-27 04:04:11 +08:00
|
|
|
# or else MySQL will return a different ordering than the other dbs.
|
2010-10-11 20:55:17 +08:00
|
|
|
qs = County.objects.order_by("name")
|
2008-08-06 02:13:06 +08:00
|
|
|
self.assertEqual(2, qs.count())
|
|
|
|
hi, co = tuple(qs)
|
|
|
|
hi_idx, co_idx = tuple(map(NAMES.index, ("Honolulu", "Pueblo")))
|
2012-05-24 15:52:06 +08:00
|
|
|
self.assertEqual("Pueblo", co.name)
|
|
|
|
self.assertEqual(NUMS[co_idx], len(co.mpoly))
|
|
|
|
self.assertEqual("Honolulu", hi.name)
|
|
|
|
self.assertEqual(NUMS[hi_idx], len(hi.mpoly))
|
2008-08-06 02:13:06 +08:00
|
|
|
|
|
|
|
# Testing the `step` keyword -- should get the same counties
|
|
|
|
# regardless of we use a step that divides equally, that is odd,
|
|
|
|
# or that is larger than the dataset.
|
2013-10-27 01:50:40 +08:00
|
|
|
for st in (4, 7, 1000):
|
2008-08-06 02:13:06 +08:00
|
|
|
clear_counties()
|
|
|
|
lm.save(step=st, strict=True)
|
|
|
|
self.county_helper(county_feat=False)
|
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_model_inheritance(self):
|
2009-11-02 10:17:56 +08:00
|
|
|
"Tests LayerMapping on inherited models. See #12093."
|
2019-01-03 07:18:19 +08:00
|
|
|
icity_mapping = {
|
|
|
|
"name": "Name",
|
|
|
|
"population": "Population",
|
|
|
|
"density": "Density",
|
|
|
|
"point": "POINT",
|
|
|
|
"dt": "Created",
|
|
|
|
}
|
2009-11-02 10:17:56 +08:00
|
|
|
# Parent model has geometry field.
|
|
|
|
lm1 = LayerMapping(ICity1, city_shp, icity_mapping)
|
|
|
|
lm1.save()
|
|
|
|
|
|
|
|
# Grandparent has geometry field.
|
|
|
|
lm2 = LayerMapping(ICity2, city_shp, icity_mapping)
|
|
|
|
lm2.save()
|
|
|
|
|
|
|
|
self.assertEqual(6, ICity1.objects.count())
|
|
|
|
self.assertEqual(3, ICity2.objects.count())
|
2011-03-15 10:26:37 +08:00
|
|
|
|
2012-05-24 15:52:06 +08:00
|
|
|
def test_invalid_layer(self):
|
2011-03-15 10:26:37 +08:00
|
|
|
"Tests LayerMapping on invalid geometries. See #15378."
|
|
|
|
invalid_mapping = {"point": "POINT"}
|
|
|
|
lm = LayerMapping(Invalid, invalid_shp, invalid_mapping, source_srs=4326)
|
|
|
|
lm.save(silent=True)
|
2012-05-24 15:57:22 +08:00
|
|
|
|
2014-09-19 00:54:38 +08:00
|
|
|
def test_charfield_too_short(self):
|
|
|
|
mapping = copy(city_mapping)
|
|
|
|
mapping["name_short"] = "Name"
|
|
|
|
lm = LayerMapping(City, city_shp, mapping)
|
|
|
|
with self.assertRaises(InvalidString):
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
|
2012-05-24 15:57:22 +08:00
|
|
|
def test_textfield(self):
|
2016-10-27 15:53:39 +08:00
|
|
|
"String content fits also in a TextField"
|
2012-05-24 15:57:22 +08:00
|
|
|
mapping = copy(city_mapping)
|
|
|
|
mapping["name_txt"] = "Name"
|
|
|
|
lm = LayerMapping(City, city_shp, mapping)
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
self.assertEqual(City.objects.count(), 3)
|
2014-09-19 00:41:58 +08:00
|
|
|
self.assertEqual(City.objects.get(name="Houston").name_txt, "Houston")
|
2012-09-20 16:08:08 +08:00
|
|
|
|
2012-10-07 04:56:47 +08:00
|
|
|
def test_encoded_name(self):
|
|
|
|
"""Test a layer containing utf-8-encoded name"""
|
2020-11-28 14:28:03 +08:00
|
|
|
city_shp = shp_path / "ch-city" / "ch-city.shp"
|
2012-10-07 04:56:47 +08:00
|
|
|
lm = LayerMapping(City, city_shp, city_mapping)
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
self.assertEqual(City.objects.count(), 1)
|
|
|
|
self.assertEqual(City.objects.all()[0].name, "Zürich")
|
2012-09-20 16:08:08 +08:00
|
|
|
|
2018-06-19 20:21:39 +08:00
|
|
|
def test_null_geom_with_unique(self):
|
|
|
|
"""LayerMapping may be created with a unique and a null geometry."""
|
|
|
|
State.objects.bulk_create(
|
|
|
|
[State(name="Colorado"), State(name="Hawaii"), State(name="Texas")]
|
|
|
|
)
|
|
|
|
hw = State.objects.get(name="Hawaii")
|
|
|
|
hu = County.objects.create(name="Honolulu", state=hw, mpoly=None)
|
|
|
|
lm = LayerMapping(County, co_shp, co_mapping, transform=False, unique="name")
|
|
|
|
lm.save(silent=True, strict=True)
|
|
|
|
hu.refresh_from_db()
|
|
|
|
self.assertIsNotNone(hu.mpoly)
|
|
|
|
self.assertEqual(hu.mpoly.ogr.num_coords, 449)
|
|
|
|
|
2019-02-01 08:50:16 +08:00
|
|
|
def test_null_number_imported(self):
|
|
|
|
"""LayerMapping import of GeoJSON with a null numeric value."""
|
|
|
|
lm = LayerMapping(HasNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(HasNulls.objects.count(), 3)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(num=0).count(), 1)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(num__isnull=True).count(), 1)
|
|
|
|
|
|
|
|
def test_null_string_imported(self):
|
|
|
|
"Test LayerMapping import of GeoJSON with a null string value."
|
|
|
|
lm = LayerMapping(HasNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(HasNulls.objects.filter(name="None").count(), 0)
|
|
|
|
num_empty = 1 if connection.features.interprets_empty_strings_as_nulls else 0
|
|
|
|
self.assertEqual(HasNulls.objects.filter(name="").count(), num_empty)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(name__isnull=True).count(), 1)
|
|
|
|
|
|
|
|
def test_nullable_boolean_imported(self):
|
|
|
|
"""LayerMapping import of GeoJSON with a nullable boolean value."""
|
|
|
|
lm = LayerMapping(HasNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(HasNulls.objects.filter(boolean=True).count(), 1)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(boolean=False).count(), 1)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(boolean__isnull=True).count(), 1)
|
|
|
|
|
|
|
|
def test_nullable_datetime_imported(self):
|
|
|
|
"""LayerMapping import of GeoJSON with a nullable date/time value."""
|
|
|
|
lm = LayerMapping(HasNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(
|
|
|
|
HasNulls.objects.filter(datetime__lt=datetime.date(1994, 8, 15)).count(), 1
|
|
|
|
)
|
|
|
|
self.assertEqual(
|
|
|
|
HasNulls.objects.filter(datetime="2018-11-29T03:02:52").count(), 1
|
|
|
|
)
|
|
|
|
self.assertEqual(HasNulls.objects.filter(datetime__isnull=True).count(), 1)
|
|
|
|
|
|
|
|
def test_uuids_imported(self):
|
|
|
|
"""LayerMapping import of GeoJSON with UUIDs."""
|
|
|
|
lm = LayerMapping(HasNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save()
|
|
|
|
self.assertEqual(
|
|
|
|
HasNulls.objects.filter(
|
|
|
|
uuid="1378c26f-cbe6-44b0-929f-eb330d4991f5"
|
|
|
|
).count(),
|
|
|
|
1,
|
|
|
|
)
|
|
|
|
|
|
|
|
def test_null_number_imported_not_allowed(self):
|
|
|
|
"""
|
|
|
|
LayerMapping import of GeoJSON with nulls to fields that don't permit
|
|
|
|
them.
|
|
|
|
"""
|
|
|
|
lm = LayerMapping(DoesNotAllowNulls, has_nulls_geojson, has_nulls_mapping)
|
|
|
|
lm.save(silent=True)
|
|
|
|
# When a model fails to save due to IntegrityError (null in non-null
|
|
|
|
# column), subsequent saves fail with "An error occurred in the current
|
|
|
|
# transaction. You can't execute queries until the end of the 'atomic'
|
|
|
|
# block." On Oracle and MySQL, the one object that did load appears in
|
|
|
|
# this count. On other databases, no records appear.
|
|
|
|
self.assertLessEqual(DoesNotAllowNulls.objects.count(), 1)
|
|
|
|
|
2013-11-03 04:12:09 +08:00
|
|
|
|
2017-01-19 15:39:46 +08:00
|
|
|
class OtherRouter:
|
2012-09-20 16:08:08 +08:00
|
|
|
def db_for_read(self, model, **hints):
|
|
|
|
return "other"
|
|
|
|
|
|
|
|
def db_for_write(self, model, **hints):
|
|
|
|
return self.db_for_read(model, **hints)
|
|
|
|
|
|
|
|
def allow_relation(self, obj1, obj2, **hints):
|
2018-05-11 08:42:44 +08:00
|
|
|
# ContentType objects are created during a post-migrate signal while
|
|
|
|
# performing fixture teardown using the default database alias and
|
|
|
|
# don't abide by the database specified by this router.
|
|
|
|
return True
|
2012-09-20 16:08:08 +08:00
|
|
|
|
2015-02-19 15:27:58 +08:00
|
|
|
def allow_migrate(self, db, app_label, **hints):
|
2012-09-20 16:08:08 +08:00
|
|
|
return True
|
|
|
|
|
|
|
|
|
2014-11-30 01:09:40 +08:00
|
|
|
@override_settings(DATABASE_ROUTERS=[OtherRouter()])
|
2012-09-20 16:08:08 +08:00
|
|
|
class LayerMapRouterTest(TestCase):
|
2018-07-12 12:12:20 +08:00
|
|
|
databases = {"default", "other"}
|
2012-09-20 16:08:08 +08:00
|
|
|
|
2012-10-06 06:26:33 +08:00
|
|
|
@unittest.skipUnless(len(settings.DATABASES) > 1, "multiple databases required")
|
2012-09-20 16:08:08 +08:00
|
|
|
def test_layermapping_default_db(self):
|
|
|
|
lm = LayerMapping(City, city_shp, city_mapping)
|
|
|
|
self.assertEqual(lm.using, "other")
|