2018-10-25 02:02:33 +08:00
|
|
|
from unittest import mock
|
|
|
|
|
2015-01-28 20:35:27 +08:00
|
|
|
from django.apps.registry import apps as global_apps
|
2021-01-21 09:59:20 +08:00
|
|
|
from django.db import DatabaseError, connection, migrations, models
|
2015-08-27 07:49:35 +08:00
|
|
|
from django.db.migrations.exceptions import InvalidMigrationPlan
|
2013-05-31 01:08:58 +08:00
|
|
|
from django.db.migrations.executor import MigrationExecutor
|
2014-11-18 01:13:47 +08:00
|
|
|
from django.db.migrations.graph import MigrationGraph
|
2015-06-03 04:23:07 +08:00
|
|
|
from django.db.migrations.recorder import MigrationRecorder
|
2021-01-21 09:59:20 +08:00
|
|
|
from django.db.migrations.state import ProjectState
|
2018-10-25 02:02:33 +08:00
|
|
|
from django.test import (
|
2018-11-27 03:05:02 +08:00
|
|
|
SimpleTestCase, modify_settings, override_settings, skipUnlessDBFeature,
|
2018-10-25 02:02:33 +08:00
|
|
|
)
|
2021-10-08 21:51:04 +08:00
|
|
|
from django.test.utils import isolate_lru_cache
|
2013-12-23 23:01:13 +08:00
|
|
|
|
2013-10-30 23:17:28 +08:00
|
|
|
from .test_base import MigrationTestBase
|
2013-05-31 01:08:58 +08:00
|
|
|
|
|
|
|
|
2014-06-16 22:15:19 +08:00
|
|
|
@modify_settings(INSTALLED_APPS={'append': 'migrations2'})
|
2013-10-30 23:17:28 +08:00
|
|
|
class ExecutorTests(MigrationTestBase):
|
2013-05-31 01:08:58 +08:00
|
|
|
"""
|
|
|
|
Tests the migration executor (full end-to-end running).
|
|
|
|
|
|
|
|
Bear in mind that if these are failing you should fix the other
|
|
|
|
test failures first, as they may be propagating into here.
|
|
|
|
"""
|
|
|
|
|
2014-07-26 00:35:25 +08:00
|
|
|
available_apps = ["migrations", "migrations2", "django.contrib.auth", "django.contrib.contenttypes"]
|
2013-06-19 22:36:02 +08:00
|
|
|
|
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations"})
|
2013-05-31 01:08:58 +08:00
|
|
|
def test_run(self):
|
|
|
|
"""
|
|
|
|
Tests running a simple set of migrations.
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
# Let's look at the plan first and make sure it's up to scratch
|
|
|
|
plan = executor.migration_plan([("migrations", "0002_second")])
|
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], False),
|
|
|
|
(executor.loader.graph.nodes["migrations", "0002_second"], False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
# Were the tables there before?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_book")
|
2013-05-31 01:08:58 +08:00
|
|
|
# Alright, let's try running it
|
|
|
|
executor.migrate([("migrations", "0002_second")])
|
|
|
|
# Are the tables there now?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableExists("migrations_author")
|
|
|
|
self.assertTableExists("migrations_book")
|
2013-10-24 05:56:54 +08:00
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
2013-08-23 19:07:43 +08:00
|
|
|
# Alright, let's undo what we did
|
2013-10-24 05:56:54 +08:00
|
|
|
plan = executor.migration_plan([("migrations", None)])
|
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0002_second"], True),
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], True),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
executor.migrate([("migrations", None)])
|
|
|
|
# Are the tables gone?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_book")
|
2013-10-24 05:56:54 +08:00
|
|
|
|
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations_squashed"})
|
|
|
|
def test_run_with_squashed(self):
|
|
|
|
"""
|
|
|
|
Tests running a squashed migration from zero (should ignore what it replaces)
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
# Check our leaf node is the squashed one
|
|
|
|
leaves = [key for key in executor.loader.graph.leaf_nodes() if key[0] == "migrations"]
|
|
|
|
self.assertEqual(leaves, [("migrations", "0001_squashed_0002")])
|
|
|
|
# Check the plan
|
|
|
|
plan = executor.migration_plan([("migrations", "0001_squashed_0002")])
|
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_squashed_0002"], False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
# Were the tables there before?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_book")
|
2013-10-24 05:56:54 +08:00
|
|
|
# Alright, let's try running it
|
|
|
|
executor.migrate([("migrations", "0001_squashed_0002")])
|
|
|
|
# Are the tables there now?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableExists("migrations_author")
|
|
|
|
self.assertTableExists("migrations_book")
|
2013-10-24 05:56:54 +08:00
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
# Alright, let's undo what we did. Should also just use squashed.
|
|
|
|
plan = executor.migration_plan([("migrations", None)])
|
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_squashed_0002"], True),
|
|
|
|
],
|
|
|
|
)
|
2013-08-23 19:07:43 +08:00
|
|
|
executor.migrate([("migrations", None)])
|
|
|
|
# Are the tables gone?
|
2013-10-30 23:17:28 +08:00
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_book")
|
2013-08-11 22:28:51 +08:00
|
|
|
|
2021-06-05 13:06:17 +08:00
|
|
|
@override_settings(
|
|
|
|
MIGRATION_MODULES={'migrations': 'migrations.test_migrations_squashed'},
|
|
|
|
)
|
|
|
|
def test_migrate_backward_to_squashed_migration(self):
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
try:
|
|
|
|
self.assertTableNotExists('migrations_author')
|
|
|
|
self.assertTableNotExists('migrations_book')
|
|
|
|
executor.migrate([('migrations', '0001_squashed_0002')])
|
|
|
|
self.assertTableExists('migrations_author')
|
|
|
|
self.assertTableExists('migrations_book')
|
|
|
|
executor.loader.build_graph()
|
|
|
|
# Migrate backward to a squashed migration.
|
|
|
|
executor.migrate([('migrations', '0001_initial')])
|
|
|
|
self.assertTableExists('migrations_author')
|
|
|
|
self.assertTableNotExists('migrations_book')
|
|
|
|
finally:
|
|
|
|
# Unmigrate everything.
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
executor.migrate([('migrations', None)])
|
|
|
|
self.assertTableNotExists('migrations_author')
|
|
|
|
self.assertTableNotExists('migrations_book')
|
|
|
|
|
2016-01-31 04:46:28 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations_non_atomic"})
|
|
|
|
def test_non_atomic_migration(self):
|
|
|
|
"""
|
|
|
|
Applying a non-atomic migration works as expected.
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
with self.assertRaisesMessage(RuntimeError, "Abort migration"):
|
|
|
|
executor.migrate([("migrations", "0001_initial")])
|
|
|
|
self.assertTableExists("migrations_publisher")
|
|
|
|
migrations_apps = executor.loader.project_state(("migrations", "0001_initial")).apps
|
|
|
|
Publisher = migrations_apps.get_model("migrations", "Publisher")
|
|
|
|
self.assertTrue(Publisher.objects.exists())
|
|
|
|
self.assertTableNotExists("migrations_book")
|
|
|
|
|
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations_atomic_operation"})
|
|
|
|
def test_atomic_operation_in_non_atomic_migration(self):
|
|
|
|
"""
|
|
|
|
An atomic operation is properly rolled back inside a non-atomic
|
|
|
|
migration.
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
with self.assertRaisesMessage(RuntimeError, "Abort migration"):
|
|
|
|
executor.migrate([("migrations", "0001_initial")])
|
|
|
|
migrations_apps = executor.loader.project_state(("migrations", "0001_initial")).apps
|
|
|
|
Editor = migrations_apps.get_model("migrations", "Editor")
|
|
|
|
self.assertFalse(Editor.objects.exists())
|
2017-07-12 00:40:18 +08:00
|
|
|
# Record previous migration as successful.
|
|
|
|
executor.migrate([("migrations", "0001_initial")], fake=True)
|
|
|
|
# Rebuild the graph to reflect the new DB state.
|
|
|
|
executor.loader.build_graph()
|
|
|
|
# Migrating backwards is also atomic.
|
|
|
|
with self.assertRaisesMessage(RuntimeError, "Abort migration"):
|
|
|
|
executor.migrate([("migrations", None)])
|
|
|
|
self.assertFalse(Editor.objects.exists())
|
2016-01-31 04:46:28 +08:00
|
|
|
|
2014-06-16 22:15:19 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={
|
|
|
|
"migrations": "migrations.test_migrations",
|
|
|
|
"migrations2": "migrations2.test_migrations_2",
|
|
|
|
})
|
2013-08-11 22:28:51 +08:00
|
|
|
def test_empty_plan(self):
|
|
|
|
"""
|
2016-10-27 15:53:39 +08:00
|
|
|
Re-planning a full migration of a fully-migrated set doesn't
|
2013-08-11 22:28:51 +08:00
|
|
|
perform spurious unmigrations and remigrations.
|
|
|
|
|
|
|
|
There was previously a bug where the executor just always performed the
|
|
|
|
backwards plan for applied migrations - which even for the most recent
|
|
|
|
migration in an app, might include other, dependent apps, and these
|
|
|
|
were being unmigrated.
|
|
|
|
"""
|
|
|
|
# Make the initial plan, check it
|
|
|
|
executor = MigrationExecutor(connection)
|
2014-06-16 22:15:19 +08:00
|
|
|
plan = executor.migration_plan([
|
|
|
|
("migrations", "0002_second"),
|
|
|
|
("migrations2", "0001_initial"),
|
|
|
|
])
|
2013-08-11 22:28:51 +08:00
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], False),
|
|
|
|
(executor.loader.graph.nodes["migrations", "0002_second"], False),
|
2014-06-16 22:15:19 +08:00
|
|
|
(executor.loader.graph.nodes["migrations2", "0001_initial"], False),
|
2013-08-11 22:28:51 +08:00
|
|
|
],
|
|
|
|
)
|
|
|
|
# Fake-apply all migrations
|
2014-06-16 22:15:19 +08:00
|
|
|
executor.migrate([
|
|
|
|
("migrations", "0002_second"),
|
|
|
|
("migrations2", "0001_initial")
|
|
|
|
], fake=True)
|
2013-10-24 05:56:54 +08:00
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
2013-08-11 22:28:51 +08:00
|
|
|
# Now plan a second time and make sure it's empty
|
2014-06-16 22:15:19 +08:00
|
|
|
plan = executor.migration_plan([
|
|
|
|
("migrations", "0002_second"),
|
|
|
|
("migrations2", "0001_initial"),
|
|
|
|
])
|
2013-08-11 22:28:51 +08:00
|
|
|
self.assertEqual(plan, [])
|
2016-08-21 04:34:06 +08:00
|
|
|
# The resulting state should include applied migrations.
|
|
|
|
state = executor.migrate([
|
|
|
|
("migrations", "0002_second"),
|
|
|
|
("migrations2", "0001_initial"),
|
|
|
|
])
|
|
|
|
self.assertIn(('migrations', 'book'), state.models)
|
|
|
|
self.assertIn(('migrations', 'author'), state.models)
|
|
|
|
self.assertIn(('migrations2', 'otherauthor'), state.models)
|
2013-08-11 22:28:51 +08:00
|
|
|
# Erase all the fake records
|
2014-06-16 22:15:19 +08:00
|
|
|
executor.recorder.record_unapplied("migrations2", "0001_initial")
|
2014-06-10 00:52:05 +08:00
|
|
|
executor.recorder.record_unapplied("migrations", "0002_second")
|
|
|
|
executor.recorder.record_unapplied("migrations", "0001_initial")
|
2013-10-30 23:17:28 +08:00
|
|
|
|
2015-08-27 07:49:35 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={
|
|
|
|
"migrations": "migrations.test_migrations",
|
|
|
|
"migrations2": "migrations2.test_migrations_2_no_deps",
|
|
|
|
})
|
|
|
|
def test_mixed_plan_not_supported(self):
|
|
|
|
"""
|
|
|
|
Although the MigrationExecutor interfaces allows for mixed migration
|
|
|
|
plans (combined forwards and backwards migrations) this is not
|
|
|
|
supported.
|
|
|
|
"""
|
|
|
|
# Prepare for mixed plan
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
plan = executor.migration_plan([("migrations", "0002_second")])
|
|
|
|
self.assertEqual(
|
|
|
|
plan,
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], False),
|
|
|
|
(executor.loader.graph.nodes["migrations", "0002_second"], False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
executor.migrate(None, plan)
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
self.assertIn(('migrations', '0001_initial'), executor.loader.applied_migrations)
|
|
|
|
self.assertIn(('migrations', '0002_second'), executor.loader.applied_migrations)
|
|
|
|
self.assertNotIn(('migrations2', '0001_initial'), executor.loader.applied_migrations)
|
|
|
|
|
|
|
|
# Generate mixed plan
|
|
|
|
plan = executor.migration_plan([
|
|
|
|
("migrations", None),
|
|
|
|
("migrations2", "0001_initial"),
|
|
|
|
])
|
|
|
|
msg = (
|
|
|
|
'Migration plans with both forwards and backwards migrations are '
|
|
|
|
'not supported. Please split your migration process into separate '
|
|
|
|
'plans of only forwards OR backwards migrations.'
|
|
|
|
)
|
|
|
|
with self.assertRaisesMessage(InvalidMigrationPlan, msg) as cm:
|
|
|
|
executor.migrate(None, plan)
|
|
|
|
self.assertEqual(
|
|
|
|
cm.exception.args[1],
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0002_second"], True),
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], True),
|
|
|
|
(executor.loader.graph.nodes["migrations2", "0001_initial"], False),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
executor.migrate([
|
|
|
|
("migrations", None),
|
|
|
|
("migrations2", None),
|
|
|
|
])
|
|
|
|
# Are the tables gone?
|
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_book")
|
|
|
|
self.assertTableNotExists("migrations2_otherauthor")
|
|
|
|
|
2013-10-30 23:17:28 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations"})
|
|
|
|
def test_soft_apply(self):
|
|
|
|
"""
|
|
|
|
Tests detection of initial migrations already having been applied.
|
|
|
|
"""
|
|
|
|
state = {"faked": None}
|
2013-10-31 01:42:35 +08:00
|
|
|
|
2015-01-12 04:07:45 +08:00
|
|
|
def fake_storer(phase, migration=None, fake=None):
|
2013-10-30 23:17:28 +08:00
|
|
|
state["faked"] = fake
|
|
|
|
executor = MigrationExecutor(connection, progress_callback=fake_storer)
|
|
|
|
# Were the tables there before?
|
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_tribble")
|
|
|
|
# Run it normally
|
2014-06-23 09:29:27 +08:00
|
|
|
self.assertEqual(
|
|
|
|
executor.migration_plan([("migrations", "0001_initial")]),
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], False),
|
|
|
|
],
|
|
|
|
)
|
2013-10-30 23:17:28 +08:00
|
|
|
executor.migrate([("migrations", "0001_initial")])
|
|
|
|
# Are the tables there now?
|
|
|
|
self.assertTableExists("migrations_author")
|
|
|
|
self.assertTableExists("migrations_tribble")
|
|
|
|
# We shouldn't have faked that one
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(state["faked"], False)
|
2013-10-30 23:17:28 +08:00
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
# Fake-reverse that
|
|
|
|
executor.migrate([("migrations", None)], fake=True)
|
|
|
|
# Are the tables still there?
|
|
|
|
self.assertTableExists("migrations_author")
|
|
|
|
self.assertTableExists("migrations_tribble")
|
|
|
|
# Make sure that was faked
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(state["faked"], True)
|
2013-10-30 23:17:28 +08:00
|
|
|
# Finally, migrate forwards; this should fake-apply our initial migration
|
2014-06-23 09:29:27 +08:00
|
|
|
executor.loader.build_graph()
|
|
|
|
self.assertEqual(
|
|
|
|
executor.migration_plan([("migrations", "0001_initial")]),
|
|
|
|
[
|
|
|
|
(executor.loader.graph.nodes["migrations", "0001_initial"], False),
|
|
|
|
],
|
|
|
|
)
|
2015-02-12 19:48:28 +08:00
|
|
|
# Applying the migration should raise a database level error
|
|
|
|
# because we haven't given the --fake-initial option
|
|
|
|
with self.assertRaises(DatabaseError):
|
|
|
|
executor.migrate([("migrations", "0001_initial")])
|
|
|
|
# Reset the faked state
|
|
|
|
state = {"faked": None}
|
|
|
|
# Allow faking of initial CreateModel operations
|
|
|
|
executor.migrate([("migrations", "0001_initial")], fake_initial=True)
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(state["faked"], True)
|
2013-10-30 23:17:28 +08:00
|
|
|
# And migrate back to clean up the database
|
2014-06-23 09:29:27 +08:00
|
|
|
executor.loader.build_graph()
|
2013-10-30 23:17:28 +08:00
|
|
|
executor.migrate([("migrations", None)])
|
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_tribble")
|
2014-05-09 12:48:10 +08:00
|
|
|
|
|
|
|
@override_settings(
|
2014-07-26 00:35:25 +08:00
|
|
|
MIGRATION_MODULES={
|
|
|
|
"migrations": "migrations.test_migrations_custom_user",
|
2014-09-06 02:06:02 +08:00
|
|
|
"django.contrib.auth": "django.contrib.auth.migrations",
|
2014-07-26 00:35:25 +08:00
|
|
|
},
|
2014-05-09 12:48:10 +08:00
|
|
|
AUTH_USER_MODEL="migrations.Author",
|
|
|
|
)
|
|
|
|
def test_custom_user(self):
|
|
|
|
"""
|
|
|
|
Regression test for #22325 - references to a custom user model defined in the
|
|
|
|
same app are not resolved correctly.
|
|
|
|
"""
|
2021-10-08 21:51:04 +08:00
|
|
|
with isolate_lru_cache(global_apps.get_swappable_settings_name):
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
self.assertTableNotExists('migrations_author')
|
|
|
|
self.assertTableNotExists('migrations_tribble')
|
|
|
|
# Migrate forwards
|
|
|
|
executor.migrate([('migrations', '0001_initial')])
|
|
|
|
self.assertTableExists('migrations_author')
|
|
|
|
self.assertTableExists('migrations_tribble')
|
|
|
|
# The soft-application detection works.
|
|
|
|
# Change table_names to not return auth_user during this as it
|
|
|
|
# wouldn't be there in a normal run, and ensure migrations.Author
|
|
|
|
# exists in the global app registry temporarily.
|
|
|
|
old_table_names = connection.introspection.table_names
|
|
|
|
connection.introspection.table_names = lambda c: [
|
|
|
|
x for x in old_table_names(c) if x != 'auth_user'
|
|
|
|
]
|
|
|
|
migrations_apps = executor.loader.project_state(
|
|
|
|
('migrations', '0001_initial'),
|
|
|
|
).apps
|
|
|
|
global_apps.get_app_config('migrations').models['author'] = (
|
|
|
|
migrations_apps.get_model('migrations', 'author')
|
|
|
|
)
|
|
|
|
try:
|
|
|
|
migration = executor.loader.get_migration('auth', '0001_initial')
|
|
|
|
self.assertIs(executor.detect_soft_applied(None, migration)[0], True)
|
|
|
|
finally:
|
|
|
|
connection.introspection.table_names = old_table_names
|
|
|
|
del global_apps.get_app_config('migrations').models['author']
|
|
|
|
# Migrate back to clean up the database.
|
|
|
|
executor.loader.build_graph()
|
|
|
|
executor.migrate([('migrations', None)])
|
|
|
|
self.assertTableNotExists('migrations_author')
|
|
|
|
self.assertTableNotExists('migrations_tribble')
|
2014-11-18 01:13:47 +08:00
|
|
|
|
2015-12-19 21:53:53 +08:00
|
|
|
@override_settings(
|
|
|
|
MIGRATION_MODULES={
|
|
|
|
"migrations": "migrations.test_add_many_to_many_field_initial",
|
|
|
|
},
|
|
|
|
)
|
|
|
|
def test_detect_soft_applied_add_field_manytomanyfield(self):
|
|
|
|
"""
|
|
|
|
executor.detect_soft_applied() detects ManyToManyField tables from an
|
|
|
|
AddField operation. This checks the case of AddField in a migration
|
|
|
|
with other operations (0001) and the case of AddField in its own
|
|
|
|
migration (0002).
|
|
|
|
"""
|
|
|
|
tables = [
|
|
|
|
# from 0001
|
|
|
|
"migrations_project",
|
|
|
|
"migrations_task",
|
|
|
|
"migrations_project_tasks",
|
|
|
|
# from 0002
|
|
|
|
"migrations_task_projects",
|
|
|
|
]
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
# Create the tables for 0001 but make it look like the migration hasn't
|
|
|
|
# been applied.
|
|
|
|
executor.migrate([("migrations", "0001_initial")])
|
|
|
|
executor.migrate([("migrations", None)], fake=True)
|
|
|
|
for table in tables[:3]:
|
|
|
|
self.assertTableExists(table)
|
|
|
|
# Table detection sees 0001 is applied but not 0002.
|
|
|
|
migration = executor.loader.get_migration("migrations", "0001_initial")
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(executor.detect_soft_applied(None, migration)[0], True)
|
2015-12-19 21:53:53 +08:00
|
|
|
migration = executor.loader.get_migration("migrations", "0002_initial")
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(executor.detect_soft_applied(None, migration)[0], False)
|
2015-12-19 21:53:53 +08:00
|
|
|
|
|
|
|
# Create the tables for both migrations but make it look like neither
|
|
|
|
# has been applied.
|
|
|
|
executor.loader.build_graph()
|
|
|
|
executor.migrate([("migrations", "0001_initial")], fake=True)
|
|
|
|
executor.migrate([("migrations", "0002_initial")])
|
|
|
|
executor.loader.build_graph()
|
|
|
|
executor.migrate([("migrations", None)], fake=True)
|
|
|
|
# Table detection sees 0002 is applied.
|
|
|
|
migration = executor.loader.get_migration("migrations", "0002_initial")
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(executor.detect_soft_applied(None, migration)[0], True)
|
2015-12-19 21:53:53 +08:00
|
|
|
|
|
|
|
# Leave the tables for 0001 except the many-to-many table. That missing
|
|
|
|
# table should cause detect_soft_applied() to return False.
|
|
|
|
with connection.schema_editor() as editor:
|
|
|
|
for table in tables[2:]:
|
|
|
|
editor.execute(editor.sql_delete_table % {"table": table})
|
|
|
|
migration = executor.loader.get_migration("migrations", "0001_initial")
|
2016-06-17 02:19:18 +08:00
|
|
|
self.assertIs(executor.detect_soft_applied(None, migration)[0], False)
|
2015-12-19 21:53:53 +08:00
|
|
|
|
|
|
|
# Cleanup by removing the remaining tables.
|
|
|
|
with connection.schema_editor() as editor:
|
|
|
|
for table in tables[:2]:
|
|
|
|
editor.execute(editor.sql_delete_table % {"table": table})
|
|
|
|
for table in tables:
|
|
|
|
self.assertTableNotExists(table)
|
|
|
|
|
2015-01-12 03:13:31 +08:00
|
|
|
@override_settings(
|
|
|
|
INSTALLED_APPS=[
|
|
|
|
"migrations.migrations_test_apps.lookuperror_a",
|
|
|
|
"migrations.migrations_test_apps.lookuperror_b",
|
|
|
|
"migrations.migrations_test_apps.lookuperror_c"
|
|
|
|
]
|
|
|
|
)
|
|
|
|
def test_unrelated_model_lookups_forwards(self):
|
|
|
|
"""
|
2016-10-27 15:53:39 +08:00
|
|
|
#24123 - All models of apps already applied which are
|
2015-01-12 03:13:31 +08:00
|
|
|
unrelated to the first app being applied are part of the initial model
|
|
|
|
state.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
self.assertTableNotExists("lookuperror_a_a1")
|
|
|
|
self.assertTableNotExists("lookuperror_b_b1")
|
|
|
|
self.assertTableNotExists("lookuperror_c_c1")
|
|
|
|
executor.migrate([("lookuperror_b", "0003_b3")])
|
|
|
|
self.assertTableExists("lookuperror_b_b3")
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
|
|
|
|
# Migrate forwards -- This led to a lookup LookupErrors because
|
|
|
|
# lookuperror_b.B2 is already applied
|
|
|
|
executor.migrate([
|
|
|
|
("lookuperror_a", "0004_a4"),
|
|
|
|
("lookuperror_c", "0003_c3"),
|
|
|
|
])
|
|
|
|
self.assertTableExists("lookuperror_a_a4")
|
|
|
|
self.assertTableExists("lookuperror_c_c3")
|
|
|
|
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
finally:
|
|
|
|
# Cleanup
|
|
|
|
executor.migrate([
|
|
|
|
("lookuperror_a", None),
|
|
|
|
("lookuperror_b", None),
|
|
|
|
("lookuperror_c", None),
|
|
|
|
])
|
|
|
|
self.assertTableNotExists("lookuperror_a_a1")
|
|
|
|
self.assertTableNotExists("lookuperror_b_b1")
|
|
|
|
self.assertTableNotExists("lookuperror_c_c1")
|
|
|
|
|
|
|
|
@override_settings(
|
|
|
|
INSTALLED_APPS=[
|
|
|
|
"migrations.migrations_test_apps.lookuperror_a",
|
|
|
|
"migrations.migrations_test_apps.lookuperror_b",
|
|
|
|
"migrations.migrations_test_apps.lookuperror_c"
|
|
|
|
]
|
|
|
|
)
|
|
|
|
def test_unrelated_model_lookups_backwards(self):
|
|
|
|
"""
|
2016-10-27 15:53:39 +08:00
|
|
|
#24123 - All models of apps being unapplied which are
|
2015-01-12 03:13:31 +08:00
|
|
|
unrelated to the first app being unapplied are part of the initial
|
|
|
|
model state.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
self.assertTableNotExists("lookuperror_a_a1")
|
|
|
|
self.assertTableNotExists("lookuperror_b_b1")
|
|
|
|
self.assertTableNotExists("lookuperror_c_c1")
|
|
|
|
executor.migrate([
|
|
|
|
("lookuperror_a", "0004_a4"),
|
|
|
|
("lookuperror_b", "0003_b3"),
|
|
|
|
("lookuperror_c", "0003_c3"),
|
|
|
|
])
|
|
|
|
self.assertTableExists("lookuperror_b_b3")
|
|
|
|
self.assertTableExists("lookuperror_a_a4")
|
|
|
|
self.assertTableExists("lookuperror_c_c3")
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
|
|
|
|
# Migrate backwards -- This led to a lookup LookupErrors because
|
|
|
|
# lookuperror_b.B2 is not in the initial state (unrelated to app c)
|
|
|
|
executor.migrate([("lookuperror_a", None)])
|
|
|
|
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
finally:
|
|
|
|
# Cleanup
|
|
|
|
executor.migrate([
|
|
|
|
("lookuperror_b", None),
|
|
|
|
("lookuperror_c", None)
|
|
|
|
])
|
|
|
|
self.assertTableNotExists("lookuperror_a_a1")
|
|
|
|
self.assertTableNotExists("lookuperror_b_b1")
|
|
|
|
self.assertTableNotExists("lookuperror_c_c1")
|
|
|
|
|
2016-05-23 08:35:10 +08:00
|
|
|
@override_settings(
|
|
|
|
INSTALLED_APPS=[
|
|
|
|
'migrations.migrations_test_apps.mutate_state_a',
|
|
|
|
'migrations.migrations_test_apps.mutate_state_b',
|
|
|
|
]
|
|
|
|
)
|
|
|
|
def test_unrelated_applied_migrations_mutate_state(self):
|
|
|
|
"""
|
|
|
|
#26647 - Unrelated applied migrations should be part of the final
|
|
|
|
state in both directions.
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
executor.migrate([
|
|
|
|
('mutate_state_b', '0002_add_field'),
|
|
|
|
])
|
|
|
|
# Migrate forward.
|
|
|
|
executor.loader.build_graph()
|
|
|
|
state = executor.migrate([
|
|
|
|
('mutate_state_a', '0001_initial'),
|
|
|
|
])
|
2020-04-22 11:14:34 +08:00
|
|
|
self.assertIn('added', state.models['mutate_state_b', 'b'].fields)
|
2016-05-23 08:35:10 +08:00
|
|
|
executor.loader.build_graph()
|
|
|
|
# Migrate backward.
|
|
|
|
state = executor.migrate([
|
|
|
|
('mutate_state_a', None),
|
|
|
|
])
|
2020-04-22 11:14:34 +08:00
|
|
|
self.assertIn('added', state.models['mutate_state_b', 'b'].fields)
|
2017-01-09 21:02:09 +08:00
|
|
|
executor.migrate([
|
|
|
|
('mutate_state_b', None),
|
|
|
|
])
|
2016-05-23 08:35:10 +08:00
|
|
|
|
2015-01-12 04:07:45 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations"})
|
|
|
|
def test_process_callback(self):
|
|
|
|
"""
|
|
|
|
#24129 - Tests callback process
|
|
|
|
"""
|
|
|
|
call_args_list = []
|
|
|
|
|
|
|
|
def callback(*args):
|
|
|
|
call_args_list.append(args)
|
|
|
|
|
|
|
|
executor = MigrationExecutor(connection, progress_callback=callback)
|
|
|
|
# Were the tables there before?
|
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_tribble")
|
|
|
|
executor.migrate([
|
|
|
|
("migrations", "0001_initial"),
|
|
|
|
("migrations", "0002_second"),
|
|
|
|
])
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
|
|
|
|
executor.migrate([
|
|
|
|
("migrations", None),
|
|
|
|
("migrations", None),
|
|
|
|
])
|
|
|
|
self.assertTableNotExists("migrations_author")
|
|
|
|
self.assertTableNotExists("migrations_tribble")
|
|
|
|
|
|
|
|
migrations = executor.loader.graph.nodes
|
|
|
|
expected = [
|
2017-12-29 04:07:29 +08:00
|
|
|
("render_start",),
|
|
|
|
("render_success",),
|
2015-01-12 04:07:45 +08:00
|
|
|
("apply_start", migrations['migrations', '0001_initial'], False),
|
|
|
|
("apply_success", migrations['migrations', '0001_initial'], False),
|
|
|
|
("apply_start", migrations['migrations', '0002_second'], False),
|
|
|
|
("apply_success", migrations['migrations', '0002_second'], False),
|
2017-12-29 04:07:29 +08:00
|
|
|
("render_start",),
|
|
|
|
("render_success",),
|
2015-01-12 04:07:45 +08:00
|
|
|
("unapply_start", migrations['migrations', '0002_second'], False),
|
|
|
|
("unapply_success", migrations['migrations', '0002_second'], False),
|
|
|
|
("unapply_start", migrations['migrations', '0001_initial'], False),
|
|
|
|
("unapply_success", migrations['migrations', '0001_initial'], False),
|
|
|
|
]
|
|
|
|
self.assertEqual(call_args_list, expected)
|
|
|
|
|
2015-02-11 17:14:53 +08:00
|
|
|
@override_settings(
|
|
|
|
INSTALLED_APPS=[
|
|
|
|
"migrations.migrations_test_apps.alter_fk.author_app",
|
|
|
|
"migrations.migrations_test_apps.alter_fk.book_app",
|
|
|
|
]
|
|
|
|
)
|
|
|
|
def test_alter_id_type_with_fk(self):
|
|
|
|
try:
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
self.assertTableNotExists("author_app_author")
|
|
|
|
self.assertTableNotExists("book_app_book")
|
|
|
|
# Apply initial migrations
|
|
|
|
executor.migrate([
|
|
|
|
("author_app", "0001_initial"),
|
|
|
|
("book_app", "0001_initial"),
|
|
|
|
])
|
|
|
|
self.assertTableExists("author_app_author")
|
|
|
|
self.assertTableExists("book_app_book")
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
|
|
|
|
# Apply PK type alteration
|
|
|
|
executor.migrate([("author_app", "0002_alter_id")])
|
|
|
|
|
|
|
|
# Rebuild the graph to reflect the new DB state
|
|
|
|
executor.loader.build_graph()
|
|
|
|
finally:
|
|
|
|
# We can't simply unapply the migrations here because there is no
|
|
|
|
# implicit cast from VARCHAR to INT on the database level.
|
|
|
|
with connection.schema_editor() as editor:
|
|
|
|
editor.execute(editor.sql_delete_table % {"table": "book_app_book"})
|
|
|
|
editor.execute(editor.sql_delete_table % {"table": "author_app_author"})
|
|
|
|
self.assertTableNotExists("author_app_author")
|
|
|
|
self.assertTableNotExists("book_app_book")
|
2017-01-09 21:02:09 +08:00
|
|
|
executor.migrate([("author_app", None)], fake=True)
|
2015-02-11 17:14:53 +08:00
|
|
|
|
2015-06-03 04:23:07 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations_squashed"})
|
|
|
|
def test_apply_all_replaced_marks_replacement_as_applied(self):
|
|
|
|
"""
|
2015-06-04 09:20:12 +08:00
|
|
|
Applying all replaced migrations marks replacement as applied (#24628).
|
2015-06-03 04:23:07 +08:00
|
|
|
"""
|
|
|
|
recorder = MigrationRecorder(connection)
|
|
|
|
# Place the database in a state where the replaced migrations are
|
|
|
|
# partially applied: 0001 is applied, 0002 is not.
|
|
|
|
recorder.record_applied("migrations", "0001_initial")
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
# Use fake because we don't actually have the first migration
|
|
|
|
# applied, so the second will fail. And there's no need to actually
|
|
|
|
# create/modify tables here, we're just testing the
|
|
|
|
# MigrationRecord, which works the same with or without fake.
|
|
|
|
executor.migrate([("migrations", "0002_second")], fake=True)
|
|
|
|
|
|
|
|
# Because we've now applied 0001 and 0002 both, their squashed
|
|
|
|
# replacement should be marked as applied.
|
|
|
|
self.assertIn(
|
|
|
|
("migrations", "0001_squashed_0002"),
|
|
|
|
recorder.applied_migrations(),
|
|
|
|
)
|
|
|
|
|
2015-06-04 03:46:01 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={"migrations": "migrations.test_migrations_squashed"})
|
|
|
|
def test_migrate_marks_replacement_applied_even_if_it_did_nothing(self):
|
|
|
|
"""
|
|
|
|
A new squash migration will be marked as applied even if all its
|
2015-06-04 09:20:12 +08:00
|
|
|
replaced migrations were previously already applied (#24628).
|
2015-06-04 03:46:01 +08:00
|
|
|
"""
|
|
|
|
recorder = MigrationRecorder(connection)
|
|
|
|
# Record all replaced migrations as applied
|
|
|
|
recorder.record_applied("migrations", "0001_initial")
|
|
|
|
recorder.record_applied("migrations", "0002_second")
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
executor.migrate([("migrations", "0001_squashed_0002")])
|
|
|
|
|
|
|
|
# Because 0001 and 0002 are both applied, even though this migrate run
|
|
|
|
# didn't apply anything new, their squashed replacement should be
|
|
|
|
# marked as applied.
|
|
|
|
self.assertIn(
|
|
|
|
("migrations", "0001_squashed_0002"),
|
|
|
|
recorder.applied_migrations(),
|
|
|
|
)
|
|
|
|
|
2021-06-08 12:59:04 +08:00
|
|
|
@override_settings(MIGRATION_MODULES={'migrations': 'migrations.test_migrations_squashed'})
|
|
|
|
def test_migrate_marks_replacement_unapplied(self):
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
executor.migrate([('migrations', '0001_squashed_0002')])
|
|
|
|
try:
|
|
|
|
self.assertIn(
|
|
|
|
('migrations', '0001_squashed_0002'),
|
|
|
|
executor.recorder.applied_migrations(),
|
|
|
|
)
|
|
|
|
finally:
|
|
|
|
executor.loader.build_graph()
|
|
|
|
executor.migrate([('migrations', None)])
|
|
|
|
self.assertNotIn(
|
|
|
|
('migrations', '0001_squashed_0002'),
|
|
|
|
executor.recorder.applied_migrations(),
|
|
|
|
)
|
|
|
|
|
2018-10-25 02:02:33 +08:00
|
|
|
# When the feature is False, the operation and the record won't be
|
|
|
|
# performed in a transaction and the test will systematically pass.
|
|
|
|
@skipUnlessDBFeature('can_rollback_ddl')
|
|
|
|
def test_migrations_applied_and_recorded_atomically(self):
|
|
|
|
"""Migrations are applied and recorded atomically."""
|
2021-01-21 09:59:20 +08:00
|
|
|
class Migration(migrations.Migration):
|
|
|
|
operations = [
|
|
|
|
migrations.CreateModel('model', [
|
|
|
|
('id', models.AutoField(primary_key=True)),
|
|
|
|
]),
|
|
|
|
]
|
|
|
|
|
2018-10-25 02:02:33 +08:00
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
with mock.patch('django.db.migrations.executor.MigrationExecutor.record_migration') as record_migration:
|
|
|
|
record_migration.side_effect = RuntimeError('Recording migration failed.')
|
|
|
|
with self.assertRaisesMessage(RuntimeError, 'Recording migration failed.'):
|
2021-01-21 09:59:20 +08:00
|
|
|
executor.apply_migration(
|
|
|
|
ProjectState(),
|
|
|
|
Migration('0001_initial', 'record_migration'),
|
|
|
|
)
|
2018-10-25 02:02:33 +08:00
|
|
|
executor.migrate([('migrations', '0001_initial')])
|
|
|
|
# The migration isn't recorded as applied since it failed.
|
|
|
|
migration_recorder = MigrationRecorder(connection)
|
2021-01-21 09:59:20 +08:00
|
|
|
self.assertIs(
|
|
|
|
migration_recorder.migration_qs.filter(
|
|
|
|
app='record_migration', name='0001_initial',
|
|
|
|
).exists(),
|
|
|
|
False,
|
|
|
|
)
|
|
|
|
self.assertTableNotExists('record_migration_model')
|
2018-10-25 02:02:33 +08:00
|
|
|
|
2021-01-21 08:00:36 +08:00
|
|
|
def test_migrations_not_applied_on_deferred_sql_failure(self):
|
|
|
|
"""Migrations are not recorded if deferred SQL application fails."""
|
|
|
|
class DeferredSQL:
|
|
|
|
def __str__(self):
|
|
|
|
raise DatabaseError('Failed to apply deferred SQL')
|
|
|
|
|
|
|
|
class Migration(migrations.Migration):
|
|
|
|
atomic = False
|
|
|
|
|
|
|
|
def apply(self, project_state, schema_editor, collect_sql=False):
|
|
|
|
schema_editor.deferred_sql.append(DeferredSQL())
|
|
|
|
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
with self.assertRaisesMessage(DatabaseError, 'Failed to apply deferred SQL'):
|
|
|
|
executor.apply_migration(
|
|
|
|
ProjectState(),
|
|
|
|
Migration('0001_initial', 'deferred_sql'),
|
|
|
|
)
|
|
|
|
# The migration isn't recorded as applied since it failed.
|
|
|
|
migration_recorder = MigrationRecorder(connection)
|
|
|
|
self.assertIs(
|
|
|
|
migration_recorder.migration_qs.filter(
|
|
|
|
app='deferred_sql', name='0001_initial',
|
|
|
|
).exists(),
|
|
|
|
False,
|
|
|
|
)
|
|
|
|
|
2021-12-28 23:46:42 +08:00
|
|
|
@mock.patch.object(MigrationRecorder, 'has_table', return_value=False)
|
|
|
|
def test_migrate_skips_schema_creation(self, mocked_has_table):
|
|
|
|
"""
|
|
|
|
The django_migrations table is not created if there are no migrations
|
|
|
|
to record.
|
|
|
|
"""
|
|
|
|
executor = MigrationExecutor(connection)
|
|
|
|
# 0 queries, since the query for has_table is being mocked.
|
|
|
|
with self.assertNumQueries(0):
|
|
|
|
executor.migrate([], plan=[])
|
|
|
|
|
2014-11-18 01:13:47 +08:00
|
|
|
|
2017-01-19 15:39:46 +08:00
|
|
|
class FakeLoader:
|
2014-11-18 01:13:47 +08:00
|
|
|
def __init__(self, graph, applied):
|
|
|
|
self.graph = graph
|
|
|
|
self.applied_migrations = applied
|
2021-06-05 13:06:17 +08:00
|
|
|
self.replace_migrations = True
|
2014-11-18 01:13:47 +08:00
|
|
|
|
|
|
|
|
2017-01-19 15:39:46 +08:00
|
|
|
class FakeMigration:
|
2014-11-18 01:13:47 +08:00
|
|
|
"""Really all we need is any object with a debug-useful repr."""
|
|
|
|
def __init__(self, name):
|
|
|
|
self.name = name
|
|
|
|
|
|
|
|
def __repr__(self):
|
|
|
|
return 'M<%s>' % self.name
|
|
|
|
|
|
|
|
|
2018-11-27 03:05:02 +08:00
|
|
|
class ExecutorUnitTests(SimpleTestCase):
|
2014-11-18 01:13:47 +08:00
|
|
|
"""(More) isolated unit tests for executor methods."""
|
|
|
|
def test_minimize_rollbacks(self):
|
|
|
|
"""
|
|
|
|
Minimize unnecessary rollbacks in connected apps.
|
|
|
|
|
|
|
|
When you say "./manage.py migrate appA 0001", rather than migrating to
|
|
|
|
just after appA-0001 in the linearized migration plan (which could roll
|
|
|
|
back migrations in other apps that depend on appA 0001, but don't need
|
|
|
|
to be rolled back since we're not rolling back appA 0001), we migrate
|
|
|
|
to just before appA-0002.
|
|
|
|
"""
|
|
|
|
a1_impl = FakeMigration('a1')
|
|
|
|
a1 = ('a', '1')
|
|
|
|
a2_impl = FakeMigration('a2')
|
|
|
|
a2 = ('a', '2')
|
|
|
|
b1_impl = FakeMigration('b1')
|
|
|
|
b1 = ('b', '1')
|
|
|
|
graph = MigrationGraph()
|
|
|
|
graph.add_node(a1, a1_impl)
|
|
|
|
graph.add_node(a2, a2_impl)
|
|
|
|
graph.add_node(b1, b1_impl)
|
|
|
|
graph.add_dependency(None, b1, a1)
|
|
|
|
graph.add_dependency(None, a2, a1)
|
|
|
|
|
|
|
|
executor = MigrationExecutor(None)
|
2019-03-08 08:36:55 +08:00
|
|
|
executor.loader = FakeLoader(graph, {
|
|
|
|
a1: a1_impl,
|
|
|
|
b1: b1_impl,
|
|
|
|
a2: a2_impl,
|
|
|
|
})
|
2014-11-18 01:13:47 +08:00
|
|
|
|
|
|
|
plan = executor.migration_plan({a1})
|
|
|
|
|
|
|
|
self.assertEqual(plan, [(a2_impl, True)])
|
|
|
|
|
|
|
|
def test_minimize_rollbacks_branchy(self):
|
2016-09-17 00:15:00 +08:00
|
|
|
r"""
|
2014-11-18 01:13:47 +08:00
|
|
|
Minimize rollbacks when target has multiple in-app children.
|
|
|
|
|
|
|
|
a: 1 <---- 3 <--\
|
|
|
|
\ \- 2 <--- 4
|
|
|
|
\ \
|
|
|
|
b: \- 1 <--- 2
|
|
|
|
"""
|
|
|
|
a1_impl = FakeMigration('a1')
|
|
|
|
a1 = ('a', '1')
|
|
|
|
a2_impl = FakeMigration('a2')
|
|
|
|
a2 = ('a', '2')
|
|
|
|
a3_impl = FakeMigration('a3')
|
|
|
|
a3 = ('a', '3')
|
|
|
|
a4_impl = FakeMigration('a4')
|
|
|
|
a4 = ('a', '4')
|
|
|
|
b1_impl = FakeMigration('b1')
|
|
|
|
b1 = ('b', '1')
|
|
|
|
b2_impl = FakeMigration('b2')
|
|
|
|
b2 = ('b', '2')
|
|
|
|
graph = MigrationGraph()
|
|
|
|
graph.add_node(a1, a1_impl)
|
|
|
|
graph.add_node(a2, a2_impl)
|
|
|
|
graph.add_node(a3, a3_impl)
|
|
|
|
graph.add_node(a4, a4_impl)
|
|
|
|
graph.add_node(b1, b1_impl)
|
|
|
|
graph.add_node(b2, b2_impl)
|
|
|
|
graph.add_dependency(None, a2, a1)
|
|
|
|
graph.add_dependency(None, a3, a1)
|
|
|
|
graph.add_dependency(None, a4, a2)
|
|
|
|
graph.add_dependency(None, a4, a3)
|
|
|
|
graph.add_dependency(None, b2, b1)
|
|
|
|
graph.add_dependency(None, b1, a1)
|
|
|
|
graph.add_dependency(None, b2, a2)
|
|
|
|
|
|
|
|
executor = MigrationExecutor(None)
|
2019-03-08 08:36:55 +08:00
|
|
|
executor.loader = FakeLoader(graph, {
|
|
|
|
a1: a1_impl,
|
|
|
|
b1: b1_impl,
|
|
|
|
a2: a2_impl,
|
|
|
|
b2: b2_impl,
|
|
|
|
a3: a3_impl,
|
|
|
|
a4: a4_impl,
|
|
|
|
})
|
2014-11-18 01:13:47 +08:00
|
|
|
|
|
|
|
plan = executor.migration_plan({a1})
|
|
|
|
|
|
|
|
should_be_rolled_back = [b2_impl, a4_impl, a2_impl, a3_impl]
|
|
|
|
exp = [(m, True) for m in should_be_rolled_back]
|
|
|
|
self.assertEqual(plan, exp)
|
2014-11-20 10:43:12 +08:00
|
|
|
|
|
|
|
def test_backwards_nothing_to_do(self):
|
2016-09-17 00:15:00 +08:00
|
|
|
r"""
|
2014-11-20 10:43:12 +08:00
|
|
|
If the current state satisfies the given target, do nothing.
|
|
|
|
|
|
|
|
a: 1 <--- 2
|
|
|
|
b: \- 1
|
|
|
|
c: \- 1
|
|
|
|
|
|
|
|
If a1 is applied already and a2 is not, and we're asked to migrate to
|
|
|
|
a1, don't apply or unapply b1 or c1, regardless of their current state.
|
|
|
|
"""
|
|
|
|
a1_impl = FakeMigration('a1')
|
|
|
|
a1 = ('a', '1')
|
|
|
|
a2_impl = FakeMigration('a2')
|
|
|
|
a2 = ('a', '2')
|
|
|
|
b1_impl = FakeMigration('b1')
|
|
|
|
b1 = ('b', '1')
|
|
|
|
c1_impl = FakeMigration('c1')
|
|
|
|
c1 = ('c', '1')
|
|
|
|
graph = MigrationGraph()
|
|
|
|
graph.add_node(a1, a1_impl)
|
|
|
|
graph.add_node(a2, a2_impl)
|
|
|
|
graph.add_node(b1, b1_impl)
|
|
|
|
graph.add_node(c1, c1_impl)
|
|
|
|
graph.add_dependency(None, a2, a1)
|
|
|
|
graph.add_dependency(None, b1, a1)
|
|
|
|
graph.add_dependency(None, c1, a1)
|
|
|
|
|
|
|
|
executor = MigrationExecutor(None)
|
2019-03-08 08:36:55 +08:00
|
|
|
executor.loader = FakeLoader(graph, {
|
|
|
|
a1: a1_impl,
|
|
|
|
b1: b1_impl,
|
|
|
|
})
|
2014-11-20 10:43:12 +08:00
|
|
|
|
|
|
|
plan = executor.migration_plan({a1})
|
|
|
|
|
|
|
|
self.assertEqual(plan, [])
|