You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
bazarr/libs/alembic/script/base.py

1061 lines
37 KiB

from __future__ import annotations
from contextlib import contextmanager
import datetime
import os
import re
import shutil
import sys
from types import ModuleType
from typing import Any
from typing import cast
from typing import Iterator
from typing import List
from typing import Mapping
from typing import Optional
from typing import Sequence
from typing import Set
from typing import Tuple
from typing import TYPE_CHECKING
from typing import Union
from . import revision
from . import write_hooks
from .. import util
from ..runtime import migration
from ..util import compat
from ..util import not_none
if TYPE_CHECKING:
from .revision import _GetRevArg
from .revision import _RevIdType
from .revision import Revision
from ..config import Config
from ..config import MessagingOptions
from ..runtime.migration import RevisionStep
from ..runtime.migration import StampStep
try:
if compat.py39:
from zoneinfo import ZoneInfo
from zoneinfo import ZoneInfoNotFoundError
else:
from backports.zoneinfo import ZoneInfo # type: ignore[import-not-found,no-redef] # noqa: E501
from backports.zoneinfo import ZoneInfoNotFoundError # type: ignore[no-redef] # noqa: E501
except ImportError:
ZoneInfo = None # type: ignore[assignment, misc]
_sourceless_rev_file = re.compile(r"(?!\.\#|__init__)(.*\.py)(c|o)?$")
_only_source_rev_file = re.compile(r"(?!\.\#|__init__)(.*\.py)$")
_legacy_rev = re.compile(r"([a-f0-9]+)\.py$")
_slug_re = re.compile(r"\w+")
_default_file_template = "%(rev)s_%(slug)s"
_split_on_space_comma = re.compile(r", *|(?: +)")
_split_on_space_comma_colon = re.compile(r", *|(?: +)|\:")
class ScriptDirectory:
"""Provides operations upon an Alembic script directory.
This object is useful to get information as to current revisions,
most notably being able to get at the "head" revision, for schemes
that want to test if the current revision in the database is the most
recent::
from alembic.script import ScriptDirectory
from alembic.config import Config
config = Config()
config.set_main_option("script_location", "myapp:migrations")
script = ScriptDirectory.from_config(config)
head_revision = script.get_current_head()
"""
def __init__(
self,
dir: str, # noqa
file_template: str = _default_file_template,
truncate_slug_length: Optional[int] = 40,
version_locations: Optional[List[str]] = None,
sourceless: bool = False,
output_encoding: str = "utf-8",
timezone: Optional[str] = None,
hook_config: Optional[Mapping[str, str]] = None,
recursive_version_locations: bool = False,
messaging_opts: MessagingOptions = cast(
"MessagingOptions", util.EMPTY_DICT
),
) -> None:
self.dir = dir
self.file_template = file_template
self.version_locations = version_locations
self.truncate_slug_length = truncate_slug_length or 40
self.sourceless = sourceless
self.output_encoding = output_encoding
self.revision_map = revision.RevisionMap(self._load_revisions)
self.timezone = timezone
self.hook_config = hook_config
self.recursive_version_locations = recursive_version_locations
self.messaging_opts = messaging_opts
if not os.access(dir, os.F_OK):
raise util.CommandError(
"Path doesn't exist: %r. Please use "
"the 'init' command to create a new "
"scripts folder." % os.path.abspath(dir)
)
@property
def versions(self) -> str:
loc = self._version_locations
if len(loc) > 1:
raise util.CommandError("Multiple version_locations present")
else:
return loc[0]
@util.memoized_property
def _version_locations(self) -> Sequence[str]:
if self.version_locations:
return [
os.path.abspath(util.coerce_resource_to_filename(location))
for location in self.version_locations
]
else:
return (os.path.abspath(os.path.join(self.dir, "versions")),)
def _load_revisions(self) -> Iterator[Script]:
if self.version_locations:
paths = [
vers
for vers in self._version_locations
if os.path.exists(vers)
]
else:
paths = [self.versions]
dupes = set()
for vers in paths:
for file_path in Script._list_py_dir(self, vers):
real_path = os.path.realpath(file_path)
if real_path in dupes:
util.warn(
"File %s loaded twice! ignoring. Please ensure "
"version_locations is unique." % real_path
)
continue
dupes.add(real_path)
filename = os.path.basename(real_path)
dir_name = os.path.dirname(real_path)
script = Script._from_filename(self, dir_name, filename)
if script is None:
continue
yield script
@classmethod
def from_config(cls, config: Config) -> ScriptDirectory:
"""Produce a new :class:`.ScriptDirectory` given a :class:`.Config`
instance.
The :class:`.Config` need only have the ``script_location`` key
present.
"""
script_location = config.get_main_option("script_location")
if script_location is None:
raise util.CommandError(
"No 'script_location' key " "found in configuration."
)
truncate_slug_length: Optional[int]
tsl = config.get_main_option("truncate_slug_length")
if tsl is not None:
truncate_slug_length = int(tsl)
else:
truncate_slug_length = None
version_locations_str = config.get_main_option("version_locations")
version_locations: Optional[List[str]]
if version_locations_str:
version_path_separator = config.get_main_option(
"version_path_separator"
)
split_on_path = {
None: None,
"space": " ",
"os": os.pathsep,
":": ":",
";": ";",
}
try:
split_char: Optional[str] = split_on_path[
version_path_separator
]
except KeyError as ke:
raise ValueError(
"'%s' is not a valid value for "
"version_path_separator; "
"expected 'space', 'os', ':', ';'" % version_path_separator
) from ke
else:
if split_char is None:
# legacy behaviour for backwards compatibility
version_locations = _split_on_space_comma.split(
version_locations_str
)
else:
version_locations = [
x for x in version_locations_str.split(split_char) if x
]
else:
version_locations = None
prepend_sys_path = config.get_main_option("prepend_sys_path")
if prepend_sys_path:
sys.path[:0] = list(
_split_on_space_comma_colon.split(prepend_sys_path)
)
rvl = config.get_main_option("recursive_version_locations") == "true"
return ScriptDirectory(
util.coerce_resource_to_filename(script_location),
file_template=config.get_main_option(
"file_template", _default_file_template
),
truncate_slug_length=truncate_slug_length,
sourceless=config.get_main_option("sourceless") == "true",
output_encoding=config.get_main_option("output_encoding", "utf-8"),
version_locations=version_locations,
timezone=config.get_main_option("timezone"),
hook_config=config.get_section("post_write_hooks", {}),
recursive_version_locations=rvl,
messaging_opts=config.messaging_opts,
)
@contextmanager
def _catch_revision_errors(
self,
ancestor: Optional[str] = None,
multiple_heads: Optional[str] = None,
start: Optional[str] = None,
end: Optional[str] = None,
resolution: Optional[str] = None,
) -> Iterator[None]:
try:
yield
except revision.RangeNotAncestorError as rna:
if start is None:
start = cast(Any, rna.lower)
if end is None:
end = cast(Any, rna.upper)
if not ancestor:
ancestor = (
"Requested range %(start)s:%(end)s does not refer to "
"ancestor/descendant revisions along the same branch"
)
ancestor = ancestor % {"start": start, "end": end}
raise util.CommandError(ancestor) from rna
except revision.MultipleHeads as mh:
if not multiple_heads:
multiple_heads = (
"Multiple head revisions are present for given "
"argument '%(head_arg)s'; please "
"specify a specific target revision, "
"'<branchname>@%(head_arg)s' to "
"narrow to a specific head, or 'heads' for all heads"
)
multiple_heads = multiple_heads % {
"head_arg": end or mh.argument,
"heads": util.format_as_comma(mh.heads),
}
raise util.CommandError(multiple_heads) from mh
except revision.ResolutionError as re:
if resolution is None:
resolution = "Can't locate revision identified by '%s'" % (
re.argument
)
raise util.CommandError(resolution) from re
except revision.RevisionError as err:
raise util.CommandError(err.args[0]) from err
def walk_revisions(
self, base: str = "base", head: str = "heads"
) -> Iterator[Script]:
"""Iterate through all revisions.
:param base: the base revision, or "base" to start from the
empty revision.
:param head: the head revision; defaults to "heads" to indicate
all head revisions. May also be "head" to indicate a single
head revision.
"""
with self._catch_revision_errors(start=base, end=head):
for rev in self.revision_map.iterate_revisions(
head, base, inclusive=True, assert_relative_length=False
):
yield cast(Script, rev)
def get_revisions(self, id_: _GetRevArg) -> Tuple[Script, ...]:
"""Return the :class:`.Script` instance with the given rev identifier,
symbolic name, or sequence of identifiers.
"""
with self._catch_revision_errors():
return cast(
Tuple[Script, ...],
self.revision_map.get_revisions(id_),
)
def get_all_current(self, id_: Tuple[str, ...]) -> Set[Script]:
with self._catch_revision_errors():
return cast(Set[Script], self.revision_map._get_all_current(id_))
def get_revision(self, id_: str) -> Script:
"""Return the :class:`.Script` instance with the given rev id.
.. seealso::
:meth:`.ScriptDirectory.get_revisions`
"""
with self._catch_revision_errors():
return cast(Script, self.revision_map.get_revision(id_))
def as_revision_number(
self, id_: Optional[str]
) -> Optional[Union[str, Tuple[str, ...]]]:
"""Convert a symbolic revision, i.e. 'head' or 'base', into
an actual revision number."""
with self._catch_revision_errors():
rev, branch_name = self.revision_map._resolve_revision_number(id_)
if not rev:
# convert () to None
return None
elif id_ == "heads":
return rev
else:
return rev[0]
def iterate_revisions(
self,
upper: Union[str, Tuple[str, ...], None],
lower: Union[str, Tuple[str, ...], None],
**kw: Any,
) -> Iterator[Script]:
"""Iterate through script revisions, starting at the given
upper revision identifier and ending at the lower.
The traversal uses strictly the `down_revision`
marker inside each migration script, so
it is a requirement that upper >= lower,
else you'll get nothing back.
The iterator yields :class:`.Script` objects.
.. seealso::
:meth:`.RevisionMap.iterate_revisions`
"""
return cast(
Iterator[Script],
self.revision_map.iterate_revisions(upper, lower, **kw),
)
def get_current_head(self) -> Optional[str]:
"""Return the current head revision.
If the script directory has multiple heads
due to branching, an error is raised;
:meth:`.ScriptDirectory.get_heads` should be
preferred.
:return: a string revision number.
.. seealso::
:meth:`.ScriptDirectory.get_heads`
"""
with self._catch_revision_errors(
multiple_heads=(
"The script directory has multiple heads (due to branching)."
"Please use get_heads(), or merge the branches using "
"alembic merge."
)
):
return self.revision_map.get_current_head()
def get_heads(self) -> List[str]:
"""Return all "versioned head" revisions as strings.
This is normally a list of length one,
unless branches are present. The
:meth:`.ScriptDirectory.get_current_head()` method
can be used normally when a script directory
has only one head.
:return: a tuple of string revision numbers.
"""
return list(self.revision_map.heads)
def get_base(self) -> Optional[str]:
"""Return the "base" revision as a string.
This is the revision number of the script that
has a ``down_revision`` of None.
If the script directory has multiple bases, an error is raised;
:meth:`.ScriptDirectory.get_bases` should be
preferred.
"""
bases = self.get_bases()
if len(bases) > 1:
raise util.CommandError(
"The script directory has multiple bases. "
"Please use get_bases()."
)
elif bases:
return bases[0]
else:
return None
def get_bases(self) -> List[str]:
"""return all "base" revisions as strings.
This is the revision number of all scripts that
have a ``down_revision`` of None.
"""
return list(self.revision_map.bases)
def _upgrade_revs(
self, destination: str, current_rev: str
) -> List[RevisionStep]:
with self._catch_revision_errors(
ancestor="Destination %(end)s is not a valid upgrade "
"target from current head(s)",
end=destination,
):
revs = self.iterate_revisions(
destination, current_rev, implicit_base=True
)
return [
migration.MigrationStep.upgrade_from_script(
self.revision_map, script
)
for script in reversed(list(revs))
]
def _downgrade_revs(
self, destination: str, current_rev: Optional[str]
) -> List[RevisionStep]:
with self._catch_revision_errors(
ancestor="Destination %(end)s is not a valid downgrade "
"target from current head(s)",
end=destination,
):
revs = self.iterate_revisions(
current_rev, destination, select_for_downgrade=True
)
return [
migration.MigrationStep.downgrade_from_script(
self.revision_map, script
)
for script in revs
]
def _stamp_revs(
self, revision: _RevIdType, heads: _RevIdType
) -> List[StampStep]:
with self._catch_revision_errors(
multiple_heads="Multiple heads are present; please specify a "
"single target revision"
):
heads_revs = self.get_revisions(heads)
steps = []
if not revision:
revision = "base"
filtered_heads: List[Script] = []
for rev in util.to_tuple(revision):
if rev:
filtered_heads.extend(
self.revision_map.filter_for_lineage(
cast(Sequence[Script], heads_revs),
rev,
include_dependencies=True,
)
)
filtered_heads = util.unique_list(filtered_heads)
dests = self.get_revisions(revision) or [None]
for dest in dests:
if dest is None:
# dest is 'base'. Return a "delete branch" migration
# for all applicable heads.
steps.extend(
[
migration.StampStep(
head.revision,
None,
False,
True,
self.revision_map,
)
for head in filtered_heads
]
)
continue
elif dest in filtered_heads:
# the dest is already in the version table, do nothing.
continue
# figure out if the dest is a descendant or an
# ancestor of the selected nodes
descendants = set(
self.revision_map._get_descendant_nodes([dest])
)
ancestors = set(self.revision_map._get_ancestor_nodes([dest]))
if descendants.intersection(filtered_heads):
# heads are above the target, so this is a downgrade.
# we can treat them as a "merge", single step.
assert not ancestors.intersection(filtered_heads)
todo_heads = [head.revision for head in filtered_heads]
step = migration.StampStep(
todo_heads,
dest.revision,
False,
False,
self.revision_map,
)
steps.append(step)
continue
elif ancestors.intersection(filtered_heads):
# heads are below the target, so this is an upgrade.
# we can treat them as a "merge", single step.
todo_heads = [head.revision for head in filtered_heads]
step = migration.StampStep(
todo_heads,
dest.revision,
True,
False,
self.revision_map,
)
steps.append(step)
continue
else:
# destination is in a branch not represented,
# treat it as new branch
step = migration.StampStep(
(), dest.revision, True, True, self.revision_map
)
steps.append(step)
continue
return steps
def run_env(self) -> None:
"""Run the script environment.
This basically runs the ``env.py`` script present
in the migration environment. It is called exclusively
by the command functions in :mod:`alembic.command`.
"""
util.load_python_file(self.dir, "env.py")
@property
def env_py_location(self) -> str:
return os.path.abspath(os.path.join(self.dir, "env.py"))
def _generate_template(self, src: str, dest: str, **kw: Any) -> None:
with util.status(
f"Generating {os.path.abspath(dest)}", **self.messaging_opts
):
util.template_to_file(src, dest, self.output_encoding, **kw)
def _copy_file(self, src: str, dest: str) -> None:
with util.status(
f"Generating {os.path.abspath(dest)}", **self.messaging_opts
):
shutil.copy(src, dest)
def _ensure_directory(self, path: str) -> None:
path = os.path.abspath(path)
if not os.path.exists(path):
with util.status(
f"Creating directory {path}", **self.messaging_opts
):
os.makedirs(path)
def _generate_create_date(self) -> datetime.datetime:
if self.timezone is not None:
if ZoneInfo is None:
raise util.CommandError(
"Python >= 3.9 is required for timezone support or"
"the 'backports.zoneinfo' package must be installed."
)
# First, assume correct capitalization
try:
tzinfo = ZoneInfo(self.timezone)
except ZoneInfoNotFoundError:
tzinfo = None
if tzinfo is None:
try:
tzinfo = ZoneInfo(self.timezone.upper())
except ZoneInfoNotFoundError:
raise util.CommandError(
"Can't locate timezone: %s" % self.timezone
) from None
create_date = (
datetime.datetime.utcnow()
.replace(tzinfo=datetime.timezone.utc)
.astimezone(tzinfo)
)
else:
create_date = datetime.datetime.now()
return create_date
def generate_revision(
self,
revid: str,
message: Optional[str],
head: Optional[_RevIdType] = None,
splice: Optional[bool] = False,
branch_labels: Optional[_RevIdType] = None,
version_path: Optional[str] = None,
depends_on: Optional[_RevIdType] = None,
**kw: Any,
) -> Optional[Script]:
"""Generate a new revision file.
This runs the ``script.py.mako`` template, given
template arguments, and creates a new file.
:param revid: String revision id. Typically this
comes from ``alembic.util.rev_id()``.
:param message: the revision message, the one passed
by the -m argument to the ``revision`` command.
:param head: the head revision to generate against. Defaults
to the current "head" if no branches are present, else raises
an exception.
:param splice: if True, allow the "head" version to not be an
actual head; otherwise, the selected head must be a head
(e.g. endpoint) revision.
"""
if head is None:
head = "head"
try:
Script.verify_rev_id(revid)
except revision.RevisionError as err:
raise util.CommandError(err.args[0]) from err
with self._catch_revision_errors(
multiple_heads=(
"Multiple heads are present; please specify the head "
"revision on which the new revision should be based, "
"or perform a merge."
)
):
heads = cast(
Tuple[Optional["Revision"], ...],
self.revision_map.get_revisions(head),
)
for h in heads:
assert h != "base" # type: ignore[comparison-overlap]
if len(set(heads)) != len(heads):
raise util.CommandError("Duplicate head revisions specified")
create_date = self._generate_create_date()
if version_path is None:
if len(self._version_locations) > 1:
for head_ in heads:
if head_ is not None:
assert isinstance(head_, Script)
version_path = os.path.dirname(head_.path)
break
else:
raise util.CommandError(
"Multiple version locations present, "
"please specify --version-path"
)
else:
version_path = self.versions
norm_path = os.path.normpath(os.path.abspath(version_path))
for vers_path in self._version_locations:
if os.path.normpath(vers_path) == norm_path:
break
else:
raise util.CommandError(
"Path %s is not represented in current "
"version locations" % version_path
)
if self.version_locations:
self._ensure_directory(version_path)
path = self._rev_path(version_path, revid, message, create_date)
if not splice:
for head_ in heads:
if head_ is not None and not head_.is_head:
raise util.CommandError(
"Revision %s is not a head revision; please specify "
"--splice to create a new branch from this revision"
% head_.revision
)
resolved_depends_on: Optional[List[str]]
if depends_on:
with self._catch_revision_errors():
resolved_depends_on = [
dep
if dep in rev.branch_labels # maintain branch labels
else rev.revision # resolve partial revision identifiers
for rev, dep in [
(not_none(self.revision_map.get_revision(dep)), dep)
for dep in util.to_list(depends_on)
]
]
else:
resolved_depends_on = None
self._generate_template(
os.path.join(self.dir, "script.py.mako"),
path,
up_revision=str(revid),
down_revision=revision.tuple_rev_as_scalar(
tuple(h.revision if h is not None else None for h in heads)
),
branch_labels=util.to_tuple(branch_labels),
depends_on=revision.tuple_rev_as_scalar(resolved_depends_on),
create_date=create_date,
comma=util.format_as_comma,
message=message if message is not None else ("empty message"),
**kw,
)
post_write_hooks = self.hook_config
if post_write_hooks:
write_hooks._run_hooks(path, post_write_hooks)
try:
script = Script._from_path(self, path)
except revision.RevisionError as err:
raise util.CommandError(err.args[0]) from err
if script is None:
return None
if branch_labels and not script.branch_labels:
raise util.CommandError(
"Version %s specified branch_labels %s, however the "
"migration file %s does not have them; have you upgraded "
"your script.py.mako to include the "
"'branch_labels' section?"
% (script.revision, branch_labels, script.path)
)
self.revision_map.add_revision(script)
return script
def _rev_path(
self,
path: str,
rev_id: str,
message: Optional[str],
create_date: datetime.datetime,
) -> str:
epoch = int(create_date.timestamp())
slug = "_".join(_slug_re.findall(message or "")).lower()
if len(slug) > self.truncate_slug_length:
slug = slug[: self.truncate_slug_length].rsplit("_", 1)[0] + "_"
filename = "%s.py" % (
self.file_template
% {
"rev": rev_id,
"slug": slug,
"epoch": epoch,
"year": create_date.year,
"month": create_date.month,
"day": create_date.day,
"hour": create_date.hour,
"minute": create_date.minute,
"second": create_date.second,
}
)
return os.path.join(path, filename)
class Script(revision.Revision):
"""Represent a single revision file in a ``versions/`` directory.
The :class:`.Script` instance is returned by methods
such as :meth:`.ScriptDirectory.iterate_revisions`.
"""
def __init__(self, module: ModuleType, rev_id: str, path: str):
self.module = module
self.path = path
super().__init__(
rev_id,
module.down_revision,
branch_labels=util.to_tuple(
getattr(module, "branch_labels", None), default=()
),
dependencies=util.to_tuple(
getattr(module, "depends_on", None), default=()
),
)
module: ModuleType
"""The Python module representing the actual script itself."""
path: str
"""Filesystem path of the script."""
_db_current_indicator: Optional[bool] = None
"""Utility variable which when set will cause string output to indicate
this is a "current" version in some database"""
@property
def doc(self) -> str:
"""Return the docstring given in the script."""
return re.split("\n\n", self.longdoc)[0]
@property
def longdoc(self) -> str:
"""Return the docstring given in the script."""
doc = self.module.__doc__
if doc:
if hasattr(self.module, "_alembic_source_encoding"):
doc = doc.decode( # type: ignore[attr-defined]
self.module._alembic_source_encoding
)
return doc.strip() # type: ignore[union-attr]
else:
return ""
@property
def log_entry(self) -> str:
entry = "Rev: %s%s%s%s%s\n" % (
self.revision,
" (head)" if self.is_head else "",
" (branchpoint)" if self.is_branch_point else "",
" (mergepoint)" if self.is_merge_point else "",
" (current)" if self._db_current_indicator else "",
)
if self.is_merge_point:
entry += "Merges: %s\n" % (self._format_down_revision(),)
else:
entry += "Parent: %s\n" % (self._format_down_revision(),)
if self.dependencies:
entry += "Also depends on: %s\n" % (
util.format_as_comma(self.dependencies)
)
if self.is_branch_point:
entry += "Branches into: %s\n" % (
util.format_as_comma(self.nextrev)
)
if self.branch_labels:
entry += "Branch names: %s\n" % (
util.format_as_comma(self.branch_labels),
)
entry += "Path: %s\n" % (self.path,)
entry += "\n%s\n" % (
"\n".join(" %s" % para for para in self.longdoc.splitlines())
)
return entry
def __str__(self) -> str:
return "%s -> %s%s%s%s, %s" % (
self._format_down_revision(),
self.revision,
" (head)" if self.is_head else "",
" (branchpoint)" if self.is_branch_point else "",
" (mergepoint)" if self.is_merge_point else "",
self.doc,
)
def _head_only(
self,
include_branches: bool = False,
include_doc: bool = False,
include_parents: bool = False,
tree_indicators: bool = True,
head_indicators: bool = True,
) -> str:
text = self.revision
if include_parents:
if self.dependencies:
text = "%s (%s) -> %s" % (
self._format_down_revision(),
util.format_as_comma(self.dependencies),
text,
)
else:
text = "%s -> %s" % (self._format_down_revision(), text)
assert text is not None
if include_branches and self.branch_labels:
text += " (%s)" % util.format_as_comma(self.branch_labels)
if head_indicators or tree_indicators:
text += "%s%s%s" % (
" (head)" if self._is_real_head else "",
" (effective head)"
if self.is_head and not self._is_real_head
else "",
" (current)" if self._db_current_indicator else "",
)
if tree_indicators:
text += "%s%s" % (
" (branchpoint)" if self.is_branch_point else "",
" (mergepoint)" if self.is_merge_point else "",
)
if include_doc:
text += ", %s" % self.doc
return text
def cmd_format(
self,
verbose: bool,
include_branches: bool = False,
include_doc: bool = False,
include_parents: bool = False,
tree_indicators: bool = True,
) -> str:
if verbose:
return self.log_entry
else:
return self._head_only(
include_branches, include_doc, include_parents, tree_indicators
)
def _format_down_revision(self) -> str:
if not self.down_revision:
return "<base>"
else:
return util.format_as_comma(self._versioned_down_revisions)
@classmethod
def _from_path(
cls, scriptdir: ScriptDirectory, path: str
) -> Optional[Script]:
dir_, filename = os.path.split(path)
return cls._from_filename(scriptdir, dir_, filename)
@classmethod
def _list_py_dir(cls, scriptdir: ScriptDirectory, path: str) -> List[str]:
paths = []
for root, dirs, files in os.walk(path, topdown=True):
if root.endswith("__pycache__"):
# a special case - we may include these files
# if a `sourceless` option is specified
continue
for filename in sorted(files):
paths.append(os.path.join(root, filename))
if scriptdir.sourceless:
# look for __pycache__
py_cache_path = os.path.join(root, "__pycache__")
if os.path.exists(py_cache_path):
# add all files from __pycache__ whose filename is not
# already in the names we got from the version directory.
# add as relative paths including __pycache__ token
names = {filename.split(".")[0] for filename in files}
paths.extend(
os.path.join(py_cache_path, pyc)
for pyc in os.listdir(py_cache_path)
if pyc.split(".")[0] not in names
)
if not scriptdir.recursive_version_locations:
break
# the real script order is defined by revision,
# but it may be undefined if there are many files with a same
# `down_revision`, for a better user experience (ex. debugging),
# we use a deterministic order
dirs.sort()
return paths
@classmethod
def _from_filename(
cls, scriptdir: ScriptDirectory, dir_: str, filename: str
) -> Optional[Script]:
if scriptdir.sourceless:
py_match = _sourceless_rev_file.match(filename)
else:
py_match = _only_source_rev_file.match(filename)
if not py_match:
return None
py_filename = py_match.group(1)
if scriptdir.sourceless:
is_c = py_match.group(2) == "c"
is_o = py_match.group(2) == "o"
else:
is_c = is_o = False
if is_o or is_c:
py_exists = os.path.exists(os.path.join(dir_, py_filename))
pyc_exists = os.path.exists(os.path.join(dir_, py_filename + "c"))
# prefer .py over .pyc because we'd like to get the
# source encoding; prefer .pyc over .pyo because we'd like to
# have the docstrings which a -OO file would not have
if py_exists or is_o and pyc_exists:
return None
module = util.load_python_file(dir_, filename)
if not hasattr(module, "revision"):
# attempt to get the revision id from the script name,
# this for legacy only
m = _legacy_rev.match(filename)
if not m:
raise util.CommandError(
"Could not determine revision id from filename %s. "
"Be sure the 'revision' variable is "
"declared inside the script (please see 'Upgrading "
"from Alembic 0.1 to 0.2' in the documentation)."
% filename
)
else:
revision = m.group(1)
else:
revision = module.revision
return Script(module, revision, os.path.join(dir_, filename))