12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099 |
- import collections
- import json
- import re
- from functools import partial
- from itertools import chain
- from django.core.exceptions import EmptyResultSet, FieldError, FullResultSet
- from django.db import DatabaseError, NotSupportedError
- from django.db.models.constants import LOOKUP_SEP
- from django.db.models.expressions import F, OrderBy, RawSQL, Ref, Value
- from django.db.models.functions import Cast, Random
- from django.db.models.lookups import Lookup
- from django.db.models.query_utils import select_related_descend
- from django.db.models.sql.constants import (
- CURSOR,
- GET_ITERATOR_CHUNK_SIZE,
- MULTI,
- NO_RESULTS,
- ORDER_DIR,
- SINGLE,
- )
- from django.db.models.sql.query import Query, get_order_dir
- from django.db.models.sql.where import AND
- from django.db.transaction import TransactionManagementError
- from django.utils.functional import cached_property
- from django.utils.hashable import make_hashable
- from django.utils.regex_helper import _lazy_re_compile
- class PositionRef(Ref):
- def __init__(self, ordinal, refs, source):
- self.ordinal = ordinal
- super().__init__(refs, source)
- def as_sql(self, compiler, connection):
- return str(self.ordinal), ()
- class SQLCompiler:
- # Multiline ordering SQL clause may appear from RawSQL.
- ordering_parts = _lazy_re_compile(
- r"^(.*)\s(?:ASC|DESC).*",
- re.MULTILINE | re.DOTALL,
- )
- def __init__(self, query, connection, using, elide_empty=True):
- self.query = query
- self.connection = connection
- self.using = using
- # Some queries, e.g. coalesced aggregation, need to be executed even if
- # they would return an empty result set.
- self.elide_empty = elide_empty
- self.quote_cache = {"*": "*"}
- # The select, klass_info, and annotations are needed by QuerySet.iterator()
- # these are set as a side-effect of executing the query. Note that we calculate
- # separately a list of extra select columns needed for grammatical correctness
- # of the query, but these columns are not included in self.select.
- self.select = None
- self.annotation_col_map = None
- self.klass_info = None
- self._meta_ordering = None
- def __repr__(self):
- return (
- f"<{self.__class__.__qualname__} "
- f"model={self.query.model.__qualname__} "
- f"connection={self.connection!r} using={self.using!r}>"
- )
- def setup_query(self, with_col_aliases=False):
- if all(self.query.alias_refcount[a] == 0 for a in self.query.alias_map):
- self.query.get_initial_alias()
- self.select, self.klass_info, self.annotation_col_map = self.get_select(
- with_col_aliases=with_col_aliases,
- )
- self.col_count = len(self.select)
- def pre_sql_setup(self, with_col_aliases=False):
- """
- Do any necessary class setup immediately prior to producing SQL. This
- is for things that can't necessarily be done in __init__ because we
- might not have all the pieces in place at that time.
- """
- self.setup_query(with_col_aliases=with_col_aliases)
- order_by = self.get_order_by()
- self.where, self.having, self.qualify = self.query.where.split_having_qualify(
- must_group_by=self.query.group_by is not None
- )
- extra_select = self.get_extra_select(order_by, self.select)
- self.has_extra_select = bool(extra_select)
- group_by = self.get_group_by(self.select + extra_select, order_by)
- return extra_select, order_by, group_by
- def get_group_by(self, select, order_by):
- """
- Return a list of 2-tuples of form (sql, params).
- The logic of what exactly the GROUP BY clause contains is hard
- to describe in other words than "if it passes the test suite,
- then it is correct".
- """
- # Some examples:
- # SomeModel.objects.annotate(Count('somecol'))
- # GROUP BY: all fields of the model
- #
- # SomeModel.objects.values('name').annotate(Count('somecol'))
- # GROUP BY: name
- #
- # SomeModel.objects.annotate(Count('somecol')).values('name')
- # GROUP BY: all cols of the model
- #
- # SomeModel.objects.values('name', 'pk')
- # .annotate(Count('somecol')).values('pk')
- # GROUP BY: name, pk
- #
- # SomeModel.objects.values('name').annotate(Count('somecol')).values('pk')
- # GROUP BY: name, pk
- #
- # In fact, the self.query.group_by is the minimal set to GROUP BY. It
- # can't be ever restricted to a smaller set, but additional columns in
- # HAVING, ORDER BY, and SELECT clauses are added to it. Unfortunately
- # the end result is that it is impossible to force the query to have
- # a chosen GROUP BY clause - you can almost do this by using the form:
- # .values(*wanted_cols).annotate(AnAggregate())
- # but any later annotations, extra selects, values calls that
- # refer some column outside of the wanted_cols, order_by, or even
- # filter calls can alter the GROUP BY clause.
- # The query.group_by is either None (no GROUP BY at all), True
- # (group by select fields), or a list of expressions to be added
- # to the group by.
- if self.query.group_by is None:
- return []
- expressions = []
- group_by_refs = set()
- if self.query.group_by is not True:
- # If the group by is set to a list (by .values() call most likely),
- # then we need to add everything in it to the GROUP BY clause.
- # Backwards compatibility hack for setting query.group_by. Remove
- # when we have public API way of forcing the GROUP BY clause.
- # Converts string references to expressions.
- for expr in self.query.group_by:
- if not hasattr(expr, "as_sql"):
- expr = self.query.resolve_ref(expr)
- if isinstance(expr, Ref):
- if expr.refs not in group_by_refs:
- group_by_refs.add(expr.refs)
- expressions.append(expr.source)
- else:
- expressions.append(expr)
- # Note that even if the group_by is set, it is only the minimal
- # set to group by. So, we need to add cols in select, order_by, and
- # having into the select in any case.
- selected_expr_positions = {}
- for ordinal, (expr, _, alias) in enumerate(select, start=1):
- if alias:
- selected_expr_positions[expr] = ordinal
- # Skip members of the select clause that are already explicitly
- # grouped against.
- if alias in group_by_refs:
- continue
- expressions.extend(expr.get_group_by_cols())
- if not self._meta_ordering:
- for expr, (sql, params, is_ref) in order_by:
- # Skip references to the SELECT clause, as all expressions in
- # the SELECT clause are already part of the GROUP BY.
- if not is_ref:
- expressions.extend(expr.get_group_by_cols())
- having_group_by = self.having.get_group_by_cols() if self.having else ()
- for expr in having_group_by:
- expressions.append(expr)
- result = []
- seen = set()
- expressions = self.collapse_group_by(expressions, having_group_by)
- allows_group_by_select_index = (
- self.connection.features.allows_group_by_select_index
- )
- for expr in expressions:
- try:
- sql, params = self.compile(expr)
- except (EmptyResultSet, FullResultSet):
- continue
- if (
- allows_group_by_select_index
- and (position := selected_expr_positions.get(expr)) is not None
- ):
- sql, params = str(position), ()
- else:
- sql, params = expr.select_format(self, sql, params)
- params_hash = make_hashable(params)
- if (sql, params_hash) not in seen:
- result.append((sql, params))
- seen.add((sql, params_hash))
- return result
- def collapse_group_by(self, expressions, having):
- # If the database supports group by functional dependence reduction,
- # then the expressions can be reduced to the set of selected table
- # primary keys as all other columns are functionally dependent on them.
- if self.connection.features.allows_group_by_selected_pks:
- # Filter out all expressions associated with a table's primary key
- # present in the grouped columns. This is done by identifying all
- # tables that have their primary key included in the grouped
- # columns and removing non-primary key columns referring to them.
- # Unmanaged models are excluded because they could be representing
- # database views on which the optimization might not be allowed.
- pks = {
- expr
- for expr in expressions
- if (
- hasattr(expr, "target")
- and expr.target.primary_key
- and self.connection.features.allows_group_by_selected_pks_on_model(
- expr.target.model
- )
- )
- }
- aliases = {expr.alias for expr in pks}
- expressions = [
- expr
- for expr in expressions
- if expr in pks
- or expr in having
- or getattr(expr, "alias", None) not in aliases
- ]
- return expressions
- def get_select(self, with_col_aliases=False):
- """
- Return three values:
- - a list of 3-tuples of (expression, (sql, params), alias)
- - a klass_info structure,
- - a dictionary of annotations
- The (sql, params) is what the expression will produce, and alias is the
- "AS alias" for the column (possibly None).
- The klass_info structure contains the following information:
- - The base model of the query.
- - Which columns for that model are present in the query (by
- position of the select clause).
- - related_klass_infos: [f, klass_info] to descent into
- The annotations is a dictionary of {'attname': column position} values.
- """
- select = []
- klass_info = None
- annotations = {}
- select_idx = 0
- for alias, (sql, params) in self.query.extra_select.items():
- annotations[alias] = select_idx
- select.append((RawSQL(sql, params), alias))
- select_idx += 1
- assert not (self.query.select and self.query.default_cols)
- select_mask = self.query.get_select_mask()
- if self.query.default_cols:
- cols = self.get_default_columns(select_mask)
- else:
- # self.query.select is a special case. These columns never go to
- # any model.
- cols = self.query.select
- if cols:
- select_list = []
- for col in cols:
- select_list.append(select_idx)
- select.append((col, None))
- select_idx += 1
- klass_info = {
- "model": self.query.model,
- "select_fields": select_list,
- }
- for alias, annotation in self.query.annotation_select.items():
- annotations[alias] = select_idx
- select.append((annotation, alias))
- select_idx += 1
- if self.query.select_related:
- related_klass_infos = self.get_related_selections(select, select_mask)
- klass_info["related_klass_infos"] = related_klass_infos
- def get_select_from_parent(klass_info):
- for ki in klass_info["related_klass_infos"]:
- if ki["from_parent"]:
- ki["select_fields"] = (
- klass_info["select_fields"] + ki["select_fields"]
- )
- get_select_from_parent(ki)
- get_select_from_parent(klass_info)
- ret = []
- col_idx = 1
- for col, alias in select:
- try:
- sql, params = self.compile(col)
- except EmptyResultSet:
- empty_result_set_value = getattr(
- col, "empty_result_set_value", NotImplemented
- )
- if empty_result_set_value is NotImplemented:
- # Select a predicate that's always False.
- sql, params = "0", ()
- else:
- sql, params = self.compile(Value(empty_result_set_value))
- except FullResultSet:
- sql, params = self.compile(Value(True))
- else:
- sql, params = col.select_format(self, sql, params)
- if alias is None and with_col_aliases:
- alias = f"col{col_idx}"
- col_idx += 1
- ret.append((col, (sql, params), alias))
- return ret, klass_info, annotations
- def _order_by_pairs(self):
- if self.query.extra_order_by:
- ordering = self.query.extra_order_by
- elif not self.query.default_ordering:
- ordering = self.query.order_by
- elif self.query.order_by:
- ordering = self.query.order_by
- elif (meta := self.query.get_meta()) and meta.ordering:
- ordering = meta.ordering
- self._meta_ordering = ordering
- else:
- ordering = []
- if self.query.standard_ordering:
- default_order, _ = ORDER_DIR["ASC"]
- else:
- default_order, _ = ORDER_DIR["DESC"]
- selected_exprs = {}
- # Avoid computing `selected_exprs` if there is no `ordering` as it's
- # relatively expensive.
- if ordering and (select := self.select):
- for ordinal, (expr, _, alias) in enumerate(select, start=1):
- pos_expr = PositionRef(ordinal, alias, expr)
- if alias:
- selected_exprs[alias] = pos_expr
- selected_exprs[expr] = pos_expr
- for field in ordering:
- if hasattr(field, "resolve_expression"):
- if isinstance(field, Value):
- # output_field must be resolved for constants.
- field = Cast(field, field.output_field)
- if not isinstance(field, OrderBy):
- field = field.asc()
- if not self.query.standard_ordering:
- field = field.copy()
- field.reverse_ordering()
- select_ref = selected_exprs.get(field.expression)
- if select_ref or (
- isinstance(field.expression, F)
- and (select_ref := selected_exprs.get(field.expression.name))
- ):
- # Emulation of NULLS (FIRST|LAST) cannot be combined with
- # the usage of ordering by position.
- if (
- field.nulls_first is None and field.nulls_last is None
- ) or self.connection.features.supports_order_by_nulls_modifier:
- field = field.copy()
- field.expression = select_ref
- # Alias collisions are not possible when dealing with
- # combined queries so fallback to it if emulation of NULLS
- # handling is required.
- elif self.query.combinator:
- field = field.copy()
- field.expression = Ref(select_ref.refs, select_ref.source)
- yield field, select_ref is not None
- continue
- if field == "?": # random
- yield OrderBy(Random()), False
- continue
- col, order = get_order_dir(field, default_order)
- descending = order == "DESC"
- if select_ref := selected_exprs.get(col):
- # Reference to expression in SELECT clause
- yield (
- OrderBy(
- select_ref,
- descending=descending,
- ),
- True,
- )
- continue
- if col in self.query.annotations:
- # References to an expression which is masked out of the SELECT
- # clause.
- if self.query.combinator and self.select:
- # Don't use the resolved annotation because other
- # combinated queries might define it differently.
- expr = F(col)
- else:
- expr = self.query.annotations[col]
- if isinstance(expr, Value):
- # output_field must be resolved for constants.
- expr = Cast(expr, expr.output_field)
- yield OrderBy(expr, descending=descending), False
- continue
- if "." in field:
- # This came in through an extra(order_by=...) addition. Pass it
- # on verbatim.
- table, col = col.split(".", 1)
- yield (
- OrderBy(
- RawSQL(
- "%s.%s" % (self.quote_name_unless_alias(table), col), []
- ),
- descending=descending,
- ),
- False,
- )
- continue
- if self.query.extra and col in self.query.extra:
- if col in self.query.extra_select:
- yield (
- OrderBy(
- Ref(col, RawSQL(*self.query.extra[col])),
- descending=descending,
- ),
- True,
- )
- else:
- yield (
- OrderBy(RawSQL(*self.query.extra[col]), descending=descending),
- False,
- )
- else:
- if self.query.combinator and self.select:
- # Don't use the first model's field because other
- # combinated queries might define it differently.
- yield OrderBy(F(col), descending=descending), False
- else:
- # 'col' is of the form 'field' or 'field1__field2' or
- # '-field1__field2__field', etc.
- yield from self.find_ordering_name(
- field,
- self.query.get_meta(),
- default_order=default_order,
- )
- def get_order_by(self):
- """
- Return a list of 2-tuples of the form (expr, (sql, params, is_ref)) for
- the ORDER BY clause.
- The order_by clause can alter the select clause (for example it can add
- aliases to clauses that do not yet have one, or it can add totally new
- select clauses).
- """
- result = []
- seen = set()
- for expr, is_ref in self._order_by_pairs():
- resolved = expr.resolve_expression(self.query, allow_joins=True, reuse=None)
- if not is_ref and self.query.combinator and self.select:
- src = resolved.expression
- expr_src = expr.expression
- for sel_expr, _, col_alias in self.select:
- if src == sel_expr:
- # When values() is used the exact alias must be used to
- # reference annotations.
- if (
- self.query.has_select_fields
- and col_alias in self.query.annotation_select
- and not (
- isinstance(expr_src, F) and col_alias == expr_src.name
- )
- ):
- continue
- resolved.set_source_expressions(
- [Ref(col_alias if col_alias else src.target.column, src)]
- )
- break
- else:
- # Add column used in ORDER BY clause to the selected
- # columns and to each combined query.
- order_by_idx = len(self.query.select) + 1
- col_alias = f"__orderbycol{order_by_idx}"
- for q in self.query.combined_queries:
- # If fields were explicitly selected through values()
- # combined queries cannot be augmented.
- if q.has_select_fields:
- raise DatabaseError(
- "ORDER BY term does not match any column in "
- "the result set."
- )
- q.add_annotation(expr_src, col_alias)
- self.query.add_select_col(resolved, col_alias)
- resolved.set_source_expressions([Ref(col_alias, src)])
- sql, params = self.compile(resolved)
- # Don't add the same column twice, but the order direction is
- # not taken into account so we strip it. When this entire method
- # is refactored into expressions, then we can check each part as we
- # generate it.
- without_ordering = self.ordering_parts.search(sql)[1]
- params_hash = make_hashable(params)
- if (without_ordering, params_hash) in seen:
- continue
- seen.add((without_ordering, params_hash))
- result.append((resolved, (sql, params, is_ref)))
- return result
- def get_extra_select(self, order_by, select):
- extra_select = []
- if self.query.distinct and not self.query.distinct_fields:
- select_sql = [t[1] for t in select]
- for expr, (sql, params, is_ref) in order_by:
- without_ordering = self.ordering_parts.search(sql)[1]
- if not is_ref and (without_ordering, params) not in select_sql:
- extra_select.append((expr, (without_ordering, params), None))
- return extra_select
- def quote_name_unless_alias(self, name):
- """
- A wrapper around connection.ops.quote_name that doesn't quote aliases
- for table names. This avoids problems with some SQL dialects that treat
- quoted strings specially (e.g. PostgreSQL).
- """
- if name in self.quote_cache:
- return self.quote_cache[name]
- if (
- (name in self.query.alias_map and name not in self.query.table_map)
- or name in self.query.extra_select
- or (
- self.query.external_aliases.get(name)
- and name not in self.query.table_map
- )
- ):
- self.quote_cache[name] = name
- return name
- r = self.connection.ops.quote_name(name)
- self.quote_cache[name] = r
- return r
- def compile(self, node):
- vendor_impl = getattr(node, "as_" + self.connection.vendor, None)
- if vendor_impl:
- sql, params = vendor_impl(self, self.connection)
- else:
- sql, params = node.as_sql(self, self.connection)
- return sql, params
- def get_combinator_sql(self, combinator, all):
- features = self.connection.features
- compilers = [
- query.get_compiler(self.using, self.connection, self.elide_empty)
- for query in self.query.combined_queries
- ]
- if not features.supports_slicing_ordering_in_compound:
- for compiler in compilers:
- if compiler.query.is_sliced:
- raise DatabaseError(
- "LIMIT/OFFSET not allowed in subqueries of compound statements."
- )
- if compiler.get_order_by():
- raise DatabaseError(
- "ORDER BY not allowed in subqueries of compound statements."
- )
- elif self.query.is_sliced and combinator == "union":
- for compiler in compilers:
- # A sliced union cannot have its parts elided as some of them
- # might be sliced as well and in the event where only a single
- # part produces a non-empty resultset it might be impossible to
- # generate valid SQL.
- compiler.elide_empty = False
- parts = ()
- for compiler in compilers:
- try:
- # If the columns list is limited, then all combined queries
- # must have the same columns list. Set the selects defined on
- # the query on all combined queries, if not already set.
- if not compiler.query.values_select and self.query.values_select:
- compiler.query = compiler.query.clone()
- compiler.query.set_values(
- (
- *self.query.extra_select,
- *self.query.values_select,
- *self.query.annotation_select,
- )
- )
- part_sql, part_args = compiler.as_sql(with_col_aliases=True)
- if compiler.query.combinator:
- # Wrap in a subquery if wrapping in parentheses isn't
- # supported.
- if not features.supports_parentheses_in_compound:
- part_sql = "SELECT * FROM ({})".format(part_sql)
- # Add parentheses when combining with compound query if not
- # already added for all compound queries.
- elif (
- self.query.subquery
- or not features.supports_slicing_ordering_in_compound
- ):
- part_sql = "({})".format(part_sql)
- elif (
- self.query.subquery
- and features.supports_slicing_ordering_in_compound
- ):
- part_sql = "({})".format(part_sql)
- parts += ((part_sql, part_args),)
- except EmptyResultSet:
- # Omit the empty queryset with UNION and with DIFFERENCE if the
- # first queryset is nonempty.
- if combinator == "union" or (combinator == "difference" and parts):
- continue
- raise
- if not parts:
- raise EmptyResultSet
- combinator_sql = self.connection.ops.set_operators[combinator]
- if all and combinator == "union":
- combinator_sql += " ALL"
- braces = "{}"
- if not self.query.subquery and features.supports_slicing_ordering_in_compound:
- braces = "({})"
- sql_parts, args_parts = zip(
- *((braces.format(sql), args) for sql, args in parts)
- )
- result = [" {} ".format(combinator_sql).join(sql_parts)]
- params = []
- for part in args_parts:
- params.extend(part)
- return result, params
- def get_qualify_sql(self):
- where_parts = []
- if self.where:
- where_parts.append(self.where)
- if self.having:
- where_parts.append(self.having)
- inner_query = self.query.clone()
- inner_query.subquery = True
- inner_query.where = inner_query.where.__class__(where_parts)
- # Augment the inner query with any window function references that
- # might have been masked via values() and alias(). If any masked
- # aliases are added they'll be masked again to avoid fetching
- # the data in the `if qual_aliases` branch below.
- select = {
- expr: alias for expr, _, alias in self.get_select(with_col_aliases=True)[0]
- }
- select_aliases = set(select.values())
- qual_aliases = set()
- replacements = {}
- def collect_replacements(expressions):
- while expressions:
- expr = expressions.pop()
- if expr in replacements:
- continue
- elif select_alias := select.get(expr):
- replacements[expr] = select_alias
- elif isinstance(expr, Lookup):
- expressions.extend(expr.get_source_expressions())
- elif isinstance(expr, Ref):
- if expr.refs not in select_aliases:
- expressions.extend(expr.get_source_expressions())
- else:
- num_qual_alias = len(qual_aliases)
- select_alias = f"qual{num_qual_alias}"
- qual_aliases.add(select_alias)
- inner_query.add_annotation(expr, select_alias)
- replacements[expr] = select_alias
- collect_replacements(list(self.qualify.leaves()))
- self.qualify = self.qualify.replace_expressions(
- {expr: Ref(alias, expr) for expr, alias in replacements.items()}
- )
- order_by = []
- for order_by_expr, *_ in self.get_order_by():
- collect_replacements(order_by_expr.get_source_expressions())
- order_by.append(
- order_by_expr.replace_expressions(
- {expr: Ref(alias, expr) for expr, alias in replacements.items()}
- )
- )
- inner_query_compiler = inner_query.get_compiler(
- self.using, connection=self.connection, elide_empty=self.elide_empty
- )
- inner_sql, inner_params = inner_query_compiler.as_sql(
- # The limits must be applied to the outer query to avoid pruning
- # results too eagerly.
- with_limits=False,
- # Force unique aliasing of selected columns to avoid collisions
- # and make rhs predicates referencing easier.
- with_col_aliases=True,
- )
- qualify_sql, qualify_params = self.compile(self.qualify)
- result = [
- "SELECT * FROM (",
- inner_sql,
- ")",
- self.connection.ops.quote_name("qualify"),
- "WHERE",
- qualify_sql,
- ]
- if qual_aliases:
- # If some select aliases were unmasked for filtering purposes they
- # must be masked back.
- cols = [self.connection.ops.quote_name(alias) for alias in select.values()]
- result = [
- "SELECT",
- ", ".join(cols),
- "FROM (",
- *result,
- ")",
- self.connection.ops.quote_name("qualify_mask"),
- ]
- params = list(inner_params) + qualify_params
- # As the SQL spec is unclear on whether or not derived tables
- # ordering must propagate it has to be explicitly repeated on the
- # outer-most query to ensure it's preserved.
- if order_by:
- ordering_sqls = []
- for ordering in order_by:
- ordering_sql, ordering_params = self.compile(ordering)
- ordering_sqls.append(ordering_sql)
- params.extend(ordering_params)
- result.extend(["ORDER BY", ", ".join(ordering_sqls)])
- return result, params
- def as_sql(self, with_limits=True, with_col_aliases=False):
- """
- Create the SQL for this query. Return the SQL string and list of
- parameters.
- If 'with_limits' is False, any limit/offset information is not included
- in the query.
- """
- refcounts_before = self.query.alias_refcount.copy()
- try:
- combinator = self.query.combinator
- extra_select, order_by, group_by = self.pre_sql_setup(
- with_col_aliases=with_col_aliases or bool(combinator),
- )
- for_update_part = None
- # Is a LIMIT/OFFSET clause needed?
- with_limit_offset = with_limits and self.query.is_sliced
- combinator = self.query.combinator
- features = self.connection.features
- if combinator:
- if not getattr(features, "supports_select_{}".format(combinator)):
- raise NotSupportedError(
- "{} is not supported on this database backend.".format(
- combinator
- )
- )
- result, params = self.get_combinator_sql(
- combinator, self.query.combinator_all
- )
- elif self.qualify:
- result, params = self.get_qualify_sql()
- order_by = None
- else:
- distinct_fields, distinct_params = self.get_distinct()
- # This must come after 'select', 'ordering', and 'distinct'
- # (see docstring of get_from_clause() for details).
- from_, f_params = self.get_from_clause()
- try:
- where, w_params = (
- self.compile(self.where) if self.where is not None else ("", [])
- )
- except EmptyResultSet:
- if self.elide_empty:
- raise
- # Use a predicate that's always False.
- where, w_params = "0 = 1", []
- except FullResultSet:
- where, w_params = "", []
- try:
- having, h_params = (
- self.compile(self.having)
- if self.having is not None
- else ("", [])
- )
- except FullResultSet:
- having, h_params = "", []
- result = ["SELECT"]
- params = []
- if self.query.distinct:
- distinct_result, distinct_params = self.connection.ops.distinct_sql(
- distinct_fields,
- distinct_params,
- )
- result += distinct_result
- params += distinct_params
- out_cols = []
- for _, (s_sql, s_params), alias in self.select + extra_select:
- if alias:
- s_sql = "%s AS %s" % (
- s_sql,
- self.connection.ops.quote_name(alias),
- )
- params.extend(s_params)
- out_cols.append(s_sql)
- result += [", ".join(out_cols)]
- if from_:
- result += ["FROM", *from_]
- elif self.connection.features.bare_select_suffix:
- result += [self.connection.features.bare_select_suffix]
- params.extend(f_params)
- if self.query.select_for_update and features.has_select_for_update:
- if (
- self.connection.get_autocommit()
- # Don't raise an exception when database doesn't
- # support transactions, as it's a noop.
- and features.supports_transactions
- ):
- raise TransactionManagementError(
- "select_for_update cannot be used outside of a transaction."
- )
- if (
- with_limit_offset
- and not features.supports_select_for_update_with_limit
- ):
- raise NotSupportedError(
- "LIMIT/OFFSET is not supported with "
- "select_for_update on this database backend."
- )
- nowait = self.query.select_for_update_nowait
- skip_locked = self.query.select_for_update_skip_locked
- of = self.query.select_for_update_of
- no_key = self.query.select_for_no_key_update
- # If it's a NOWAIT/SKIP LOCKED/OF/NO KEY query but the
- # backend doesn't support it, raise NotSupportedError to
- # prevent a possible deadlock.
- if nowait and not features.has_select_for_update_nowait:
- raise NotSupportedError(
- "NOWAIT is not supported on this database backend."
- )
- elif skip_locked and not features.has_select_for_update_skip_locked:
- raise NotSupportedError(
- "SKIP LOCKED is not supported on this database backend."
- )
- elif of and not features.has_select_for_update_of:
- raise NotSupportedError(
- "FOR UPDATE OF is not supported on this database backend."
- )
- elif no_key and not features.has_select_for_no_key_update:
- raise NotSupportedError(
- "FOR NO KEY UPDATE is not supported on this "
- "database backend."
- )
- for_update_part = self.connection.ops.for_update_sql(
- nowait=nowait,
- skip_locked=skip_locked,
- of=self.get_select_for_update_of_arguments(),
- no_key=no_key,
- )
- if for_update_part and features.for_update_after_from:
- result.append(for_update_part)
- if where:
- result.append("WHERE %s" % where)
- params.extend(w_params)
- grouping = []
- for g_sql, g_params in group_by:
- grouping.append(g_sql)
- params.extend(g_params)
- if grouping:
- if distinct_fields:
- raise NotImplementedError(
- "annotate() + distinct(fields) is not implemented."
- )
- order_by = order_by or self.connection.ops.force_no_ordering()
- result.append("GROUP BY %s" % ", ".join(grouping))
- if self._meta_ordering:
- order_by = None
- if having:
- result.append("HAVING %s" % having)
- params.extend(h_params)
- if self.query.explain_info:
- result.insert(
- 0,
- self.connection.ops.explain_query_prefix(
- self.query.explain_info.format,
- **self.query.explain_info.options,
- ),
- )
- if order_by:
- ordering = []
- for _, (o_sql, o_params, _) in order_by:
- ordering.append(o_sql)
- params.extend(o_params)
- order_by_sql = "ORDER BY %s" % ", ".join(ordering)
- if combinator and features.requires_compound_order_by_subquery:
- result = ["SELECT * FROM (", *result, ")", order_by_sql]
- else:
- result.append(order_by_sql)
- if with_limit_offset:
- result.append(
- self.connection.ops.limit_offset_sql(
- self.query.low_mark, self.query.high_mark
- )
- )
- if for_update_part and not features.for_update_after_from:
- result.append(for_update_part)
- if self.query.subquery and extra_select:
- # If the query is used as a subquery, the extra selects would
- # result in more columns than the left-hand side expression is
- # expecting. This can happen when a subquery uses a combination
- # of order_by() and distinct(), forcing the ordering expressions
- # to be selected as well. Wrap the query in another subquery
- # to exclude extraneous selects.
- sub_selects = []
- sub_params = []
- for index, (select, _, alias) in enumerate(self.select, start=1):
- if alias:
- sub_selects.append(
- "%s.%s"
- % (
- self.connection.ops.quote_name("subquery"),
- self.connection.ops.quote_name(alias),
- )
- )
- else:
- select_clone = select.relabeled_clone(
- {select.alias: "subquery"}
- )
- subselect, subparams = select_clone.as_sql(
- self, self.connection
- )
- sub_selects.append(subselect)
- sub_params.extend(subparams)
- return "SELECT %s FROM (%s) subquery" % (
- ", ".join(sub_selects),
- " ".join(result),
- ), tuple(sub_params + params)
- return " ".join(result), tuple(params)
- finally:
- # Finally do cleanup - get rid of the joins we created above.
- self.query.reset_refcounts(refcounts_before)
- def get_default_columns(
- self, select_mask, start_alias=None, opts=None, from_parent=None
- ):
- """
- Compute the default columns for selecting every field in the base
- model. Will sometimes be called to pull in related models (e.g. via
- select_related), in which case "opts" and "start_alias" will be given
- to provide a starting point for the traversal.
- Return a list of strings, quoted appropriately for use in SQL
- directly, as well as a set of aliases used in the select statement (if
- 'as_pairs' is True, return a list of (alias, col_name) pairs instead
- of strings as the first component and None as the second component).
- """
- result = []
- if opts is None:
- if (opts := self.query.get_meta()) is None:
- return result
- start_alias = start_alias or self.query.get_initial_alias()
- # The 'seen_models' is used to optimize checking the needed parent
- # alias for a given field. This also includes None -> start_alias to
- # be used by local fields.
- seen_models = {None: start_alias}
- for field in opts.concrete_fields:
- model = field.model._meta.concrete_model
- # A proxy model will have a different model and concrete_model. We
- # will assign None if the field belongs to this model.
- if model == opts.model:
- model = None
- if (
- from_parent
- and model is not None
- and issubclass(
- from_parent._meta.concrete_model, model._meta.concrete_model
- )
- ):
- # Avoid loading data for already loaded parents.
- # We end up here in the case select_related() resolution
- # proceeds from parent model to child model. In that case the
- # parent model data is already present in the SELECT clause,
- # and we want to avoid reloading the same data again.
- continue
- if select_mask and field not in select_mask:
- continue
- alias = self.query.join_parent_model(opts, model, start_alias, seen_models)
- column = field.get_col(alias)
- result.append(column)
- return result
- def get_distinct(self):
- """
- Return a quoted list of fields to use in DISTINCT ON part of the query.
- This method can alter the tables in the query, and thus it must be
- called before get_from_clause().
- """
- result = []
- params = []
- opts = self.query.get_meta()
- for name in self.query.distinct_fields:
- parts = name.split(LOOKUP_SEP)
- _, targets, alias, joins, path, _, transform_function = self._setup_joins(
- parts, opts, None
- )
- targets, alias, _ = self.query.trim_joins(targets, joins, path)
- for target in targets:
- if name in self.query.annotation_select:
- result.append(self.connection.ops.quote_name(name))
- else:
- r, p = self.compile(transform_function(target, alias))
- result.append(r)
- params.append(p)
- return result, params
- def find_ordering_name(
- self, name, opts, alias=None, default_order="ASC", already_seen=None
- ):
- """
- Return the table alias (the name might be ambiguous, the alias will
- not be) and column name for ordering by the given 'name' parameter.
- The 'name' is of the form 'field1__field2__...__fieldN'.
- """
- name, order = get_order_dir(name, default_order)
- descending = order == "DESC"
- pieces = name.split(LOOKUP_SEP)
- (
- field,
- targets,
- alias,
- joins,
- path,
- opts,
- transform_function,
- ) = self._setup_joins(pieces, opts, alias)
- # If we get to this point and the field is a relation to another model,
- # append the default ordering for that model unless it is the pk
- # shortcut or the attribute name of the field that is specified or
- # there are transforms to process.
- if (
- field.is_relation
- and opts.ordering
- and getattr(field, "attname", None) != pieces[-1]
- and name != "pk"
- and not getattr(transform_function, "has_transforms", False)
- ):
- # Firstly, avoid infinite loops.
- already_seen = already_seen or set()
- join_tuple = tuple(
- getattr(self.query.alias_map[j], "join_cols", None) for j in joins
- )
- if join_tuple in already_seen:
- raise FieldError("Infinite loop caused by ordering.")
- already_seen.add(join_tuple)
- results = []
- for item in opts.ordering:
- if hasattr(item, "resolve_expression") and not isinstance(
- item, OrderBy
- ):
- item = item.desc() if descending else item.asc()
- if isinstance(item, OrderBy):
- results.append(
- (item.prefix_references(f"{name}{LOOKUP_SEP}"), False)
- )
- continue
- results.extend(
- (expr.prefix_references(f"{name}{LOOKUP_SEP}"), is_ref)
- for expr, is_ref in self.find_ordering_name(
- item, opts, alias, order, already_seen
- )
- )
- return results
- targets, alias, _ = self.query.trim_joins(targets, joins, path)
- return [
- (OrderBy(transform_function(t, alias), descending=descending), False)
- for t in targets
- ]
- def _setup_joins(self, pieces, opts, alias):
- """
- Helper method for get_order_by() and get_distinct().
- get_ordering() and get_distinct() must produce same target columns on
- same input, as the prefixes of get_ordering() and get_distinct() must
- match. Executing SQL where this is not true is an error.
- """
- alias = alias or self.query.get_initial_alias()
- field, targets, opts, joins, path, transform_function = self.query.setup_joins(
- pieces, opts, alias
- )
- alias = joins[-1]
- return field, targets, alias, joins, path, opts, transform_function
- def get_from_clause(self):
- """
- Return a list of strings that are joined together to go after the
- "FROM" part of the query, as well as a list any extra parameters that
- need to be included. Subclasses, can override this to create a
- from-clause via a "select".
- This should only be called after any SQL construction methods that
- might change the tables that are needed. This means the select columns,
- ordering, and distinct must be done first.
- """
- result = []
- params = []
- for alias in tuple(self.query.alias_map):
- if not self.query.alias_refcount[alias]:
- continue
- try:
- from_clause = self.query.alias_map[alias]
- except KeyError:
- # Extra tables can end up in self.tables, but not in the
- # alias_map if they aren't in a join. That's OK. We skip them.
- continue
- clause_sql, clause_params = self.compile(from_clause)
- result.append(clause_sql)
- params.extend(clause_params)
- for t in self.query.extra_tables:
- alias, _ = self.query.table_alias(t)
- # Only add the alias if it's not already present (the table_alias()
- # call increments the refcount, so an alias refcount of one means
- # this is the only reference).
- if (
- alias not in self.query.alias_map
- or self.query.alias_refcount[alias] == 1
- ):
- result.append(", %s" % self.quote_name_unless_alias(alias))
- return result, params
- def get_related_selections(
- self,
- select,
- select_mask,
- opts=None,
- root_alias=None,
- cur_depth=1,
- requested=None,
- restricted=None,
- ):
- """
- Fill in the information needed for a select_related query. The current
- depth is measured as the number of connections away from the root model
- (for example, cur_depth=1 means we are looking at models with direct
- connections to the root model).
- """
- def _get_field_choices():
- direct_choices = (f.name for f in opts.fields if f.is_relation)
- reverse_choices = (
- f.field.related_query_name()
- for f in opts.related_objects
- if f.field.unique
- )
- return chain(
- direct_choices, reverse_choices, self.query._filtered_relations
- )
- related_klass_infos = []
- if not restricted and cur_depth > self.query.max_depth:
- # We've recursed far enough; bail out.
- return related_klass_infos
- if not opts:
- opts = self.query.get_meta()
- root_alias = self.query.get_initial_alias()
- # Setup for the case when only particular related fields should be
- # included in the related selection.
- fields_found = set()
- if requested is None:
- restricted = isinstance(self.query.select_related, dict)
- if restricted:
- requested = self.query.select_related
- def get_related_klass_infos(klass_info, related_klass_infos):
- klass_info["related_klass_infos"] = related_klass_infos
- for f in opts.fields:
- fields_found.add(f.name)
- if restricted:
- next = requested.get(f.name, {})
- if not f.is_relation:
- # If a non-related field is used like a relation,
- # or if a single non-relational field is given.
- if next or f.name in requested:
- raise FieldError(
- "Non-relational field given in select_related: '%s'. "
- "Choices are: %s"
- % (
- f.name,
- ", ".join(_get_field_choices()) or "(none)",
- )
- )
- else:
- next = False
- if not select_related_descend(f, restricted, requested, select_mask):
- continue
- related_select_mask = select_mask.get(f) or {}
- klass_info = {
- "model": f.remote_field.model,
- "field": f,
- "reverse": False,
- "local_setter": f.set_cached_value,
- "remote_setter": f.remote_field.set_cached_value
- if f.unique
- else lambda x, y: None,
- "from_parent": False,
- }
- related_klass_infos.append(klass_info)
- select_fields = []
- _, _, _, joins, _, _ = self.query.setup_joins([f.name], opts, root_alias)
- alias = joins[-1]
- columns = self.get_default_columns(
- related_select_mask, start_alias=alias, opts=f.remote_field.model._meta
- )
- for col in columns:
- select_fields.append(len(select))
- select.append((col, None))
- klass_info["select_fields"] = select_fields
- next_klass_infos = self.get_related_selections(
- select,
- related_select_mask,
- f.remote_field.model._meta,
- alias,
- cur_depth + 1,
- next,
- restricted,
- )
- get_related_klass_infos(klass_info, next_klass_infos)
- if restricted:
- related_fields = [
- (o.field, o.related_model)
- for o in opts.related_objects
- if o.field.unique and not o.many_to_many
- ]
- for related_field, model in related_fields:
- related_select_mask = select_mask.get(related_field) or {}
- if not select_related_descend(
- related_field,
- restricted,
- requested,
- related_select_mask,
- reverse=True,
- ):
- continue
- related_field_name = related_field.related_query_name()
- fields_found.add(related_field_name)
- join_info = self.query.setup_joins(
- [related_field_name], opts, root_alias
- )
- alias = join_info.joins[-1]
- from_parent = issubclass(model, opts.model) and model is not opts.model
- klass_info = {
- "model": model,
- "field": related_field,
- "reverse": True,
- "local_setter": related_field.remote_field.set_cached_value,
- "remote_setter": related_field.set_cached_value,
- "from_parent": from_parent,
- }
- related_klass_infos.append(klass_info)
- select_fields = []
- columns = self.get_default_columns(
- related_select_mask,
- start_alias=alias,
- opts=model._meta,
- from_parent=opts.model,
- )
- for col in columns:
- select_fields.append(len(select))
- select.append((col, None))
- klass_info["select_fields"] = select_fields
- next = requested.get(related_field.related_query_name(), {})
- next_klass_infos = self.get_related_selections(
- select,
- related_select_mask,
- model._meta,
- alias,
- cur_depth + 1,
- next,
- restricted,
- )
- get_related_klass_infos(klass_info, next_klass_infos)
- def local_setter(final_field, obj, from_obj):
- # Set a reverse fk object when relation is non-empty.
- if from_obj:
- final_field.remote_field.set_cached_value(from_obj, obj)
- def local_setter_noop(obj, from_obj):
- pass
- def remote_setter(name, obj, from_obj):
- setattr(from_obj, name, obj)
- for name in list(requested):
- # Filtered relations work only on the topmost level.
- if cur_depth > 1:
- break
- if name in self.query._filtered_relations:
- fields_found.add(name)
- final_field, _, join_opts, joins, _, _ = self.query.setup_joins(
- [name], opts, root_alias
- )
- model = join_opts.model
- alias = joins[-1]
- from_parent = (
- issubclass(model, opts.model) and model is not opts.model
- )
- klass_info = {
- "model": model,
- "field": final_field,
- "reverse": True,
- "local_setter": (
- partial(local_setter, final_field)
- if len(joins) <= 2
- else local_setter_noop
- ),
- "remote_setter": partial(remote_setter, name),
- "from_parent": from_parent,
- }
- related_klass_infos.append(klass_info)
- select_fields = []
- field_select_mask = select_mask.get((name, final_field)) or {}
- columns = self.get_default_columns(
- field_select_mask,
- start_alias=alias,
- opts=model._meta,
- from_parent=opts.model,
- )
- for col in columns:
- select_fields.append(len(select))
- select.append((col, None))
- klass_info["select_fields"] = select_fields
- next_requested = requested.get(name, {})
- next_klass_infos = self.get_related_selections(
- select,
- field_select_mask,
- opts=model._meta,
- root_alias=alias,
- cur_depth=cur_depth + 1,
- requested=next_requested,
- restricted=restricted,
- )
- get_related_klass_infos(klass_info, next_klass_infos)
- fields_not_found = set(requested).difference(fields_found)
- if fields_not_found:
- invalid_fields = ("'%s'" % s for s in fields_not_found)
- raise FieldError(
- "Invalid field name(s) given in select_related: %s. "
- "Choices are: %s"
- % (
- ", ".join(invalid_fields),
- ", ".join(_get_field_choices()) or "(none)",
- )
- )
- return related_klass_infos
- def get_select_for_update_of_arguments(self):
- """
- Return a quoted list of arguments for the SELECT FOR UPDATE OF part of
- the query.
- """
- def _get_parent_klass_info(klass_info):
- concrete_model = klass_info["model"]._meta.concrete_model
- for parent_model, parent_link in concrete_model._meta.parents.items():
- parent_list = parent_model._meta.get_parent_list()
- yield {
- "model": parent_model,
- "field": parent_link,
- "reverse": False,
- "select_fields": [
- select_index
- for select_index in klass_info["select_fields"]
- # Selected columns from a model or its parents.
- if (
- self.select[select_index][0].target.model == parent_model
- or self.select[select_index][0].target.model in parent_list
- )
- ],
- }
- def _get_first_selected_col_from_model(klass_info):
- """
- Find the first selected column from a model. If it doesn't exist,
- don't lock a model.
- select_fields is filled recursively, so it also contains fields
- from the parent models.
- """
- concrete_model = klass_info["model"]._meta.concrete_model
- for select_index in klass_info["select_fields"]:
- if self.select[select_index][0].target.model == concrete_model:
- return self.select[select_index][0]
- def _get_field_choices():
- """Yield all allowed field paths in breadth-first search order."""
- queue = collections.deque([(None, self.klass_info)])
- while queue:
- parent_path, klass_info = queue.popleft()
- if parent_path is None:
- path = []
- yield "self"
- else:
- field = klass_info["field"]
- if klass_info["reverse"]:
- field = field.remote_field
- path = parent_path + [field.name]
- yield LOOKUP_SEP.join(path)
- queue.extend(
- (path, klass_info)
- for klass_info in _get_parent_klass_info(klass_info)
- )
- queue.extend(
- (path, klass_info)
- for klass_info in klass_info.get("related_klass_infos", [])
- )
- if not self.klass_info:
- return []
- result = []
- invalid_names = []
- for name in self.query.select_for_update_of:
- klass_info = self.klass_info
- if name == "self":
- col = _get_first_selected_col_from_model(klass_info)
- else:
- for part in name.split(LOOKUP_SEP):
- klass_infos = (
- *klass_info.get("related_klass_infos", []),
- *_get_parent_klass_info(klass_info),
- )
- for related_klass_info in klass_infos:
- field = related_klass_info["field"]
- if related_klass_info["reverse"]:
- field = field.remote_field
- if field.name == part:
- klass_info = related_klass_info
- break
- else:
- klass_info = None
- break
- if klass_info is None:
- invalid_names.append(name)
- continue
- col = _get_first_selected_col_from_model(klass_info)
- if col is not None:
- if self.connection.features.select_for_update_of_column:
- result.append(self.compile(col)[0])
- else:
- result.append(self.quote_name_unless_alias(col.alias))
- if invalid_names:
- raise FieldError(
- "Invalid field name(s) given in select_for_update(of=(...)): %s. "
- "Only relational fields followed in the query are allowed. "
- "Choices are: %s."
- % (
- ", ".join(invalid_names),
- ", ".join(_get_field_choices()),
- )
- )
- return result
- def get_converters(self, expressions):
- converters = {}
- for i, expression in enumerate(expressions):
- if expression:
- backend_converters = self.connection.ops.get_db_converters(expression)
- field_converters = expression.get_db_converters(self.connection)
- if backend_converters or field_converters:
- converters[i] = (backend_converters + field_converters, expression)
- return converters
- def apply_converters(self, rows, converters):
- connection = self.connection
- converters = list(converters.items())
- for row in map(list, rows):
- for pos, (convs, expression) in converters:
- value = row[pos]
- for converter in convs:
- value = converter(value, expression, connection)
- row[pos] = value
- yield row
- def results_iter(
- self,
- results=None,
- tuple_expected=False,
- chunked_fetch=False,
- chunk_size=GET_ITERATOR_CHUNK_SIZE,
- ):
- """Return an iterator over the results from executing this query."""
- if results is None:
- results = self.execute_sql(
- MULTI, chunked_fetch=chunked_fetch, chunk_size=chunk_size
- )
- fields = [s[0] for s in self.select[0 : self.col_count]]
- converters = self.get_converters(fields)
- rows = chain.from_iterable(results)
- if converters:
- rows = self.apply_converters(rows, converters)
- if tuple_expected:
- rows = map(tuple, rows)
- return rows
- def has_results(self):
- """
- Backends (e.g. NoSQL) can override this in order to use optimized
- versions of "query has any results."
- """
- return bool(self.execute_sql(SINGLE))
- def execute_sql(
- self, result_type=MULTI, chunked_fetch=False, chunk_size=GET_ITERATOR_CHUNK_SIZE
- ):
- """
- Run the query against the database and return the result(s). The
- return value is a single data item if result_type is SINGLE, or an
- iterator over the results if the result_type is MULTI.
- result_type is either MULTI (use fetchmany() to retrieve all rows),
- SINGLE (only retrieve a single row), or None. In this last case, the
- cursor is returned if any query is executed, since it's used by
- subclasses such as InsertQuery). It's possible, however, that no query
- is needed, as the filters describe an empty set. In that case, None is
- returned, to avoid any unnecessary database interaction.
- """
- result_type = result_type or NO_RESULTS
- try:
- sql, params = self.as_sql()
- if not sql:
- raise EmptyResultSet
- except EmptyResultSet:
- if result_type == MULTI:
- return iter([])
- else:
- return
- if chunked_fetch:
- cursor = self.connection.chunked_cursor()
- else:
- cursor = self.connection.cursor()
- try:
- cursor.execute(sql, params)
- except Exception:
- # Might fail for server-side cursors (e.g. connection closed)
- cursor.close()
- raise
- if result_type == CURSOR:
- # Give the caller the cursor to process and close.
- return cursor
- if result_type == SINGLE:
- try:
- val = cursor.fetchone()
- if val:
- return val[0 : self.col_count]
- return val
- finally:
- # done with the cursor
- cursor.close()
- if result_type == NO_RESULTS:
- cursor.close()
- return
- result = cursor_iter(
- cursor,
- self.connection.features.empty_fetchmany_value,
- self.col_count if self.has_extra_select else None,
- chunk_size,
- )
- if not chunked_fetch or not self.connection.features.can_use_chunked_reads:
- # If we are using non-chunked reads, we return the same data
- # structure as normally, but ensure it is all read into memory
- # before going any further. Use chunked_fetch if requested,
- # unless the database doesn't support it.
- return list(result)
- return result
- def as_subquery_condition(self, alias, columns, compiler):
- qn = compiler.quote_name_unless_alias
- qn2 = self.connection.ops.quote_name
- for index, select_col in enumerate(self.query.select):
- lhs_sql, lhs_params = self.compile(select_col)
- rhs = "%s.%s" % (qn(alias), qn2(columns[index]))
- self.query.where.add(RawSQL("%s = %s" % (lhs_sql, rhs), lhs_params), AND)
- sql, params = self.as_sql()
- return "EXISTS (%s)" % sql, params
- def explain_query(self):
- result = list(self.execute_sql())
- # Some backends return 1 item tuples with strings, and others return
- # tuples with integers and strings. Flatten them out into strings.
- format_ = self.query.explain_info.format
- output_formatter = json.dumps if format_ and format_.lower() == "json" else str
- for row in result[0]:
- if not isinstance(row, str):
- yield " ".join(output_formatter(c) for c in row)
- else:
- yield row
- class SQLInsertCompiler(SQLCompiler):
- returning_fields = None
- returning_params = ()
- def field_as_sql(self, field, val):
- """
- Take a field and a value intended to be saved on that field, and
- return placeholder SQL and accompanying params. Check for raw values,
- expressions, and fields with get_placeholder() defined in that order.
- When field is None, consider the value raw and use it as the
- placeholder, with no corresponding parameters returned.
- """
- if field is None:
- # A field value of None means the value is raw.
- sql, params = val, []
- elif hasattr(val, "as_sql"):
- # This is an expression, let's compile it.
- sql, params = self.compile(val)
- elif hasattr(field, "get_placeholder"):
- # Some fields (e.g. geo fields) need special munging before
- # they can be inserted.
- sql, params = field.get_placeholder(val, self, self.connection), [val]
- else:
- # Return the common case for the placeholder
- sql, params = "%s", [val]
- # The following hook is only used by Oracle Spatial, which sometimes
- # needs to yield 'NULL' and [] as its placeholder and params instead
- # of '%s' and [None]. The 'NULL' placeholder is produced earlier by
- # OracleOperations.get_geom_placeholder(). The following line removes
- # the corresponding None parameter. See ticket #10888.
- params = self.connection.ops.modify_insert_params(sql, params)
- return sql, params
- def prepare_value(self, field, value):
- """
- Prepare a value to be used in a query by resolving it if it is an
- expression and otherwise calling the field's get_db_prep_save().
- """
- if hasattr(value, "resolve_expression"):
- value = value.resolve_expression(
- self.query, allow_joins=False, for_save=True
- )
- # Don't allow values containing Col expressions. They refer to
- # existing columns on a row, but in the case of insert the row
- # doesn't exist yet.
- if value.contains_column_references:
- raise ValueError(
- 'Failed to insert expression "%s" on %s. F() expressions '
- "can only be used to update, not to insert." % (value, field)
- )
- if value.contains_aggregate:
- raise FieldError(
- "Aggregate functions are not allowed in this query "
- "(%s=%r)." % (field.name, value)
- )
- if value.contains_over_clause:
- raise FieldError(
- "Window expressions are not allowed in this query (%s=%r)."
- % (field.name, value)
- )
- return field.get_db_prep_save(value, connection=self.connection)
- def pre_save_val(self, field, obj):
- """
- Get the given field's value off the given obj. pre_save() is used for
- things like auto_now on DateTimeField. Skip it if this is a raw query.
- """
- if self.query.raw:
- return getattr(obj, field.attname)
- return field.pre_save(obj, add=True)
- def assemble_as_sql(self, fields, value_rows):
- """
- Take a sequence of N fields and a sequence of M rows of values, and
- generate placeholder SQL and parameters for each field and value.
- Return a pair containing:
- * a sequence of M rows of N SQL placeholder strings, and
- * a sequence of M rows of corresponding parameter values.
- Each placeholder string may contain any number of '%s' interpolation
- strings, and each parameter row will contain exactly as many params
- as the total number of '%s's in the corresponding placeholder row.
- """
- if not value_rows:
- return [], []
- # list of (sql, [params]) tuples for each object to be saved
- # Shape: [n_objs][n_fields][2]
- rows_of_fields_as_sql = (
- (self.field_as_sql(field, v) for field, v in zip(fields, row))
- for row in value_rows
- )
- # tuple like ([sqls], [[params]s]) for each object to be saved
- # Shape: [n_objs][2][n_fields]
- sql_and_param_pair_rows = (zip(*row) for row in rows_of_fields_as_sql)
- # Extract separate lists for placeholders and params.
- # Each of these has shape [n_objs][n_fields]
- placeholder_rows, param_rows = zip(*sql_and_param_pair_rows)
- # Params for each field are still lists, and need to be flattened.
- param_rows = [[p for ps in row for p in ps] for row in param_rows]
- return placeholder_rows, param_rows
- def as_sql(self):
- # We don't need quote_name_unless_alias() here, since these are all
- # going to be column names (so we can avoid the extra overhead).
- qn = self.connection.ops.quote_name
- opts = self.query.get_meta()
- insert_statement = self.connection.ops.insert_statement(
- on_conflict=self.query.on_conflict,
- )
- result = ["%s %s" % (insert_statement, qn(opts.db_table))]
- fields = self.query.fields or [opts.pk]
- result.append("(%s)" % ", ".join(qn(f.column) for f in fields))
- if self.query.fields:
- value_rows = [
- [
- self.prepare_value(field, self.pre_save_val(field, obj))
- for field in fields
- ]
- for obj in self.query.objs
- ]
- else:
- # An empty object.
- value_rows = [
- [self.connection.ops.pk_default_value()] for _ in self.query.objs
- ]
- fields = [None]
- # Currently the backends just accept values when generating bulk
- # queries and generate their own placeholders. Doing that isn't
- # necessary and it should be possible to use placeholders and
- # expressions in bulk inserts too.
- can_bulk = (
- not self.returning_fields and self.connection.features.has_bulk_insert
- )
- placeholder_rows, param_rows = self.assemble_as_sql(fields, value_rows)
- on_conflict_suffix_sql = self.connection.ops.on_conflict_suffix_sql(
- fields,
- self.query.on_conflict,
- (f.column for f in self.query.update_fields),
- (f.column for f in self.query.unique_fields),
- )
- if (
- self.returning_fields
- and self.connection.features.can_return_columns_from_insert
- ):
- if self.connection.features.can_return_rows_from_bulk_insert:
- result.append(
- self.connection.ops.bulk_insert_sql(fields, placeholder_rows)
- )
- params = param_rows
- else:
- result.append("VALUES (%s)" % ", ".join(placeholder_rows[0]))
- params = [param_rows[0]]
- if on_conflict_suffix_sql:
- result.append(on_conflict_suffix_sql)
- # Skip empty r_sql to allow subclasses to customize behavior for
- # 3rd party backends. Refs #19096.
- r_sql, self.returning_params = self.connection.ops.return_insert_columns(
- self.returning_fields
- )
- if r_sql:
- result.append(r_sql)
- params += [self.returning_params]
- return [(" ".join(result), tuple(chain.from_iterable(params)))]
- if can_bulk:
- result.append(self.connection.ops.bulk_insert_sql(fields, placeholder_rows))
- if on_conflict_suffix_sql:
- result.append(on_conflict_suffix_sql)
- return [(" ".join(result), tuple(p for ps in param_rows for p in ps))]
- else:
- if on_conflict_suffix_sql:
- result.append(on_conflict_suffix_sql)
- return [
- (" ".join(result + ["VALUES (%s)" % ", ".join(p)]), vals)
- for p, vals in zip(placeholder_rows, param_rows)
- ]
- def execute_sql(self, returning_fields=None):
- assert not (
- returning_fields
- and len(self.query.objs) != 1
- and not self.connection.features.can_return_rows_from_bulk_insert
- )
- opts = self.query.get_meta()
- self.returning_fields = returning_fields
- with self.connection.cursor() as cursor:
- for sql, params in self.as_sql():
- cursor.execute(sql, params)
- if not self.returning_fields:
- return []
- if (
- self.connection.features.can_return_rows_from_bulk_insert
- and len(self.query.objs) > 1
- ):
- rows = self.connection.ops.fetch_returned_insert_rows(cursor)
- elif self.connection.features.can_return_columns_from_insert:
- assert len(self.query.objs) == 1
- rows = [
- self.connection.ops.fetch_returned_insert_columns(
- cursor,
- self.returning_params,
- )
- ]
- else:
- rows = [
- (
- self.connection.ops.last_insert_id(
- cursor,
- opts.db_table,
- opts.pk.column,
- ),
- )
- ]
- cols = [field.get_col(opts.db_table) for field in self.returning_fields]
- converters = self.get_converters(cols)
- if converters:
- rows = list(self.apply_converters(rows, converters))
- return rows
- class SQLDeleteCompiler(SQLCompiler):
- @cached_property
- def single_alias(self):
- # Ensure base table is in aliases.
- self.query.get_initial_alias()
- return sum(self.query.alias_refcount[t] > 0 for t in self.query.alias_map) == 1
- @classmethod
- def _expr_refs_base_model(cls, expr, base_model):
- if isinstance(expr, Query):
- return expr.model == base_model
- if not hasattr(expr, "get_source_expressions"):
- return False
- return any(
- cls._expr_refs_base_model(source_expr, base_model)
- for source_expr in expr.get_source_expressions()
- )
- @cached_property
- def contains_self_reference_subquery(self):
- return any(
- self._expr_refs_base_model(expr, self.query.model)
- for expr in chain(
- self.query.annotations.values(), self.query.where.children
- )
- )
- def _as_sql(self, query):
- delete = "DELETE FROM %s" % self.quote_name_unless_alias(query.base_table)
- try:
- where, params = self.compile(query.where)
- except FullResultSet:
- return delete, ()
- return f"{delete} WHERE {where}", tuple(params)
- def as_sql(self):
- """
- Create the SQL for this query. Return the SQL string and list of
- parameters.
- """
- if self.single_alias and (
- self.connection.features.delete_can_self_reference_subquery
- or not self.contains_self_reference_subquery
- ):
- return self._as_sql(self.query)
- innerq = self.query.clone()
- innerq.__class__ = Query
- innerq.clear_select_clause()
- pk = self.query.model._meta.pk
- innerq.select = [pk.get_col(self.query.get_initial_alias())]
- outerq = Query(self.query.model)
- if not self.connection.features.update_can_self_select:
- # Force the materialization of the inner query to allow reference
- # to the target table on MySQL.
- sql, params = innerq.get_compiler(connection=self.connection).as_sql()
- innerq = RawSQL("SELECT * FROM (%s) subquery" % sql, params)
- outerq.add_filter("pk__in", innerq)
- return self._as_sql(outerq)
- class SQLUpdateCompiler(SQLCompiler):
- def as_sql(self):
- """
- Create the SQL for this query. Return the SQL string and list of
- parameters.
- """
- self.pre_sql_setup()
- if not self.query.values:
- return "", ()
- qn = self.quote_name_unless_alias
- values, update_params = [], []
- for field, model, val in self.query.values:
- if hasattr(val, "resolve_expression"):
- val = val.resolve_expression(
- self.query, allow_joins=False, for_save=True
- )
- if val.contains_aggregate:
- raise FieldError(
- "Aggregate functions are not allowed in this query "
- "(%s=%r)." % (field.name, val)
- )
- if val.contains_over_clause:
- raise FieldError(
- "Window expressions are not allowed in this query "
- "(%s=%r)." % (field.name, val)
- )
- elif hasattr(val, "prepare_database_save"):
- if field.remote_field:
- val = val.prepare_database_save(field)
- else:
- raise TypeError(
- "Tried to update field %s with a model instance, %r. "
- "Use a value compatible with %s."
- % (field, val, field.__class__.__name__)
- )
- val = field.get_db_prep_save(val, connection=self.connection)
- # Getting the placeholder for the field.
- if hasattr(field, "get_placeholder"):
- placeholder = field.get_placeholder(val, self, self.connection)
- else:
- placeholder = "%s"
- name = field.column
- if hasattr(val, "as_sql"):
- sql, params = self.compile(val)
- values.append("%s = %s" % (qn(name), placeholder % sql))
- update_params.extend(params)
- elif val is not None:
- values.append("%s = %s" % (qn(name), placeholder))
- update_params.append(val)
- else:
- values.append("%s = NULL" % qn(name))
- table = self.query.base_table
- result = [
- "UPDATE %s SET" % qn(table),
- ", ".join(values),
- ]
- try:
- where, params = self.compile(self.query.where)
- except FullResultSet:
- params = []
- else:
- result.append("WHERE %s" % where)
- return " ".join(result), tuple(update_params + params)
- def execute_sql(self, result_type):
- """
- Execute the specified update. Return the number of rows affected by
- the primary update query. The "primary update query" is the first
- non-empty query that is executed. Row counts for any subsequent,
- related queries are not available.
- """
- cursor = super().execute_sql(result_type)
- try:
- rows = cursor.rowcount if cursor else 0
- is_empty = cursor is None
- finally:
- if cursor:
- cursor.close()
- for query in self.query.get_related_updates():
- aux_rows = query.get_compiler(self.using).execute_sql(result_type)
- if is_empty and aux_rows:
- rows = aux_rows
- is_empty = False
- return rows
- def pre_sql_setup(self):
- """
- If the update depends on results from other tables, munge the "where"
- conditions to match the format required for (portable) SQL updates.
- If multiple updates are required, pull out the id values to update at
- this point so that they don't change as a result of the progressive
- updates.
- """
- refcounts_before = self.query.alias_refcount.copy()
- # Ensure base table is in the query
- self.query.get_initial_alias()
- count = self.query.count_active_tables()
- if not self.query.related_updates and count == 1:
- return
- query = self.query.chain(klass=Query)
- query.select_related = False
- query.clear_ordering(force=True)
- query.extra = {}
- query.select = []
- meta = query.get_meta()
- fields = [meta.pk.name]
- related_ids_index = []
- for related in self.query.related_updates:
- if all(
- path.join_field.primary_key for path in meta.get_path_to_parent(related)
- ):
- # If a primary key chain exists to the targeted related update,
- # then the meta.pk value can be used for it.
- related_ids_index.append((related, 0))
- else:
- # This branch will only be reached when updating a field of an
- # ancestor that is not part of the primary key chain of a MTI
- # tree.
- related_ids_index.append((related, len(fields)))
- fields.append(related._meta.pk.name)
- query.add_fields(fields)
- super().pre_sql_setup()
- must_pre_select = (
- count > 1 and not self.connection.features.update_can_self_select
- )
- # Now we adjust the current query: reset the where clause and get rid
- # of all the tables we don't need (since they're in the sub-select).
- self.query.clear_where()
- if self.query.related_updates or must_pre_select:
- # Either we're using the idents in multiple update queries (so
- # don't want them to change), or the db backend doesn't support
- # selecting from the updating table (e.g. MySQL).
- idents = []
- related_ids = collections.defaultdict(list)
- for rows in query.get_compiler(self.using).execute_sql(MULTI):
- idents.extend(r[0] for r in rows)
- for parent, index in related_ids_index:
- related_ids[parent].extend(r[index] for r in rows)
- self.query.add_filter("pk__in", idents)
- self.query.related_ids = related_ids
- else:
- # The fast path. Filters and updates in one query.
- self.query.add_filter("pk__in", query)
- self.query.reset_refcounts(refcounts_before)
- class SQLAggregateCompiler(SQLCompiler):
- def as_sql(self):
- """
- Create the SQL for this query. Return the SQL string and list of
- parameters.
- """
- sql, params = [], []
- for annotation in self.query.annotation_select.values():
- ann_sql, ann_params = self.compile(annotation)
- ann_sql, ann_params = annotation.select_format(self, ann_sql, ann_params)
- sql.append(ann_sql)
- params.extend(ann_params)
- self.col_count = len(self.query.annotation_select)
- sql = ", ".join(sql)
- params = tuple(params)
- inner_query_sql, inner_query_params = self.query.inner_query.get_compiler(
- self.using,
- elide_empty=self.elide_empty,
- ).as_sql(with_col_aliases=True)
- sql = "SELECT %s FROM (%s) subquery" % (sql, inner_query_sql)
- params += inner_query_params
- return sql, params
- def cursor_iter(cursor, sentinel, col_count, itersize):
- """
- Yield blocks of rows from a cursor and ensure the cursor is closed when
- done.
- """
- try:
- for rows in iter((lambda: cursor.fetchmany(itersize)), sentinel):
- yield rows if col_count is None else [r[:col_count] for r in rows]
- finally:
- cursor.close()
|