Filename | /usr/share/perl5/DBIx/Class/Storage/DBIHacks.pm |
Statements | Executed 10933 statements in 24.3ms |
Calls | P | F | Exclusive Time |
Inclusive Time |
Subroutine |
---|---|---|---|---|---|
316 | 1 | 1 | 71.2ms | 90.7ms | _resolve_column_info | DBIx::Class::Storage::DBIHacks::
315 | 1 | 1 | 12.7ms | 19.3ms | _collapse_cond | DBIx::Class::Storage::DBIHacks::
4857 | 4 | 1 | 9.03ms | 9.03ms | CORE:match (opcode) | DBIx::Class::Storage::DBIHacks::
315 | 1 | 1 | 5.41ms | 24.7ms | _extract_fixed_condition_columns | DBIx::Class::Storage::DBIHacks::
315 | 1 | 1 | 4.59ms | 5.51ms | _collapse_cond_unroll_pairs | DBIx::Class::Storage::DBIHacks::
316 | 1 | 1 | 3.99ms | 4.24ms | _resolve_ident_sources | DBIx::Class::Storage::DBIHacks::
630 | 2 | 1 | 452µs | 452µs | CORE:sort (opcode) | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 44µs | 424µs | _extract_order_criteria | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 38µs | 333µs | __ANON__[:869] | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 37µs | 49µs | BEGIN@16 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 14µs | 20µs | BEGIN@10 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 11µs | 31µs | BEGIN@19 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 11µs | 15µs | BEGIN@11 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 11µs | 29µs | BEGIN@17 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 10µs | 16µs | BEGIN@14 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 9µs | 26.4ms | BEGIN@13 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 8µs | 48µs | BEGIN@20 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 8µs | 195µs | BEGIN@21 | DBIx::Class::Storage::DBIHacks::
1 | 1 | 1 | 8µs | 27µs | BEGIN@18 | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | __ANON__[:316] | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _adjust_select_args_for_complex_prefetch | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _columns_comprise_identifying_set | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _extract_colinfo_of_stable_main_source_order_by_portion | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _find_join_path_to_node | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _group_over_selection | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _inner_join_to_node | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _order_by_is_stable | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _prune_unused_joins | DBIx::Class::Storage::DBIHacks::
0 | 0 | 0 | 0s | 0s | _resolve_aliastypes_from_select_args | DBIx::Class::Storage::DBIHacks::
Line | State ments |
Time on line |
Calls | Time in subs |
Code |
---|---|---|---|---|---|
1 | package #hide from PAUSE | ||||
2 | DBIx::Class::Storage::DBIHacks; | ||||
3 | |||||
4 | # | ||||
5 | # This module contains code that should never have seen the light of day, | ||||
6 | # does not belong in the Storage, or is otherwise unfit for public | ||||
7 | # display. The arrival of SQLA2 should immediately obsolete 90% of this | ||||
8 | # | ||||
9 | |||||
10 | 2 | 27µs | # spent 20µs (14+7) within DBIx::Class::Storage::DBIHacks::BEGIN@10 which was called:
# once (14µs+7µs) by base::import at line 10 # spent 20µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@10
# spent 6µs making 1 call to strict::import | ||
11 | 2 | 18µs | # spent 15µs (11+4) within DBIx::Class::Storage::DBIHacks::BEGIN@11 which was called:
# once (11µs+4µs) by base::import at line 11 # spent 15µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@11
# spent 4µs making 1 call to warnings::import | ||
12 | |||||
13 | 2 | 26.4ms | # spent 26.4ms (9µs+26.4) within DBIx::Class::Storage::DBIHacks::BEGIN@13 which was called:
# once (9µs+26.4ms) by base::import at line 13 # spent 26.4ms making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@13
# spent 26.4ms making 1 call to base::import, recursion: max depth 1, sum of overlapping time 26.4ms | ||
14 | 2 | 22µs | # spent 16µs (10+6) within DBIx::Class::Storage::DBIHacks::BEGIN@14 which was called:
# once (10µs+6µs) by base::import at line 14 # spent 16µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@14
# spent 6µs making 1 call to mro::import | ||
15 | |||||
16 | 2 | 60µs | # spent 49µs (37+12) within DBIx::Class::Storage::DBIHacks::BEGIN@16 which was called:
# once (37µs+12µs) by base::import at line 16 # spent 49µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@16
# spent 12µs making 1 call to List::Util::import | ||
17 | 2 | 47µs | # spent 29µs (11+18) within DBIx::Class::Storage::DBIHacks::BEGIN@17 which was called:
# once (11µs+18µs) by base::import at line 17 # spent 29µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@17
# spent 18µs making 1 call to Exporter::import | ||
18 | 2 | 45µs | # spent 27µs (8+18) within DBIx::Class::Storage::DBIHacks::BEGIN@18 which was called:
# once (8µs+18µs) by base::import at line 18 # spent 27µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@18
# spent 18µs making 1 call to Exporter::import | ||
19 | 2 | 51µs | # spent 31µs (11+20) within DBIx::Class::Storage::DBIHacks::BEGIN@19 which was called:
# once (11µs+20µs) by base::import at line 19 # spent 31µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@19
# spent 20µs making 1 call to Exporter::import | ||
20 | 2 | 88µs | # spent 48µs (8+40) within DBIx::Class::Storage::DBIHacks::BEGIN@20 which was called:
# once (8µs+40µs) by base::import at line 20 # spent 48µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@20
# spent 40µs making 1 call to DBIx::Class::Carp::import | ||
21 | 2 | 382µs | # spent 195µs (8+187) within DBIx::Class::Storage::DBIHacks::BEGIN@21 which was called:
# once (8µs+187µs) by base::import at line 21 # spent 195µs making 1 call to DBIx::Class::Storage::DBIHacks::BEGIN@21
# spent 187µs making 1 call to namespace::clean::import | ||
22 | |||||
23 | # | ||||
24 | # This code will remove non-selecting/non-restricting joins from | ||||
25 | # {from} specs, aiding the RDBMS query optimizer | ||||
26 | # | ||||
27 | sub _prune_unused_joins { | ||||
28 | my ($self, $attrs) = @_; | ||||
29 | |||||
30 | # only standard {from} specs are supported, and we could be disabled in general | ||||
31 | return ($attrs->{from}, {}) unless ( | ||||
32 | ref $attrs->{from} eq 'ARRAY' | ||||
33 | and | ||||
34 | @{$attrs->{from}} > 1 | ||||
35 | and | ||||
36 | ref $attrs->{from}[0] eq 'HASH' | ||||
37 | and | ||||
38 | ref $attrs->{from}[1] eq 'ARRAY' | ||||
39 | and | ||||
40 | $self->_use_join_optimizer | ||||
41 | ); | ||||
42 | |||||
43 | my $orig_aliastypes = $self->_resolve_aliastypes_from_select_args($attrs); | ||||
44 | |||||
45 | my $new_aliastypes = { %$orig_aliastypes }; | ||||
46 | |||||
47 | # we will be recreating this entirely | ||||
48 | my @reclassify = 'joining'; | ||||
49 | |||||
50 | # a grouped set will not be affected by amount of rows. Thus any | ||||
51 | # purely multiplicator classifications can go | ||||
52 | # (will be reintroduced below if needed by something else) | ||||
53 | push @reclassify, qw(multiplying premultiplied) | ||||
54 | if $attrs->{_force_prune_multiplying_joins} or $attrs->{group_by}; | ||||
55 | |||||
56 | # nuke what will be recalculated | ||||
57 | delete @{$new_aliastypes}{@reclassify}; | ||||
58 | |||||
59 | my @newfrom = $attrs->{from}[0]; # FROM head is always present | ||||
60 | |||||
61 | # recalculate what we need once the multipliers are potentially gone | ||||
62 | # ignore premultiplies, since they do not add any value to anything | ||||
63 | my %need_joins; | ||||
64 | for ( @{$new_aliastypes}{grep { $_ ne 'premultiplied' } keys %$new_aliastypes }) { | ||||
65 | # add all requested aliases | ||||
66 | $need_joins{$_} = 1 for keys %$_; | ||||
67 | |||||
68 | # add all their parents (as per joinpath which is an AoH { table => alias }) | ||||
69 | $need_joins{$_} = 1 for map { values %$_ } map { @{$_->{-parents}} } values %$_; | ||||
70 | } | ||||
71 | |||||
72 | for my $j (@{$attrs->{from}}[1..$#{$attrs->{from}}]) { | ||||
73 | push @newfrom, $j if ( | ||||
74 | (! defined $j->[0]{-alias}) # legacy crap | ||||
75 | || | ||||
76 | $need_joins{$j->[0]{-alias}} | ||||
77 | ); | ||||
78 | } | ||||
79 | |||||
80 | # we have a new set of joiners - for everything we nuked pull the classification | ||||
81 | # off the original stack | ||||
82 | for my $ctype (@reclassify) { | ||||
83 | $new_aliastypes->{$ctype} = { map | ||||
84 | { $need_joins{$_} ? ( $_ => $orig_aliastypes->{$ctype}{$_} ) : () } | ||||
85 | keys %{$orig_aliastypes->{$ctype}} | ||||
86 | } | ||||
87 | } | ||||
88 | |||||
89 | return ( \@newfrom, $new_aliastypes ); | ||||
90 | } | ||||
91 | |||||
92 | # | ||||
93 | # This is the code producing joined subqueries like: | ||||
94 | # SELECT me.*, other.* FROM ( SELECT me.* FROM ... ) JOIN other ON ... | ||||
95 | # | ||||
96 | sub _adjust_select_args_for_complex_prefetch { | ||||
97 | my ($self, $attrs) = @_; | ||||
98 | |||||
99 | $self->throw_exception ('Complex prefetches are not supported on resultsets with a custom from attribute') unless ( | ||||
100 | ref $attrs->{from} eq 'ARRAY' | ||||
101 | and | ||||
102 | @{$attrs->{from}} > 1 | ||||
103 | and | ||||
104 | ref $attrs->{from}[0] eq 'HASH' | ||||
105 | and | ||||
106 | ref $attrs->{from}[1] eq 'ARRAY' | ||||
107 | ); | ||||
108 | |||||
109 | my $root_alias = $attrs->{alias}; | ||||
110 | |||||
111 | # generate inner/outer attribute lists, remove stuff that doesn't apply | ||||
112 | my $outer_attrs = { %$attrs }; | ||||
113 | delete @{$outer_attrs}{qw(from bind rows offset group_by _grouped_by_distinct having)}; | ||||
114 | |||||
115 | my $inner_attrs = { %$attrs, _simple_passthrough_construction => 1 }; | ||||
116 | delete @{$inner_attrs}{qw(for collapse select as)}; | ||||
117 | |||||
118 | # there is no point of ordering the insides if there is no limit | ||||
119 | delete $inner_attrs->{order_by} if ( | ||||
120 | delete $inner_attrs->{_order_is_artificial} | ||||
121 | or | ||||
122 | ! $inner_attrs->{rows} | ||||
123 | ); | ||||
124 | |||||
125 | # generate the inner/outer select lists | ||||
126 | # for inside we consider only stuff *not* brought in by the prefetch | ||||
127 | # on the outside we substitute any function for its alias | ||||
128 | $outer_attrs->{select} = [ @{$attrs->{select}} ]; | ||||
129 | |||||
130 | my ($root_node, $root_node_offset); | ||||
131 | |||||
132 | for my $i (0 .. $#{$inner_attrs->{from}}) { | ||||
133 | my $node = $inner_attrs->{from}[$i]; | ||||
134 | my $h = (ref $node eq 'HASH') ? $node | ||||
135 | : (ref $node eq 'ARRAY' and ref $node->[0] eq 'HASH') ? $node->[0] | ||||
136 | : next | ||||
137 | ; | ||||
138 | |||||
139 | if ( ($h->{-alias}||'') eq $root_alias and $h->{-rsrc} ) { | ||||
140 | $root_node = $h; | ||||
141 | $root_node_offset = $i; | ||||
142 | last; | ||||
143 | } | ||||
144 | } | ||||
145 | |||||
146 | $self->throw_exception ('Complex prefetches are not supported on resultsets with a custom from attribute') | ||||
147 | unless $root_node; | ||||
148 | |||||
149 | # use the heavy duty resolver to take care of aliased/nonaliased naming | ||||
150 | my $colinfo = $self->_resolve_column_info($inner_attrs->{from}); | ||||
151 | my $selected_root_columns; | ||||
152 | |||||
153 | for my $i (0 .. $#{$outer_attrs->{select}}) { | ||||
154 | my $sel = $outer_attrs->{select}->[$i]; | ||||
155 | |||||
156 | next if ( | ||||
157 | $colinfo->{$sel} and $colinfo->{$sel}{-source_alias} ne $root_alias | ||||
158 | ); | ||||
159 | |||||
160 | if (ref $sel eq 'HASH' ) { | ||||
161 | $sel->{-as} ||= $attrs->{as}[$i]; | ||||
162 | $outer_attrs->{select}->[$i] = join ('.', $root_alias, ($sel->{-as} || "inner_column_$i") ); | ||||
163 | } | ||||
164 | elsif (! ref $sel and my $ci = $colinfo->{$sel}) { | ||||
165 | $selected_root_columns->{$ci->{-colname}} = 1; | ||||
166 | } | ||||
167 | |||||
168 | push @{$inner_attrs->{select}}, $sel; | ||||
169 | |||||
170 | push @{$inner_attrs->{as}}, $attrs->{as}[$i]; | ||||
171 | } | ||||
172 | |||||
173 | # We will need to fetch all native columns in the inner subquery, which may | ||||
174 | # be a part of an *outer* join condition, or an order_by (which needs to be | ||||
175 | # preserved outside), or wheres. In other words everything but the inner | ||||
176 | # selector | ||||
177 | # We can not just fetch everything because a potential has_many restricting | ||||
178 | # join collapse *will not work* on heavy data types. | ||||
179 | my $connecting_aliastypes = $self->_resolve_aliastypes_from_select_args({ | ||||
180 | %$inner_attrs, | ||||
181 | select => [], | ||||
182 | }); | ||||
183 | |||||
184 | for (sort map { keys %{$_->{-seen_columns}||{}} } map { values %$_ } values %$connecting_aliastypes) { | ||||
185 | my $ci = $colinfo->{$_} or next; | ||||
186 | if ( | ||||
187 | $ci->{-source_alias} eq $root_alias | ||||
188 | and | ||||
189 | ! $selected_root_columns->{$ci->{-colname}}++ | ||||
190 | ) { | ||||
191 | # adding it to both to keep limits not supporting dark selectors happy | ||||
192 | push @{$inner_attrs->{select}}, $ci->{-fq_colname}; | ||||
193 | push @{$inner_attrs->{as}}, $ci->{-fq_colname}; | ||||
194 | } | ||||
195 | } | ||||
196 | |||||
197 | # construct the inner {from} and lock it in a subquery | ||||
198 | # we need to prune first, because this will determine if we need a group_by below | ||||
199 | # throw away all non-selecting, non-restricting multijoins | ||||
200 | # (since we def. do not care about multiplication of the contents of the subquery) | ||||
201 | my $inner_subq = do { | ||||
202 | |||||
203 | # must use it here regardless of user requests (vastly gentler on optimizer) | ||||
204 | local $self->{_use_join_optimizer} = 1; | ||||
205 | |||||
206 | # throw away multijoins since we def. do not care about those inside the subquery | ||||
207 | ($inner_attrs->{from}, my $inner_aliastypes) = $self->_prune_unused_joins ({ | ||||
208 | %$inner_attrs, _force_prune_multiplying_joins => 1 | ||||
209 | }); | ||||
210 | |||||
211 | # uh-oh a multiplier (which is not us) left in, this is a problem for limits | ||||
212 | # we will need to add a group_by to collapse the resultset for proper counts | ||||
213 | if ( | ||||
214 | grep { $_ ne $root_alias } keys %{ $inner_aliastypes->{multiplying} || {} } | ||||
215 | and | ||||
216 | ( ! $inner_aliastypes->{grouping} or $inner_attrs->{_grouped_by_distinct} ) | ||||
217 | |||||
218 | ) { | ||||
219 | |||||
220 | my $cur_sel = { map { $_ => 1 } @{$inner_attrs->{select}} }; | ||||
221 | |||||
222 | # *possibly* supplement the main selection with pks if not already | ||||
223 | # there, as they will have to be a part of the group_by to collapse | ||||
224 | # things properly | ||||
225 | my $inner_select_with_extras; | ||||
226 | my @pks = map { "$root_alias.$_" } $root_node->{-rsrc}->primary_columns | ||||
227 | or $self->throw_exception( sprintf | ||||
228 | 'Unable to perform complex limited prefetch off %s without declared primary key', | ||||
229 | $root_node->{-rsrc}->source_name, | ||||
230 | ); | ||||
231 | for my $col (@pks) { | ||||
232 | push @{ $inner_select_with_extras ||= [ @{$inner_attrs->{select}} ] }, $col | ||||
233 | unless $cur_sel->{$col}++; | ||||
234 | } | ||||
235 | |||||
236 | ($inner_attrs->{group_by}, $inner_attrs->{order_by}) = $self->_group_over_selection({ | ||||
237 | %$inner_attrs, | ||||
238 | $inner_select_with_extras ? ( select => $inner_select_with_extras ) : (), | ||||
239 | _aliastypes => $inner_aliastypes, | ||||
240 | }); | ||||
241 | } | ||||
242 | |||||
243 | # we already optimized $inner_attrs->{from} above | ||||
244 | # and already local()ized | ||||
245 | $self->{_use_join_optimizer} = 0; | ||||
246 | |||||
247 | # generate the subquery | ||||
248 | $self->_select_args_to_query ( | ||||
249 | @{$inner_attrs}{qw(from select where)}, | ||||
250 | $inner_attrs, | ||||
251 | ); | ||||
252 | }; | ||||
253 | |||||
254 | # Generate the outer from - this is relatively easy (really just replace | ||||
255 | # the join slot with the subquery), with a major caveat - we can not | ||||
256 | # join anything that is non-selecting (not part of the prefetch), but at | ||||
257 | # the same time is a multi-type relationship, as it will explode the result. | ||||
258 | # | ||||
259 | # There are two possibilities here | ||||
260 | # - either the join is non-restricting, in which case we simply throw it away | ||||
261 | # - it is part of the restrictions, in which case we need to collapse the outer | ||||
262 | # result by tackling yet another group_by to the outside of the query | ||||
263 | |||||
264 | # work on a shallow copy | ||||
265 | my @orig_from = @{$attrs->{from}}; | ||||
266 | |||||
267 | |||||
268 | $outer_attrs->{from} = \ my @outer_from; | ||||
269 | |||||
270 | # we may not be the head | ||||
271 | if ($root_node_offset) { | ||||
272 | # first generate the outer_from, up to the substitution point | ||||
273 | @outer_from = splice @orig_from, 0, $root_node_offset; | ||||
274 | |||||
275 | # substitute the subq at the right spot | ||||
276 | push @outer_from, [ | ||||
277 | { | ||||
278 | -alias => $root_alias, | ||||
279 | -rsrc => $root_node->{-rsrc}, | ||||
280 | $root_alias => $inner_subq, | ||||
281 | }, | ||||
282 | # preserve attrs from what is now the head of the from after the splice | ||||
283 | @{$orig_from[0]}[1 .. $#{$orig_from[0]}], | ||||
284 | ]; | ||||
285 | } | ||||
286 | else { | ||||
287 | @outer_from = { | ||||
288 | -alias => $root_alias, | ||||
289 | -rsrc => $root_node->{-rsrc}, | ||||
290 | $root_alias => $inner_subq, | ||||
291 | }; | ||||
292 | } | ||||
293 | |||||
294 | shift @orig_from; # what we just replaced above | ||||
295 | |||||
296 | # scan the *remaining* from spec against different attributes, and see which joins are needed | ||||
297 | # in what role | ||||
298 | my $outer_aliastypes = $outer_attrs->{_aliastypes} = | ||||
299 | $self->_resolve_aliastypes_from_select_args({ %$outer_attrs, from => \@orig_from }); | ||||
300 | |||||
301 | # unroll parents | ||||
302 | my ($outer_select_chain, @outer_nonselecting_chains) = map { +{ | ||||
303 | map { $_ => 1 } map { values %$_} map { @{$_->{-parents}} } values %{ $outer_aliastypes->{$_} || {} } | ||||
304 | } } qw/selecting restricting grouping ordering/; | ||||
305 | |||||
306 | # see what's left - throw away if not selecting/restricting | ||||
307 | my $may_need_outer_group_by; | ||||
308 | while (my $j = shift @orig_from) { | ||||
309 | my $alias = $j->[0]{-alias}; | ||||
310 | |||||
311 | if ( | ||||
312 | $outer_select_chain->{$alias} | ||||
313 | ) { | ||||
314 | push @outer_from, $j | ||||
315 | } | ||||
316 | elsif (first { $_->{$alias} } @outer_nonselecting_chains ) { | ||||
317 | push @outer_from, $j; | ||||
318 | $may_need_outer_group_by ||= $outer_aliastypes->{multiplying}{$alias} ? 1 : 0; | ||||
319 | } | ||||
320 | } | ||||
321 | |||||
322 | # also throw in a synthetic group_by if a non-selecting multiplier, | ||||
323 | # to guard against cross-join explosions | ||||
324 | # the logic is somewhat fragile, but relies on the idea that if a user supplied | ||||
325 | # a group by on their own - they know what they were doing | ||||
326 | if ( $may_need_outer_group_by and $attrs->{_grouped_by_distinct} ) { | ||||
327 | ($outer_attrs->{group_by}, $outer_attrs->{order_by}) = $self->_group_over_selection ({ | ||||
328 | %$outer_attrs, | ||||
329 | from => \@outer_from, | ||||
330 | }); | ||||
331 | } | ||||
332 | |||||
333 | # This is totally horrific - the {where} ends up in both the inner and outer query | ||||
334 | # Unfortunately not much can be done until SQLA2 introspection arrives, and even | ||||
335 | # then if where conditions apply to the *right* side of the prefetch, you may have | ||||
336 | # to both filter the inner select (e.g. to apply a limit) and then have to re-filter | ||||
337 | # the outer select to exclude joins you didn't want in the first place | ||||
338 | # | ||||
339 | # OTOH it can be seen as a plus: <ash> (notes that this query would make a DBA cry ;) | ||||
340 | return $outer_attrs; | ||||
341 | } | ||||
342 | |||||
343 | # | ||||
344 | # I KNOW THIS SUCKS! GET SQLA2 OUT THE DOOR SO THIS CAN DIE! | ||||
345 | # | ||||
346 | # Due to a lack of SQLA2 we fall back to crude scans of all the | ||||
347 | # select/where/order/group attributes, in order to determine what | ||||
348 | # aliases are needed to fulfill the query. This information is used | ||||
349 | # throughout the code to prune unnecessary JOINs from the queries | ||||
350 | # in an attempt to reduce the execution time. | ||||
351 | # Although the method is pretty horrific, the worst thing that can | ||||
352 | # happen is for it to fail due to some scalar SQL, which in turn will | ||||
353 | # result in a vocal exception. | ||||
354 | sub _resolve_aliastypes_from_select_args { | ||||
355 | my ( $self, $attrs ) = @_; | ||||
356 | |||||
357 | $self->throw_exception ('Unable to analyze custom {from}') | ||||
358 | if ref $attrs->{from} ne 'ARRAY'; | ||||
359 | |||||
360 | # what we will return | ||||
361 | my $aliases_by_type; | ||||
362 | |||||
363 | # see what aliases are there to work with | ||||
364 | # and record who is a multiplier and who is premultiplied | ||||
365 | my $alias_list; | ||||
366 | for my $node (@{$attrs->{from}}) { | ||||
367 | |||||
368 | my $j = $node; | ||||
369 | $j = $j->[0] if ref $j eq 'ARRAY'; | ||||
370 | my $al = $j->{-alias} | ||||
371 | or next; | ||||
372 | |||||
373 | $alias_list->{$al} = $j; | ||||
374 | |||||
375 | $aliases_by_type->{multiplying}{$al} ||= { -parents => $j->{-join_path}||[] } | ||||
376 | # not array == {from} head == can't be multiplying | ||||
377 | if ref($node) eq 'ARRAY' and ! $j->{-is_single}; | ||||
378 | |||||
379 | $aliases_by_type->{premultiplied}{$al} ||= { -parents => $j->{-join_path}||[] } | ||||
380 | # parts of the path that are not us but are multiplying | ||||
381 | if grep { $aliases_by_type->{multiplying}{$_} } | ||||
382 | grep { $_ ne $al } | ||||
383 | map { values %$_ } | ||||
384 | @{ $j->{-join_path}||[] } | ||||
385 | } | ||||
386 | |||||
387 | # get a column to source/alias map (including unambiguous unqualified ones) | ||||
388 | my $colinfo = $self->_resolve_column_info ($attrs->{from}); | ||||
389 | |||||
390 | # set up a botched SQLA | ||||
391 | my $sql_maker = $self->sql_maker; | ||||
392 | |||||
393 | # these are throw away results, do not pollute the bind stack | ||||
394 | local $sql_maker->{where_bind}; | ||||
395 | local $sql_maker->{group_bind}; | ||||
396 | local $sql_maker->{having_bind}; | ||||
397 | local $sql_maker->{from_bind}; | ||||
398 | |||||
399 | # we can't scan properly without any quoting (\b doesn't cut it | ||||
400 | # everywhere), so unless there is proper quoting set - use our | ||||
401 | # own weird impossible character. | ||||
402 | # Also in the case of no quoting, we need to explicitly disable | ||||
403 | # name_sep, otherwise sorry nasty legacy syntax like | ||||
404 | # { 'count(foo.id)' => { '>' => 3 } } will stop working >:( | ||||
405 | local $sql_maker->{quote_char} = $sql_maker->{quote_char}; | ||||
406 | local $sql_maker->{name_sep} = $sql_maker->{name_sep}; | ||||
407 | |||||
408 | unless (defined $sql_maker->{quote_char} and length $sql_maker->{quote_char}) { | ||||
409 | $sql_maker->{quote_char} = ["\x00", "\xFF"]; | ||||
410 | # if we don't unset it we screw up retarded but unfortunately working | ||||
411 | # 'MAX(foo.bar)' => { '>', 3 } | ||||
412 | $sql_maker->{name_sep} = ''; | ||||
413 | } | ||||
414 | |||||
415 | my ($lquote, $rquote, $sep) = map { quotemeta $_ } ($sql_maker->_quote_chars, $sql_maker->name_sep); | ||||
416 | |||||
417 | # generate sql chunks | ||||
418 | my $to_scan = { | ||||
419 | restricting => [ | ||||
420 | ($sql_maker->_recurse_where ($attrs->{where}))[0], | ||||
421 | $sql_maker->_parse_rs_attrs ({ having => $attrs->{having} }), | ||||
422 | ], | ||||
423 | grouping => [ | ||||
424 | $sql_maker->_parse_rs_attrs ({ group_by => $attrs->{group_by} }), | ||||
425 | ], | ||||
426 | joining => [ | ||||
427 | $sql_maker->_recurse_from ( | ||||
428 | ref $attrs->{from}[0] eq 'ARRAY' ? $attrs->{from}[0][0] : $attrs->{from}[0], | ||||
429 | @{$attrs->{from}}[1 .. $#{$attrs->{from}}], | ||||
430 | ), | ||||
431 | ], | ||||
432 | selecting => [ | ||||
433 | map { ($sql_maker->_recurse_fields($_))[0] } @{$attrs->{select}}, | ||||
434 | ], | ||||
435 | ordering => [ | ||||
436 | map { $_->[0] } $self->_extract_order_criteria ($attrs->{order_by}, $sql_maker), | ||||
437 | ], | ||||
438 | }; | ||||
439 | |||||
440 | # throw away empty chunks and all 2-value arrayrefs: the thinking is that these are | ||||
441 | # bind value specs left in by the sloppy renderer above. It is ok to do this | ||||
442 | # at this point, since we are going to end up rewriting this crap anyway | ||||
443 | for my $v (values %$to_scan) { | ||||
444 | my @nv; | ||||
445 | for (@$v) { | ||||
446 | next if ( | ||||
447 | ! defined $_ | ||||
448 | or | ||||
449 | ( | ||||
450 | ref $_ eq 'ARRAY' | ||||
451 | and | ||||
452 | ( @$_ == 0 or @$_ == 2 ) | ||||
453 | ) | ||||
454 | ); | ||||
455 | |||||
456 | if (ref $_) { | ||||
457 | require Data::Dumper::Concise; | ||||
458 | $self->throw_exception("Unexpected ref in scan-plan: " . Data::Dumper::Concise::Dumper($v) ); | ||||
459 | } | ||||
460 | |||||
461 | push @nv, $_; | ||||
462 | } | ||||
463 | |||||
464 | $v = \@nv; | ||||
465 | } | ||||
466 | |||||
467 | # kill all selectors which look like a proper subquery | ||||
468 | # this is a sucky heuristic *BUT* - if we get it wrong the query will simply | ||||
469 | # fail to run, so we are relatively safe | ||||
470 | $to_scan->{selecting} = [ grep { | ||||
471 | $_ !~ / \A \s* \( \s* SELECT \s+ .+? \s+ FROM \s+ .+? \) \s* \z /xsi | ||||
472 | } @{ $to_scan->{selecting} || [] } ]; | ||||
473 | |||||
474 | # first see if we have any exact matches (qualified or unqualified) | ||||
475 | for my $type (keys %$to_scan) { | ||||
476 | for my $piece (@{$to_scan->{$type}}) { | ||||
477 | if ($colinfo->{$piece} and my $alias = $colinfo->{$piece}{-source_alias}) { | ||||
478 | $aliases_by_type->{$type}{$alias} ||= { -parents => $alias_list->{$alias}{-join_path}||[] }; | ||||
479 | $aliases_by_type->{$type}{$alias}{-seen_columns}{$colinfo->{$piece}{-fq_colname}} = $piece; | ||||
480 | } | ||||
481 | } | ||||
482 | } | ||||
483 | |||||
484 | # now loop through all fully qualified columns and get the corresponding | ||||
485 | # alias (should work even if they are in scalarrefs) | ||||
486 | for my $alias (keys %$alias_list) { | ||||
487 | my $al_re = qr/ | ||||
488 | $lquote $alias $rquote $sep (?: $lquote ([^$rquote]+) $rquote )? | ||||
489 | | | ||||
490 | \b $alias \. ([^\s\)\($rquote]+)? | ||||
491 | /x; | ||||
492 | |||||
493 | for my $type (keys %$to_scan) { | ||||
494 | for my $piece (@{$to_scan->{$type}}) { | ||||
495 | if (my @matches = $piece =~ /$al_re/g) { | ||||
496 | $aliases_by_type->{$type}{$alias} ||= { -parents => $alias_list->{$alias}{-join_path}||[] }; | ||||
497 | $aliases_by_type->{$type}{$alias}{-seen_columns}{"$alias.$_"} = "$alias.$_" | ||||
498 | for grep { defined $_ } @matches; | ||||
499 | } | ||||
500 | } | ||||
501 | } | ||||
502 | } | ||||
503 | |||||
504 | # now loop through unqualified column names, and try to locate them within | ||||
505 | # the chunks | ||||
506 | for my $col (keys %$colinfo) { | ||||
507 | next if $col =~ / \. /x; # if column is qualified it was caught by the above | ||||
508 | |||||
509 | my $col_re = qr/ $lquote ($col) $rquote /x; | ||||
510 | |||||
511 | for my $type (keys %$to_scan) { | ||||
512 | for my $piece (@{$to_scan->{$type}}) { | ||||
513 | if ( my @matches = $piece =~ /$col_re/g) { | ||||
514 | my $alias = $colinfo->{$col}{-source_alias}; | ||||
515 | $aliases_by_type->{$type}{$alias} ||= { -parents => $alias_list->{$alias}{-join_path}||[] }; | ||||
516 | $aliases_by_type->{$type}{$alias}{-seen_columns}{"$alias.$_"} = $_ | ||||
517 | for grep { defined $_ } @matches; | ||||
518 | } | ||||
519 | } | ||||
520 | } | ||||
521 | } | ||||
522 | |||||
523 | # Add any non-left joins to the restriction list (such joins are indeed restrictions) | ||||
524 | for my $j (values %$alias_list) { | ||||
525 | my $alias = $j->{-alias} or next; | ||||
526 | $aliases_by_type->{restricting}{$alias} ||= { -parents => $j->{-join_path}||[] } if ( | ||||
527 | (not $j->{-join_type}) | ||||
528 | or | ||||
529 | ($j->{-join_type} !~ /^left (?: \s+ outer)? $/xi) | ||||
530 | ); | ||||
531 | } | ||||
532 | |||||
533 | for (keys %$aliases_by_type) { | ||||
534 | delete $aliases_by_type->{$_} unless keys %{$aliases_by_type->{$_}}; | ||||
535 | } | ||||
536 | |||||
537 | return $aliases_by_type; | ||||
538 | } | ||||
539 | |||||
540 | # This is the engine behind { distinct => 1 } and the general | ||||
541 | # complex prefetch grouper | ||||
542 | sub _group_over_selection { | ||||
543 | my ($self, $attrs) = @_; | ||||
544 | |||||
545 | my $colinfos = $self->_resolve_column_info ($attrs->{from}); | ||||
546 | |||||
547 | my (@group_by, %group_index); | ||||
548 | |||||
549 | # the logic is: if it is a { func => val } we assume an aggregate, | ||||
550 | # otherwise if \'...' or \[...] we assume the user knows what is | ||||
551 | # going on thus group over it | ||||
552 | for (@{$attrs->{select}}) { | ||||
553 | if (! ref($_) or ref ($_) ne 'HASH' ) { | ||||
554 | push @group_by, $_; | ||||
555 | $group_index{$_}++; | ||||
556 | if ($colinfos->{$_} and $_ !~ /\./ ) { | ||||
557 | # add a fully qualified version as well | ||||
558 | $group_index{"$colinfos->{$_}{-source_alias}.$_"}++; | ||||
559 | } | ||||
560 | } | ||||
561 | } | ||||
562 | |||||
563 | my @order_by = $self->_extract_order_criteria($attrs->{order_by}) | ||||
564 | or return (\@group_by, $attrs->{order_by}); | ||||
565 | |||||
566 | # add any order_by parts that are not already present in the group_by | ||||
567 | # to maintain SQL cross-compatibility and general sanity | ||||
568 | # | ||||
569 | # also in case the original selection is *not* unique, or in case part | ||||
570 | # of the ORDER BY refers to a multiplier - we will need to replace the | ||||
571 | # skipped order_by elements with their MIN/MAX equivalents as to maintain | ||||
572 | # the proper overall order without polluting the group criteria (and | ||||
573 | # possibly changing the outcome entirely) | ||||
574 | |||||
575 | my ($leftovers, $sql_maker, @new_order_by, $order_chunks, $aliastypes); | ||||
576 | |||||
577 | my $group_already_unique = $self->_columns_comprise_identifying_set($colinfos, \@group_by); | ||||
578 | |||||
579 | for my $o_idx (0 .. $#order_by) { | ||||
580 | |||||
581 | # if the chunk is already a min/max function - there is nothing left to touch | ||||
582 | next if $order_by[$o_idx][0] =~ /^ (?: min | max ) \s* \( .+ \) $/ix; | ||||
583 | |||||
584 | # only consider real columns (for functions the user got to do an explicit group_by) | ||||
585 | my $chunk_ci; | ||||
586 | if ( | ||||
587 | @{$order_by[$o_idx]} != 1 | ||||
588 | or | ||||
589 | ( ! ( $chunk_ci = $colinfos->{$order_by[$o_idx][0]} ) and $attrs->{_aliastypes} ) | ||||
590 | |||||
- - | |||||
593 | ) { | ||||
594 | push @$leftovers, $order_by[$o_idx][0]; | ||||
595 | } | ||||
596 | |||||
597 | next unless $chunk_ci; | ||||
598 | |||||
599 | # no duplication of group criteria | ||||
600 | next if $group_index{$chunk_ci->{-fq_colname}}; | ||||
601 | |||||
602 | $aliastypes ||= ( | ||||
603 | $attrs->{_aliastypes} | ||||
604 | or | ||||
605 | $self->_resolve_aliastypes_from_select_args({ | ||||
606 | from => $attrs->{from}, | ||||
607 | order_by => $attrs->{order_by}, | ||||
608 | }) | ||||
609 | ) if $group_already_unique; | ||||
610 | |||||
611 | # check that we are not ordering by a multiplier (if a check is requested at all) | ||||
612 | if ( | ||||
613 | $group_already_unique | ||||
614 | and | ||||
615 | ! $aliastypes->{multiplying}{$chunk_ci->{-source_alias}} | ||||
616 | and | ||||
617 | ! $aliastypes->{premultiplied}{$chunk_ci->{-source_alias}} | ||||
618 | ) { | ||||
619 | push @group_by, $chunk_ci->{-fq_colname}; | ||||
620 | $group_index{$chunk_ci->{-fq_colname}}++ | ||||
621 | } | ||||
622 | else { | ||||
623 | # We need to order by external columns without adding them to the group | ||||
624 | # (eiehter a non-unique selection, or a multi-external) | ||||
625 | # | ||||
626 | # This doesn't really make sense in SQL, however from DBICs point | ||||
627 | # of view is rather valid (e.g. order the leftmost objects by whatever | ||||
628 | # criteria and get the offset/rows many). There is a way around | ||||
629 | # this however in SQL - we simply tae the direction of each piece | ||||
630 | # of the external order and convert them to MIN(X) for ASC or MAX(X) | ||||
631 | # for DESC, and group_by the root columns. The end result should be | ||||
632 | # exactly what we expect | ||||
633 | |||||
634 | # FIXME - this code is a joke, will need to be completely rewritten in | ||||
635 | # the DQ branch. But I need to push a POC here, otherwise the | ||||
636 | # pesky tests won't pass | ||||
637 | # wrap any part of the order_by that "responds" to an ordering alias | ||||
638 | # into a MIN/MAX | ||||
639 | $sql_maker ||= $self->sql_maker; | ||||
640 | $order_chunks ||= [ | ||||
641 | map { ref $_ eq 'ARRAY' ? $_ : [ $_ ] } $sql_maker->_order_by_chunks($attrs->{order_by}) | ||||
642 | ]; | ||||
643 | |||||
644 | my ($chunk, $is_desc) = $sql_maker->_split_order_chunk($order_chunks->[$o_idx][0]); | ||||
645 | |||||
646 | $new_order_by[$o_idx] = \[ | ||||
647 | sprintf( '%s( %s )%s', | ||||
648 | ($is_desc ? 'MAX' : 'MIN'), | ||||
649 | $chunk, | ||||
650 | ($is_desc ? ' DESC' : ''), | ||||
651 | ), | ||||
652 | @ {$order_chunks->[$o_idx]} [ 1 .. $#{$order_chunks->[$o_idx]} ] | ||||
653 | ]; | ||||
654 | } | ||||
655 | } | ||||
656 | |||||
657 | $self->throw_exception ( sprintf | ||||
658 | 'Unable to programatically derive a required group_by from the supplied ' | ||||
659 | . 'order_by criteria. To proceed either add an explicit group_by, or ' | ||||
660 | . 'simplify your order_by to only include plain columns ' | ||||
661 | . '(supplied order_by: %s)', | ||||
662 | join ', ', map { "'$_'" } @$leftovers, | ||||
663 | ) if $leftovers; | ||||
664 | |||||
665 | # recreate the untouched order parts | ||||
666 | if (@new_order_by) { | ||||
667 | $new_order_by[$_] ||= \ $order_chunks->[$_] for ( 0 .. $#$order_chunks ); | ||||
668 | } | ||||
669 | |||||
670 | return ( | ||||
671 | \@group_by, | ||||
672 | (@new_order_by ? \@new_order_by : $attrs->{order_by} ), # same ref as original == unchanged | ||||
673 | ); | ||||
674 | } | ||||
675 | |||||
676 | # spent 4.24ms (3.99+255µs) within DBIx::Class::Storage::DBIHacks::_resolve_ident_sources which was called 316 times, avg 13µs/call:
# 316 times (3.99ms+255µs) by DBIx::Class::Storage::DBIHacks::_resolve_column_info at line 718, avg 13µs/call | ||||
677 | 84 | 31µs | my ($self, $ident) = @_; | ||
678 | |||||
679 | 84 | 43µs | my $alias2source = {}; | ||
680 | |||||
681 | # the reason this is so contrived is that $ident may be a {from} | ||||
682 | # structure, specifying multiple tables to join | ||||
683 | 84 | 397µs | 316 | 255µs | if ( blessed $ident && $ident->isa("DBIx::Class::ResultSource") ) { # spent 255µs making 316 calls to Scalar::Util::blessed, avg 809ns/call |
684 | # this is compat mode for insert/update/delete which do not deal with aliases | ||||
685 | $alias2source->{me} = $ident; | ||||
686 | } | ||||
687 | elsif (ref $ident eq 'ARRAY') { | ||||
688 | |||||
689 | 84 | 71µs | for (@$ident) { | ||
690 | 84 | 15µs | my $tabinfo; | ||
691 | 84 | 59µs | if (ref $_ eq 'HASH') { | ||
692 | $tabinfo = $_; | ||||
693 | } | ||||
694 | 84 | 32µs | if (ref $_ eq 'ARRAY' and ref $_->[0] eq 'HASH') { | ||
695 | $tabinfo = $_->[0]; | ||||
696 | } | ||||
697 | |||||
698 | 84 | 267µs | $alias2source->{$tabinfo->{-alias}} = $tabinfo->{-rsrc} | ||
699 | if ($tabinfo->{-rsrc}); | ||||
700 | } | ||||
701 | } | ||||
702 | |||||
703 | 84 | 210µs | return $alias2source; | ||
704 | } | ||||
705 | |||||
706 | # Takes $ident, \@column_names | ||||
707 | # | ||||
708 | # returns { $column_name => \%column_info, ... } | ||||
709 | # also note: this adds -result_source => $rsrc to the column info | ||||
710 | # | ||||
711 | # If no columns_names are supplied returns info about *all* columns | ||||
712 | # for all sources | ||||
713 | # spent 90.7ms (71.2+19.5) within DBIx::Class::Storage::DBIHacks::_resolve_column_info which was called 316 times, avg 287µs/call:
# 316 times (71.2ms+19.5ms) by DBIx::Class::Storage::DBI::__ANON__[/usr/share/perl5/DBIx/Class/Storage/DBI.pm:1698] at line 1684 of DBIx/Class/Storage/DBI.pm, avg 287µs/call | ||||
714 | 84 | 35µs | my ($self, $ident, $colnames) = @_; | ||
715 | |||||
716 | 84 | 15µs | return {} if $colnames and ! @$colnames; | ||
717 | |||||
718 | 84 | 173µs | 316 | 4.24ms | my $alias2src = $self->_resolve_ident_sources($ident); # spent 4.24ms making 316 calls to DBIx::Class::Storage::DBIHacks::_resolve_ident_sources, avg 13µs/call |
719 | |||||
720 | 84 | 15µs | my (%seen_cols, @auto_colnames); | ||
721 | |||||
722 | # compile a global list of column names, to be able to properly | ||||
723 | # disambiguate unqualified column names (if at all possible) | ||||
724 | 84 | 170µs | for my $alias (keys %$alias2src) { | ||
725 | 84 | 32µs | my $rsrc = $alias2src->{$alias}; | ||
726 | 84 | 284µs | 316 | 1.38ms | for my $colname ($rsrc->columns) { # spent 1.38ms making 316 calls to DBIx::Class::ResultSource::columns, avg 4µs/call |
727 | 420 | 401µs | push @{$seen_cols{$colname}}, $alias; | ||
728 | 420 | 445µs | push @auto_colnames, "$alias.$colname" unless $colnames; | ||
729 | } | ||||
730 | } | ||||
731 | |||||
732 | $colnames ||= [ | ||||
733 | @auto_colnames, | ||||
734 | 84 | 379µs | grep { @{$seen_cols{$_}} == 1 } (keys %seen_cols), | ||
735 | ]; | ||||
736 | |||||
737 | 84 | 25µs | my (%return, $colinfos); | ||
738 | 84 | 96µs | foreach my $col (@$colnames) { | ||
739 | 840 | 8.46ms | 3282 | 7.54ms | my ($source_alias, $colname) = $col =~ m/^ (?: ([^\.]+) \. )? (.+) $/x; # spent 7.54ms making 3282 calls to DBIx::Class::Storage::DBIHacks::CORE:match, avg 2µs/call |
740 | |||||
741 | # if the column was seen exactly once - we know which rsrc it came from | ||||
742 | $source_alias ||= $seen_cols{$colname}[0] | ||||
743 | 840 | 672µs | if ($seen_cols{$colname} and @{$seen_cols{$colname}} == 1); | ||
744 | |||||
745 | 840 | 110µs | next unless $source_alias; | ||
746 | |||||
747 | 840 | 279µs | my $rsrc = $alias2src->{$source_alias} | ||
748 | or next; | ||||
749 | |||||
750 | $return{$col} = { | ||||
751 | %{ | ||||
752 | 840 | 3.85ms | 316 | 6.30ms | ( $colinfos->{$source_alias} ||= $rsrc->columns_info )->{$colname} # spent 6.30ms making 316 calls to DBIx::Class::ResultSource::columns_info, avg 20µs/call |
753 | || | ||||
754 | $self->throw_exception( | ||||
755 | "No such column '$colname' on source " . $rsrc->source_name | ||||
756 | ); | ||||
757 | }, | ||||
758 | -result_source => $rsrc, | ||||
759 | -source_alias => $source_alias, | ||||
760 | -fq_colname => $col eq $colname ? "$source_alias.$col" : $col, | ||||
761 | -colname => $colname, | ||||
762 | }; | ||||
763 | |||||
764 | 840 | 1.03ms | $return{"$source_alias.$colname"} = $return{$col} if $col eq $colname; | ||
765 | } | ||||
766 | |||||
767 | 84 | 524µs | return \%return; | ||
768 | } | ||||
769 | |||||
770 | # The DBIC relationship chaining implementation is pretty simple - every | ||||
771 | # new related_relationship is pushed onto the {from} stack, and the {select} | ||||
772 | # window simply slides further in. This means that when we count somewhere | ||||
773 | # in the middle, we got to make sure that everything in the join chain is an | ||||
774 | # actual inner join, otherwise the count will come back with unpredictable | ||||
775 | # results (a resultset may be generated with _some_ rows regardless of if | ||||
776 | # the relation which the $rs currently selects has rows or not). E.g. | ||||
777 | # $artist_rs->cds->count - normally generates: | ||||
778 | # SELECT COUNT( * ) FROM artist me LEFT JOIN cd cds ON cds.artist = me.artistid | ||||
779 | # which actually returns the number of artists * (number of cds || 1) | ||||
780 | # | ||||
781 | # So what we do here is crawl {from}, determine if the current alias is at | ||||
782 | # the top of the stack, and if not - make sure the chain is inner-joined down | ||||
783 | # to the root. | ||||
784 | # | ||||
785 | sub _inner_join_to_node { | ||||
786 | my ($self, $from, $alias) = @_; | ||||
787 | |||||
788 | my $switch_branch = $self->_find_join_path_to_node($from, $alias); | ||||
789 | |||||
790 | return $from unless @{$switch_branch||[]}; | ||||
791 | |||||
792 | # So it looks like we will have to switch some stuff around. | ||||
793 | # local() is useless here as we will be leaving the scope | ||||
794 | # anyway, and deep cloning is just too fucking expensive | ||||
795 | # So replace the first hashref in the node arrayref manually | ||||
796 | my @new_from = ($from->[0]); | ||||
797 | my $sw_idx = { map { (values %$_), 1 } @$switch_branch }; #there's one k/v per join-path | ||||
798 | |||||
799 | for my $j (@{$from}[1 .. $#$from]) { | ||||
800 | my $jalias = $j->[0]{-alias}; | ||||
801 | |||||
802 | if ($sw_idx->{$jalias}) { | ||||
803 | my %attrs = %{$j->[0]}; | ||||
804 | delete $attrs{-join_type}; | ||||
805 | push @new_from, [ | ||||
806 | \%attrs, | ||||
807 | @{$j}[ 1 .. $#$j ], | ||||
808 | ]; | ||||
809 | } | ||||
810 | else { | ||||
811 | push @new_from, $j; | ||||
812 | } | ||||
813 | } | ||||
814 | |||||
815 | return \@new_from; | ||||
816 | } | ||||
817 | |||||
818 | sub _find_join_path_to_node { | ||||
819 | my ($self, $from, $target_alias) = @_; | ||||
820 | |||||
821 | # subqueries and other oddness are naturally not supported | ||||
822 | return undef if ( | ||||
823 | ref $from ne 'ARRAY' | ||||
824 | || | ||||
825 | ref $from->[0] ne 'HASH' | ||||
826 | || | ||||
827 | ! defined $from->[0]{-alias} | ||||
828 | ); | ||||
829 | |||||
830 | # no path - the head is the alias | ||||
831 | return [] if $from->[0]{-alias} eq $target_alias; | ||||
832 | |||||
833 | for my $i (1 .. $#$from) { | ||||
834 | return $from->[$i][0]{-join_path} if ( ($from->[$i][0]{-alias}||'') eq $target_alias ); | ||||
835 | } | ||||
836 | |||||
837 | # something else went quite wrong | ||||
838 | return undef; | ||||
839 | } | ||||
840 | |||||
841 | # spent 424µs (44+380) within DBIx::Class::Storage::DBIHacks::_extract_order_criteria which was called:
# once (44µs+380µs) by DBIx::Class::ResultSetColumn::new at line 77 of DBIx/Class/ResultSetColumn.pm | ||||
842 | 1 | 1µs | my ($self, $order_by, $sql_maker) = @_; | ||
843 | |||||
844 | # spent 333µs (38+295) within DBIx::Class::Storage::DBIHacks::__ANON__[/usr/share/perl5/DBIx/Class/Storage/DBIHacks.pm:869] which was called:
# once (38µs+295µs) by DBIx::Class::Storage::DBIHacks::_extract_order_criteria at line 882 | ||||
845 | 1 | 500ns | my ($sql_maker, $order_by, $orig_quote_chars) = @_; | ||
846 | |||||
847 | 1 | 300ns | return scalar $sql_maker->_order_by_chunks ($order_by) | ||
848 | unless wantarray; | ||||
849 | |||||
850 | 1 | 8µs | 1 | 236µs | my ($lq, $rq, $sep) = map { quotemeta($_) } ( # spent 236µs making 1 call to DBIx::Class::SQLMaker::name_sep |
851 | ($orig_quote_chars ? @$orig_quote_chars : $sql_maker->_quote_chars), | ||||
852 | $sql_maker->name_sep | ||||
853 | ); | ||||
854 | |||||
855 | 1 | 100ns | my @chunks; | ||
856 | 1 | 19µs | 1 | 58µs | for ($sql_maker->_order_by_chunks ($order_by) ) { # spent 58µs making 1 call to SQL::Abstract::_order_by_chunks |
857 | my $chunk = ref $_ ? [ @$_ ] : [ $_ ]; | ||||
858 | ($chunk->[0]) = $sql_maker->_split_order_chunk($chunk->[0]); | ||||
859 | |||||
860 | # order criteria may have come back pre-quoted (literals and whatnot) | ||||
861 | # this is fragile, but the best we can currently do | ||||
862 | $chunk->[0] =~ s/^ $lq (.+?) $rq $sep $lq (.+?) $rq $/"$1.$2"/xe | ||||
863 | or $chunk->[0] =~ s/^ $lq (.+) $rq $/$1/x; | ||||
864 | |||||
865 | push @chunks, $chunk; | ||||
866 | } | ||||
867 | |||||
868 | 1 | 4µs | return @chunks; | ||
869 | 1 | 6µs | }; | ||
870 | |||||
871 | 1 | 400ns | if ($sql_maker) { | ||
872 | return $parser->($sql_maker, $order_by); | ||||
873 | } | ||||
874 | else { | ||||
875 | 1 | 2µs | 1 | 40µs | $sql_maker = $self->sql_maker; # spent 40µs making 1 call to DBIx::Class::Storage::DBI::mysql::sql_maker |
876 | |||||
877 | # pass these in to deal with literals coming from | ||||
878 | # the user or the deep guts of prefetch | ||||
879 | 1 | 11µs | 1 | 8µs | my $orig_quote_chars = [$sql_maker->_quote_chars]; # spent 8µs making 1 call to DBIx::Class::SQLMaker::_quote_chars |
880 | |||||
881 | 1 | 2µs | local $sql_maker->{quote_char}; | ||
882 | 1 | 13µs | 1 | 333µs | return $parser->($sql_maker, $order_by, $orig_quote_chars); # spent 333µs making 1 call to DBIx::Class::Storage::DBIHacks::__ANON__[DBIx/Class/Storage/DBIHacks.pm:869] |
883 | } | ||||
884 | } | ||||
885 | |||||
886 | sub _order_by_is_stable { | ||||
887 | my ($self, $ident, $order_by, $where) = @_; | ||||
888 | |||||
889 | my @cols = ( | ||||
890 | ( map { $_->[0] } $self->_extract_order_criteria($order_by) ), | ||||
891 | ( $where ? keys %{ $self->_extract_fixed_condition_columns($where) } : () ), | ||||
892 | ) or return 0; | ||||
893 | |||||
894 | my $colinfo = $self->_resolve_column_info($ident, \@cols); | ||||
895 | |||||
896 | return keys %$colinfo | ||||
897 | ? $self->_columns_comprise_identifying_set( $colinfo, \@cols ) | ||||
898 | : 0 | ||||
899 | ; | ||||
900 | } | ||||
901 | |||||
902 | sub _columns_comprise_identifying_set { | ||||
903 | my ($self, $colinfo, $columns) = @_; | ||||
904 | |||||
905 | my $cols_per_src; | ||||
906 | $cols_per_src -> {$_->{-source_alias}} -> {$_->{-colname}} = $_ | ||||
907 | for grep { defined $_ } @{$colinfo}{@$columns}; | ||||
908 | |||||
909 | for (values %$cols_per_src) { | ||||
910 | my $src = (values %$_)[0]->{-result_source}; | ||||
911 | return 1 if $src->_identifying_column_set($_); | ||||
912 | } | ||||
913 | |||||
914 | return 0; | ||||
915 | } | ||||
916 | |||||
917 | # this is almost similar to _order_by_is_stable, except it takes | ||||
918 | # a single rsrc, and will succeed only if the first portion of the order | ||||
919 | # by is stable. | ||||
920 | # returns that portion as a colinfo hashref on success | ||||
921 | sub _extract_colinfo_of_stable_main_source_order_by_portion { | ||||
922 | my ($self, $attrs) = @_; | ||||
923 | |||||
924 | my $nodes = $self->_find_join_path_to_node($attrs->{from}, $attrs->{alias}); | ||||
925 | |||||
926 | return unless defined $nodes; | ||||
927 | |||||
928 | my @ord_cols = map | ||||
929 | { $_->[0] } | ||||
930 | ( $self->_extract_order_criteria($attrs->{order_by}) ) | ||||
931 | ; | ||||
932 | return unless @ord_cols; | ||||
933 | |||||
934 | my $valid_aliases = { map { $_ => 1 } ( | ||||
935 | $attrs->{from}[0]{-alias}, | ||||
936 | map { values %$_ } @$nodes, | ||||
937 | ) }; | ||||
938 | |||||
939 | my $colinfos = $self->_resolve_column_info($attrs->{from}); | ||||
940 | |||||
941 | my ($colinfos_to_return, $seen_main_src_cols); | ||||
942 | |||||
943 | for my $col (@ord_cols) { | ||||
944 | # if order criteria is unresolvable - there is nothing we can do | ||||
945 | my $colinfo = $colinfos->{$col} or last; | ||||
946 | |||||
947 | # if we reached the end of the allowed aliases - also nothing we can do | ||||
948 | last unless $valid_aliases->{$colinfo->{-source_alias}}; | ||||
949 | |||||
950 | $colinfos_to_return->{$col} = $colinfo; | ||||
951 | |||||
952 | $seen_main_src_cols->{$colinfo->{-colname}} = 1 | ||||
953 | if $colinfo->{-source_alias} eq $attrs->{alias}; | ||||
954 | } | ||||
955 | |||||
956 | # FIXME the condition may be singling out things on its own, so we | ||||
957 | # conceivable could come back wi "stable-ordered by nothing" | ||||
958 | # not confient enough in the parser yet, so punt for the time being | ||||
959 | return unless $seen_main_src_cols; | ||||
960 | |||||
961 | my $main_src_fixed_cols_from_cond = [ $attrs->{where} | ||||
962 | ? ( | ||||
963 | map | ||||
964 | { | ||||
965 | ( $colinfos->{$_} and $colinfos->{$_}{-source_alias} eq $attrs->{alias} ) | ||||
966 | ? $colinfos->{$_}{-colname} | ||||
967 | : () | ||||
968 | } | ||||
969 | keys %{ $self->_extract_fixed_condition_columns($attrs->{where}) } | ||||
970 | ) | ||||
971 | : () | ||||
972 | ]; | ||||
973 | |||||
974 | return $attrs->{result_source}->_identifying_column_set([ | ||||
975 | keys %$seen_main_src_cols, | ||||
976 | @$main_src_fixed_cols_from_cond, | ||||
977 | ]) ? $colinfos_to_return : (); | ||||
978 | } | ||||
979 | |||||
980 | # Attempts to flatten a passed in SQLA condition as much as possible towards | ||||
981 | # a plain hashref, *without* altering its semantics. Required by | ||||
982 | # create/populate being able to extract definitive conditions from preexisting | ||||
983 | # resultset {where} stacks | ||||
984 | # | ||||
985 | # FIXME - while relatively robust, this is still imperfect, one of the first | ||||
986 | # things to tackle with DQ | ||||
987 | # spent 19.3ms (12.7+6.54) within DBIx::Class::Storage::DBIHacks::_collapse_cond which was called 315 times, avg 61µs/call:
# 315 times (12.7ms+6.54ms) by DBIx::Class::Storage::DBIHacks::_extract_fixed_condition_columns at line 1342, avg 61µs/call | ||||
988 | 84 | 36µs | my ($self, $where, $where_is_anded_array) = @_; | ||
989 | |||||
990 | 84 | 18µs | my $fin; | ||
991 | |||||
992 | 84 | 157µs | if (! $where) { | ||
993 | return; | ||||
994 | } | ||||
995 | elsif ($where_is_anded_array or ref $where eq 'HASH') { | ||||
996 | |||||
997 | 84 | 17µs | my @pairs; | ||
998 | |||||
999 | 84 | 75µs | my @pieces = $where_is_anded_array ? @$where : $where; | ||
1000 | 84 | 63µs | while (@pieces) { | ||
1001 | 84 | 39µs | my $chunk = shift @pieces; | ||
1002 | |||||
1003 | 84 | 121µs | if (ref $chunk eq 'HASH') { | ||
1004 | 84 | 422µs | 315 | 360µs | for (sort keys %$chunk) { # spent 360µs making 315 calls to DBIx::Class::Storage::DBIHacks::CORE:sort, avg 1µs/call |
1005 | |||||
1006 | # Match SQLA 1.79 behavior | ||||
1007 | 84 | 27µs | if ($_ eq '') { | ||
1008 | is_literal_value($chunk->{$_}) | ||||
1009 | ? carp 'Hash-pairs consisting of an empty string with a literal are deprecated, use -and => [ $literal ] instead' | ||||
1010 | : $self->throw_exception("Supplying an empty left hand side argument is not supported in hash-pairs") | ||||
1011 | ; | ||||
1012 | } | ||||
1013 | |||||
1014 | 84 | 134µs | push @pairs, $_ => $chunk->{$_}; | ||
1015 | } | ||||
1016 | } | ||||
1017 | elsif (ref $chunk eq 'ARRAY') { | ||||
1018 | push @pairs, -or => $chunk | ||||
1019 | if @$chunk; | ||||
1020 | } | ||||
1021 | elsif ( ! length ref $chunk) { | ||||
1022 | |||||
1023 | # Match SQLA 1.79 behavior | ||||
1024 | $self->throw_exception("Supplying an empty left hand side argument is not supported in array-pairs") | ||||
1025 | if $where_is_anded_array and (! defined $chunk or $chunk eq ''); | ||||
1026 | |||||
1027 | push @pairs, $chunk, shift @pieces; | ||||
1028 | } | ||||
1029 | else { | ||||
1030 | push @pairs, '', $chunk; | ||||
1031 | } | ||||
1032 | } | ||||
1033 | |||||
1034 | 84 | 29µs | return unless @pairs; | ||
1035 | |||||
1036 | 84 | 264µs | 315 | 5.51ms | my @conds = $self->_collapse_cond_unroll_pairs(\@pairs) # spent 5.51ms making 315 calls to DBIx::Class::Storage::DBIHacks::_collapse_cond_unroll_pairs, avg 17µs/call |
1037 | or return; | ||||
1038 | |||||
1039 | # Consolidate various @conds back into something more compact | ||||
1040 | 84 | 93µs | for my $c (@conds) { | ||
1041 | 84 | 99µs | if (ref $c ne 'HASH') { | ||
1042 | push @{$fin->{-and}}, $c; | ||||
1043 | } | ||||
1044 | else { | ||||
1045 | 84 | 278µs | 315 | 92µs | for my $col (sort keys %$c) { # spent 92µs making 315 calls to DBIx::Class::Storage::DBIHacks::CORE:sort, avg 292ns/call |
1046 | |||||
1047 | # consolidate all -and nodes | ||||
1048 | 84 | 502µs | 630 | 395µs | if ($col =~ /^\-and$/i) { # spent 395µs making 630 calls to DBIx::Class::Storage::DBIHacks::CORE:match, avg 627ns/call |
1049 | push @{$fin->{-and}}, | ||||
1050 | ref $c->{$col} eq 'ARRAY' ? @{$c->{$col}} | ||||
1051 | : ref $c->{$col} eq 'HASH' ? %{$c->{$col}} | ||||
1052 | : { $col => $c->{$col} } | ||||
1053 | ; | ||||
1054 | } | ||||
1055 | elsif ($col =~ /^\-/) { | ||||
1056 | push @{$fin->{-and}}, { $col => $c->{$col} }; | ||||
1057 | } | ||||
1058 | elsif (exists $fin->{$col}) { | ||||
1059 | $fin->{$col} = [ -and => map { | ||||
1060 | (ref $_ eq 'ARRAY' and ($_->[0]||'') =~ /^\-and$/i ) | ||||
1061 | ? @{$_}[1..$#$_] | ||||
1062 | : $_ | ||||
1063 | ; | ||||
1064 | } ($fin->{$col}, $c->{$col}) ]; | ||||
1065 | } | ||||
1066 | else { | ||||
1067 | 84 | 79µs | $fin->{$col} = $c->{$col}; | ||
1068 | } | ||||
1069 | } | ||||
1070 | } | ||||
1071 | } | ||||
1072 | } | ||||
1073 | elsif (ref $where eq 'ARRAY') { | ||||
1074 | # we are always at top-level here, it is safe to dump empty *standalone* pieces | ||||
1075 | my $fin_idx; | ||||
1076 | |||||
1077 | for (my $i = 0; $i <= $#$where; $i++ ) { | ||||
1078 | |||||
1079 | # Match SQLA 1.79 behavior | ||||
1080 | $self->throw_exception( | ||||
1081 | "Supplying an empty left hand side argument is not supported in array-pairs" | ||||
1082 | ) if (! defined $where->[$i] or ! length $where->[$i]); | ||||
1083 | |||||
1084 | my $logic_mod = lc ( ($where->[$i] =~ /^(\-(?:and|or))$/i)[0] || '' ); | ||||
1085 | |||||
1086 | if ($logic_mod) { | ||||
1087 | $i++; | ||||
1088 | $self->throw_exception("Unsupported top-level op/arg pair: [ $logic_mod => $where->[$i] ]") | ||||
1089 | unless ref $where->[$i] eq 'HASH' or ref $where->[$i] eq 'ARRAY'; | ||||
1090 | |||||
1091 | my $sub_elt = $self->_collapse_cond({ $logic_mod => $where->[$i] }) | ||||
1092 | or next; | ||||
1093 | |||||
1094 | my @keys = keys %$sub_elt; | ||||
1095 | if ( @keys == 1 and $keys[0] !~ /^\-/ ) { | ||||
1096 | $fin_idx->{ "COL_$keys[0]_" . serialize $sub_elt } = $sub_elt; | ||||
1097 | } | ||||
1098 | else { | ||||
1099 | $fin_idx->{ "SER_" . serialize $sub_elt } = $sub_elt; | ||||
1100 | } | ||||
1101 | } | ||||
1102 | elsif (! length ref $where->[$i] ) { | ||||
1103 | my $sub_elt = $self->_collapse_cond({ @{$where}[$i, $i+1] }) | ||||
1104 | or next; | ||||
1105 | |||||
1106 | $fin_idx->{ "COL_$where->[$i]_" . serialize $sub_elt } = $sub_elt; | ||||
1107 | $i++; | ||||
1108 | } | ||||
1109 | else { | ||||
1110 | $fin_idx->{ "SER_" . serialize $where->[$i] } = $self->_collapse_cond( $where->[$i] ) || next; | ||||
1111 | } | ||||
1112 | } | ||||
1113 | |||||
1114 | if (! $fin_idx) { | ||||
1115 | return; | ||||
1116 | } | ||||
1117 | elsif ( keys %$fin_idx == 1 ) { | ||||
1118 | $fin = (values %$fin_idx)[0]; | ||||
1119 | } | ||||
1120 | else { | ||||
1121 | my @or; | ||||
1122 | |||||
1123 | # at this point everything is at most one level deep - unroll if needed | ||||
1124 | for (sort keys %$fin_idx) { | ||||
1125 | if ( ref $fin_idx->{$_} eq 'HASH' and keys %{$fin_idx->{$_}} == 1 ) { | ||||
1126 | my ($l, $r) = %{$fin_idx->{$_}}; | ||||
1127 | |||||
1128 | if ( | ||||
1129 | ref $r eq 'ARRAY' | ||||
1130 | and | ||||
1131 | ( | ||||
1132 | ( @$r == 1 and $l =~ /^\-and$/i ) | ||||
1133 | or | ||||
1134 | $l =~ /^\-or$/i | ||||
1135 | ) | ||||
1136 | ) { | ||||
1137 | push @or, @$r | ||||
1138 | } | ||||
1139 | |||||
1140 | elsif ( | ||||
1141 | ref $r eq 'HASH' | ||||
1142 | and | ||||
1143 | keys %$r == 1 | ||||
1144 | and | ||||
1145 | $l =~ /^\-(?:and|or)$/i | ||||
1146 | ) { | ||||
1147 | push @or, %$r; | ||||
1148 | } | ||||
1149 | |||||
1150 | else { | ||||
1151 | push @or, $l, $r; | ||||
1152 | } | ||||
1153 | } | ||||
1154 | else { | ||||
1155 | push @or, $fin_idx->{$_}; | ||||
1156 | } | ||||
1157 | } | ||||
1158 | |||||
1159 | $fin->{-or} = \@or; | ||||
1160 | } | ||||
1161 | } | ||||
1162 | else { | ||||
1163 | # not a hash not an array | ||||
1164 | $fin = { -and => [ $where ] }; | ||||
1165 | } | ||||
1166 | |||||
1167 | # unroll single-element -and's | ||||
1168 | 84 | 138µs | while ( | ||
1169 | $fin->{-and} | ||||
1170 | and | ||||
1171 | @{$fin->{-and}} < 2 | ||||
1172 | ) { | ||||
1173 | my $and = delete $fin->{-and}; | ||||
1174 | last if @$and == 0; | ||||
1175 | |||||
1176 | # at this point we have @$and == 1 | ||||
1177 | if ( | ||||
1178 | ref $and->[0] eq 'HASH' | ||||
1179 | and | ||||
1180 | ! grep { exists $fin->{$_} } keys %{$and->[0]} | ||||
1181 | ) { | ||||
1182 | $fin = { | ||||
1183 | %$fin, %{$and->[0]} | ||||
1184 | }; | ||||
1185 | } | ||||
1186 | else { | ||||
1187 | $fin->{-and} = $and; | ||||
1188 | last; | ||||
1189 | } | ||||
1190 | } | ||||
1191 | |||||
1192 | # compress same-column conds found in $fin | ||||
1193 | 168 | 455µs | 315 | 189µs | for my $col ( grep { $_ !~ /^\-/ } keys %$fin ) { # spent 189µs making 315 calls to DBIx::Class::Storage::DBIHacks::CORE:match, avg 600ns/call |
1194 | 84 | 75µs | next unless ref $fin->{$col} eq 'ARRAY' and ($fin->{$col}[0]||'') =~ /^\-and$/i; | ||
1195 | my $val_bag = { map { | ||||
1196 | (! defined $_ ) ? ( UNDEF => undef ) | ||||
1197 | : ( ! length ref $_ or is_plain_value $_ ) ? ( "VAL_$_" => $_ ) | ||||
1198 | : ( ( 'SER_' . serialize $_ ) => $_ ) | ||||
1199 | } @{$fin->{$col}}[1 .. $#{$fin->{$col}}] }; | ||||
1200 | |||||
1201 | if (keys %$val_bag == 1 ) { | ||||
1202 | ($fin->{$col}) = values %$val_bag; | ||||
1203 | } | ||||
1204 | else { | ||||
1205 | $fin->{$col} = [ -and => map { $val_bag->{$_} } sort keys %$val_bag ]; | ||||
1206 | } | ||||
1207 | } | ||||
1208 | |||||
1209 | 84 | 266µs | return keys %$fin ? $fin : (); | ||
1210 | } | ||||
1211 | |||||
1212 | # spent 5.51ms (4.59+913µs) within DBIx::Class::Storage::DBIHacks::_collapse_cond_unroll_pairs which was called 315 times, avg 17µs/call:
# 315 times (4.59ms+913µs) by DBIx::Class::Storage::DBIHacks::_collapse_cond at line 1036, avg 17µs/call | ||||
1213 | 84 | 35µs | my ($self, $pairs) = @_; | ||
1214 | |||||
1215 | 84 | 16µs | my @conds; | ||
1216 | |||||
1217 | 84 | 83µs | while (@$pairs) { | ||
1218 | 84 | 109µs | my ($lhs, $rhs) = splice @$pairs, 0, 2; | ||
1219 | |||||
1220 | 84 | 652µs | 630 | 913µs | if ($lhs eq '') { # spent 913µs making 630 calls to DBIx::Class::Storage::DBIHacks::CORE:match, avg 1µs/call |
1221 | push @conds, $self->_collapse_cond($rhs); | ||||
1222 | } | ||||
1223 | elsif ( $lhs =~ /^\-and$/i ) { | ||||
1224 | push @conds, $self->_collapse_cond($rhs, (ref $rhs eq 'ARRAY')); | ||||
1225 | } | ||||
1226 | elsif ( $lhs =~ /^\-or$/i ) { | ||||
1227 | push @conds, $self->_collapse_cond( | ||||
1228 | (ref $rhs eq 'HASH') ? [ map { $_ => $rhs->{$_} } sort keys %$rhs ] : $rhs | ||||
1229 | ); | ||||
1230 | } | ||||
1231 | else { | ||||
1232 | 84 | 226µs | if (ref $rhs eq 'HASH' and ! keys %$rhs) { | ||
1233 | # FIXME - SQLA seems to be doing... nothing...? | ||||
1234 | } | ||||
1235 | # normalize top level -ident, for saner extract_fixed_condition_columns code | ||||
1236 | elsif (ref $rhs eq 'HASH' and keys %$rhs == 1 and exists $rhs->{-ident}) { | ||||
1237 | push @conds, { $lhs => { '=', $rhs } }; | ||||
1238 | } | ||||
1239 | elsif (ref $rhs eq 'HASH' and keys %$rhs == 1 and exists $rhs->{-value} and is_plain_value $rhs->{-value}) { | ||||
1240 | push @conds, { $lhs => $rhs->{-value} }; | ||||
1241 | } | ||||
1242 | elsif (ref $rhs eq 'HASH' and keys %$rhs == 1 and exists $rhs->{'='}) { | ||||
1243 | if ( length ref $rhs->{'='} and is_literal_value $rhs->{'='} ) { | ||||
1244 | push @conds, { $lhs => $rhs }; | ||||
1245 | } | ||||
1246 | else { | ||||
1247 | for my $p ($self->_collapse_cond_unroll_pairs([ $lhs => $rhs->{'='} ])) { | ||||
1248 | |||||
1249 | # extra sanity check | ||||
1250 | if (keys %$p > 1) { | ||||
1251 | require Data::Dumper::Concise; | ||||
1252 | local $Data::Dumper::Deepcopy = 1; | ||||
1253 | $self->throw_exception( | ||||
1254 | "Internal error: unexpected collapse unroll:" | ||||
1255 | . Data::Dumper::Concise::Dumper { in => { $lhs => $rhs }, out => $p } | ||||
1256 | ); | ||||
1257 | } | ||||
1258 | |||||
1259 | my ($l, $r) = %$p; | ||||
1260 | |||||
1261 | push @conds, ( | ||||
1262 | ! length ref $r | ||||
1263 | or | ||||
1264 | ref $r eq 'HASH' and keys %$rhs == 1 and exists $rhs->{'='} | ||||
1265 | |||||
1266 | or | ||||
1267 | is_plain_value($r) | ||||
1268 | ) | ||||
1269 | ? { $l => $r } | ||||
1270 | : { $l => { '=' => $r } } | ||||
1271 | ; | ||||
1272 | } | ||||
1273 | } | ||||
1274 | } | ||||
1275 | elsif (ref $rhs eq 'ARRAY') { | ||||
1276 | # some of these conditionals encounter multi-values - roll them out using | ||||
1277 | # an unshift, which will cause extra looping in the while{} above | ||||
1278 | if (! @$rhs ) { | ||||
1279 | push @conds, { $lhs => [] }; | ||||
1280 | } | ||||
1281 | elsif ( ($rhs->[0]||'') =~ /^\-(?:and|or)$/i ) { | ||||
1282 | $self->throw_exception("Value modifier not followed by any values: $lhs => [ $rhs->[0] ] ") | ||||
1283 | if @$rhs == 1; | ||||
1284 | |||||
1285 | if( $rhs->[0] =~ /^\-and$/i ) { | ||||
1286 | unshift @$pairs, map { $lhs => $_ } @{$rhs}[1..$#$rhs]; | ||||
1287 | } | ||||
1288 | # if not an AND then it's an OR | ||||
1289 | elsif(@$rhs == 2) { | ||||
1290 | unshift @$pairs, $lhs => $rhs->[1]; | ||||
1291 | } | ||||
1292 | else { | ||||
1293 | push @conds, { $lhs => [ @{$rhs}[1..$#$rhs] ] }; | ||||
1294 | } | ||||
1295 | } | ||||
1296 | elsif (@$rhs == 1) { | ||||
1297 | unshift @$pairs, $lhs => $rhs->[0]; | ||||
1298 | } | ||||
1299 | else { | ||||
1300 | push @conds, { $lhs => $rhs }; | ||||
1301 | } | ||||
1302 | } | ||||
1303 | # unroll func + { -value => ... } | ||||
1304 | elsif ( | ||||
1305 | ref $rhs eq 'HASH' | ||||
1306 | and | ||||
1307 | ( my ($subop) = keys %$rhs ) == 1 | ||||
1308 | and | ||||
1309 | length ref ((values %$rhs)[0]) | ||||
1310 | and | ||||
1311 | my $vref = is_plain_value( (values %$rhs)[0] ) | ||||
1312 | ) { | ||||
1313 | push @conds, { $lhs => { $subop => $$vref } } | ||||
1314 | } | ||||
1315 | else { | ||||
1316 | 84 | 93µs | push @conds, { $lhs => $rhs }; | ||
1317 | } | ||||
1318 | } | ||||
1319 | } | ||||
1320 | |||||
1321 | 84 | 217µs | return @conds; | ||
1322 | } | ||||
1323 | |||||
1324 | # Analyzes a given condition and attempts to extract all columns | ||||
1325 | # with a definitive fixed-condition criteria. Returns a hashref | ||||
1326 | # of k/v pairs suitable to be passed to set_columns(), with a | ||||
1327 | # MAJOR CAVEAT - multi-value (contradictory) equalities are still | ||||
1328 | # represented as a reference to the UNRESOVABLE_CONDITION constant | ||||
1329 | # The reason we do this is that some codepaths only care about the | ||||
1330 | # codition being stable, as opposed to actually making sense | ||||
1331 | # | ||||
1332 | # The normal mode is used to figure out if a resultset is constrained | ||||
1333 | # to a column which is part of a unique constraint, which in turn | ||||
1334 | # allows us to better predict how ordering will behave etc. | ||||
1335 | # | ||||
1336 | # With the optional "consider_nulls" boolean argument, the function | ||||
1337 | # is instead used to infer inambiguous values from conditions | ||||
1338 | # (e.g. the inheritance of resultset conditions on new_result) | ||||
1339 | # | ||||
1340 | # spent 24.7ms (5.41+19.3) within DBIx::Class::Storage::DBIHacks::_extract_fixed_condition_columns which was called 315 times, avg 78µs/call:
# 315 times (5.41ms+19.3ms) by DBIx::Class::ResultSource::_minimal_valueset_satisfying_constraint at line 1581 of DBIx/Class/ResultSource.pm, avg 78µs/call | ||||
1341 | 84 | 56µs | my ($self, $where, $consider_nulls) = @_; | ||
1342 | 84 | 242µs | 315 | 19.3ms | my $where_hash = $self->_collapse_cond($_[1]); # spent 19.3ms making 315 calls to DBIx::Class::Storage::DBIHacks::_collapse_cond, avg 61µs/call |
1343 | |||||
1344 | 84 | 33µs | my $res = {}; | ||
1345 | 84 | 22µs | my ($c, $v); | ||
1346 | 84 | 85µs | for $c (keys %$where_hash) { | ||
1347 | 84 | 15µs | my $vals; | ||
1348 | |||||
1349 | 84 | 281µs | if (!defined ($v = $where_hash->{$c}) ) { | ||
1350 | $vals->{UNDEF} = $v if $consider_nulls | ||||
1351 | } | ||||
1352 | elsif ( | ||||
1353 | ref $v eq 'HASH' | ||||
1354 | and | ||||
1355 | keys %$v == 1 | ||||
1356 | ) { | ||||
1357 | if (exists $v->{-value}) { | ||||
1358 | if (defined $v->{-value}) { | ||||
1359 | $vals->{"VAL_$v->{-value}"} = $v->{-value} | ||||
1360 | } | ||||
1361 | elsif( $consider_nulls ) { | ||||
1362 | $vals->{UNDEF} = $v->{-value}; | ||||
1363 | } | ||||
1364 | } | ||||
1365 | # do not need to check for plain values - _collapse_cond did it for us | ||||
1366 | elsif( | ||||
1367 | length ref $v->{'='} | ||||
1368 | and | ||||
1369 | ( | ||||
1370 | ( ref $v->{'='} eq 'HASH' and keys %{$v->{'='}} == 1 and exists $v->{'='}{-ident} ) | ||||
1371 | or | ||||
1372 | is_literal_value($v->{'='}) | ||||
1373 | ) | ||||
1374 | ) { | ||||
1375 | $vals->{ 'SER_' . serialize $v->{'='} } = $v->{'='}; | ||||
1376 | } | ||||
1377 | } | ||||
1378 | elsif ( | ||||
1379 | ! length ref $v | ||||
1380 | or | ||||
1381 | is_plain_value ($v) | ||||
1382 | ) { | ||||
1383 | $vals->{"VAL_$v"} = $v; | ||||
1384 | } | ||||
1385 | elsif (ref $v eq 'ARRAY' and ($v->[0]||'') eq '-and') { | ||||
1386 | for ( @{$v}[1..$#$v] ) { | ||||
1387 | my $subval = $self->_extract_fixed_condition_columns({ $c => $_ }, 'consider nulls'); # always fish nulls out on recursion | ||||
1388 | next unless exists $subval->{$c}; # didn't find anything | ||||
1389 | $vals->{ | ||||
1390 | ! defined $subval->{$c} ? 'UNDEF' | ||||
1391 | : ( ! length ref $subval->{$c} or is_plain_value $subval->{$c} ) ? "VAL_$subval->{$c}" | ||||
1392 | : ( 'SER_' . serialize $subval->{$c} ) | ||||
1393 | } = $subval->{$c}; | ||||
1394 | } | ||||
1395 | } | ||||
1396 | |||||
1397 | 84 | 249µs | if (keys %$vals == 1) { | ||
1398 | ($res->{$c}) = (values %$vals) | ||||
1399 | unless !$consider_nulls and exists $vals->{UNDEF}; | ||||
1400 | } | ||||
1401 | elsif (keys %$vals > 1) { | ||||
1402 | $res->{$c} = UNRESOLVABLE_CONDITION; | ||||
1403 | } | ||||
1404 | } | ||||
1405 | |||||
1406 | 84 | 265µs | $res; | ||
1407 | } | ||||
1408 | |||||
1409 | 1 | 267µs | 1; # spent 267µs making 1 call to B::Hooks::EndOfScope::XS::__ANON__ | ||
# spent 9.03ms within DBIx::Class::Storage::DBIHacks::CORE:match which was called 4857 times, avg 2µs/call:
# 3282 times (7.54ms+0s) by DBIx::Class::Storage::DBIHacks::_resolve_column_info at line 739, avg 2µs/call
# 630 times (913µs+0s) by DBIx::Class::Storage::DBIHacks::_collapse_cond_unroll_pairs at line 1220, avg 1µs/call
# 630 times (395µs+0s) by DBIx::Class::Storage::DBIHacks::_collapse_cond at line 1048, avg 627ns/call
# 315 times (189µs+0s) by DBIx::Class::Storage::DBIHacks::_collapse_cond at line 1193, avg 600ns/call | |||||
# spent 452µs within DBIx::Class::Storage::DBIHacks::CORE:sort which was called 630 times, avg 717ns/call:
# 315 times (360µs+0s) by DBIx::Class::Storage::DBIHacks::_collapse_cond at line 1004, avg 1µs/call
# 315 times (92µs+0s) by DBIx::Class::Storage::DBIHacks::_collapse_cond at line 1045, avg 292ns/call |