home *** CD-ROM | disk | FTP | other *** search
- /* Convert tree expression to rtl instructions, for GNU compiler.
- Copyright (C) 1988-1991 Free Software Foundation, Inc.
-
- This file is part of GNU CC.
-
- GNU CC is free software; you can redistribute it and/or modify
- it under the terms of the GNU General Public License as published by
- the Free Software Foundation; either version 2, or (at your option)
- any later version.
-
- GNU CC is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
-
- You should have received a copy of the GNU General Public License
- along with GNU CC; see the file COPYING. If not, write to
- the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. */
-
-
- #include "config.h"
- #include "rtl.h"
- #include "tree.h"
- #include "flags.h"
- #include "function.h"
- #include "insn-flags.h"
- #include "insn-codes.h"
- #include "expr.h"
- #include "insn-config.h"
- #include "recog.h"
- #include "output.h"
- #include "gvarargs.h"
- #include "typeclass.h"
-
- #define MIN(x, y) ((x) < (y) ? (x) : (y))
- #define MAX(x, y) ((x) > (y) ? (x) : (y))
-
- /* Decide whether a function's arguments should be processed
- from first to last or from last to first. */
-
- #ifdef STACK_GROWS_DOWNWARD
- #ifdef PUSH_ROUNDING
- #define PUSH_ARGS_REVERSED /* If it's last to first */
- #endif
- #endif
-
- #ifndef STACK_PUSH_CODE
- #ifdef STACK_GROWS_DOWNWARD
- #define STACK_PUSH_CODE PRE_DEC
- #else
- #define STACK_PUSH_CODE PRE_INC
- #endif
- #endif
-
- /* Like STACK_BOUNDARY but in units of bytes, not bits. */
- #define STACK_BYTES (STACK_BOUNDARY / BITS_PER_UNIT)
-
- /* If this is nonzero, we do not bother generating VOLATILE
- around volatile memory references, and we are willing to
- output indirect addresses. If cse is to follow, we reject
- indirect addresses so a useful potential cse is generated;
- if it is used only once, instruction combination will produce
- the same indirect address eventually. */
- int cse_not_expected;
-
- /* Nonzero to generate code for all the subroutines within an
- expression before generating the upper levels of the expression.
- Nowadays this is never zero. */
- int do_preexpand_calls = 1;
-
- /* Number of units that we should eventually pop off the stack.
- These are the arguments to function calls that have already returned. */
- int pending_stack_adjust;
-
- /* Nonzero means stack pops must not be deferred, and deferred stack
- pops must not be output. It is nonzero inside a function call,
- inside a conditional expression, inside a statement expression,
- and in other cases as well. */
- int inhibit_defer_pop;
-
- /* A list of all cleanups which belong to the arguments of
- function calls being expanded by expand_call. */
- tree cleanups_this_call;
-
- /* Nonzero means __builtin_saveregs has already been done in this function.
- The value is the pseudoreg containing the value __builtin_saveregs
- returned. */
- static rtx saveregs_value;
-
- rtx store_expr ();
- static void store_constructor ();
- static rtx store_field ();
- static rtx expand_builtin ();
- static rtx compare ();
- static rtx compare_constants ();
- static rtx do_store_flag ();
- static void preexpand_calls ();
- static rtx expand_increment ();
- static void init_queue ();
-
- void do_pending_stack_adjust ();
- static void do_jump_for_compare ();
- static void do_jump_by_parts_equality ();
- static void do_jump_by_parts_equality_rtx ();
- static void do_jump_by_parts_greater ();
-
- /* MOVE_RATIO is the number of move instructions that is better than
- a block move. */
-
- #ifndef MOVE_RATIO
- #if defined (HAVE_movstrqi) || defined (HAVE_movstrhi) || defined (HAVE_movstrsi)
- #define MOVE_RATIO 2
- #else
- /* A value of around 6 would minimize code size; infinity would minimize
- execution time. */
- #define MOVE_RATIO 15
- #endif
- #endif
-
- /* This is run at the start of compiling a function. */
-
- void
- init_expr ()
- {
- init_queue ();
-
- pending_stack_adjust = 0;
- inhibit_defer_pop = 0;
- cleanups_this_call = 0;
- saveregs_value = 0;
- }
-
- /* Save all variables describing the current status into the structure *P.
- This is used before starting a nested function. */
-
- void
- save_expr_status (p)
- struct function *p;
- {
- /* Instead of saving the postincrement queue, empty it. */
- emit_queue ();
-
- p->pending_stack_adjust = pending_stack_adjust;
- p->inhibit_defer_pop = inhibit_defer_pop;
- p->cleanups_this_call = cleanups_this_call;
- p->saveregs_value = saveregs_value;
-
- pending_stack_adjust = 0;
- inhibit_defer_pop = 0;
- cleanups_this_call = 0;
- saveregs_value = 0;
- }
-
- /* Restore all variables describing the current status from the structure *P.
- This is used after a nested function. */
-
- void
- restore_expr_status (p)
- struct function *p;
- {
- pending_stack_adjust = p->pending_stack_adjust;
- inhibit_defer_pop = p->inhibit_defer_pop;
- cleanups_this_call = p->cleanups_this_call;
- saveregs_value = p->saveregs_value;
- }
-
- /* Manage the queue of increment instructions to be output
- for POSTINCREMENT_EXPR expressions, etc. */
-
- static rtx pending_chain;
-
- /* Queue up to increment (or change) VAR later. BODY says how:
- BODY should be the same thing you would pass to emit_insn
- to increment right away. It will go to emit_insn later on.
-
- The value is a QUEUED expression to be used in place of VAR
- where you want to guarantee the pre-incrementation value of VAR. */
-
- static rtx
- enqueue_insn (var, body)
- rtx var, body;
- {
- pending_chain = gen_rtx (QUEUED, GET_MODE (var),
- var, 0, 0, body, pending_chain);
- return pending_chain;
- }
-
- /* Use protect_from_queue to convert a QUEUED expression
- into something that you can put immediately into an instruction.
- If the queued incrementation has not happened yet,
- protect_from_queue returns the variable itself.
- If the incrementation has happened, protect_from_queue returns a temp
- that contains a copy of the old value of the variable.
-
- Any time an rtx which might possibly be a QUEUED is to be put
- into an instruction, it must be passed through protect_from_queue first.
- QUEUED expressions are not meaningful in instructions.
-
- Do not pass a value through protect_from_queue and then hold
- on to it for a while before putting it in an instruction!
- If the queue is flushed in between, incorrect code will result. */
-
- rtx
- protect_from_queue (x, modify)
- register rtx x;
- int modify;
- {
- register RTX_CODE code = GET_CODE (x);
-
- #if 0 /* A QUEUED can hang around after the queue is forced out. */
- /* Shortcut for most common case. */
- if (pending_chain == 0)
- return x;
- #endif
-
- if (code != QUEUED)
- {
- /* A special hack for read access to (MEM (QUEUED ...))
- to facilitate use of autoincrement.
- Make a copy of the contents of the memory location
- rather than a copy of the address, but not
- if the value is of mode BLKmode. */
- if (code == MEM && GET_MODE (x) != BLKmode
- && GET_CODE (XEXP (x, 0)) == QUEUED && !modify)
- {
- register rtx y = XEXP (x, 0);
- XEXP (x, 0) = QUEUED_VAR (y);
- if (QUEUED_INSN (y))
- {
- register rtx temp = gen_reg_rtx (GET_MODE (x));
- emit_insn_before (gen_move_insn (temp, x),
- QUEUED_INSN (y));
- return temp;
- }
- return x;
- }
- /* Otherwise, recursively protect the subexpressions of all
- the kinds of rtx's that can contain a QUEUED. */
- if (code == MEM)
- XEXP (x, 0) = protect_from_queue (XEXP (x, 0), 0);
- else if (code == PLUS || code == MULT)
- {
- XEXP (x, 0) = protect_from_queue (XEXP (x, 0), 0);
- XEXP (x, 1) = protect_from_queue (XEXP (x, 1), 0);
- }
- return x;
- }
- /* If the increment has not happened, use the variable itself. */
- if (QUEUED_INSN (x) == 0)
- return QUEUED_VAR (x);
- /* If the increment has happened and a pre-increment copy exists,
- use that copy. */
- if (QUEUED_COPY (x) != 0)
- return QUEUED_COPY (x);
- /* The increment has happened but we haven't set up a pre-increment copy.
- Set one up now, and use it. */
- QUEUED_COPY (x) = gen_reg_rtx (GET_MODE (QUEUED_VAR (x)));
- emit_insn_before (gen_move_insn (QUEUED_COPY (x), QUEUED_VAR (x)),
- QUEUED_INSN (x));
- return QUEUED_COPY (x);
- }
-
- /* Return nonzero if X contains a QUEUED expression:
- if it contains anything that will be altered by a queued increment.
- We handle only combinations of MEM, PLUS, MINUS and MULT operators
- since memory addresses generally contain only those. */
-
- static int
- queued_subexp_p (x)
- rtx x;
- {
- register enum rtx_code code = GET_CODE (x);
- switch (code)
- {
- case QUEUED:
- return 1;
- case MEM:
- return queued_subexp_p (XEXP (x, 0));
- case MULT:
- case PLUS:
- case MINUS:
- return queued_subexp_p (XEXP (x, 0))
- || queued_subexp_p (XEXP (x, 1));
- }
- return 0;
- }
-
- /* Perform all the pending incrementations. */
-
- void
- emit_queue ()
- {
- register rtx p;
- while (p = pending_chain)
- {
- QUEUED_INSN (p) = emit_insn (QUEUED_BODY (p));
- pending_chain = QUEUED_NEXT (p);
- }
- }
-
- static void
- init_queue ()
- {
- if (pending_chain)
- abort ();
- }
-
- /* Copy data from FROM to TO, where the machine modes are not the same.
- Both modes may be integer, or both may be floating.
- UNSIGNEDP should be nonzero if FROM is an unsigned type.
- This causes zero-extension instead of sign-extension. */
-
- void
- convert_move (to, from, unsignedp)
- register rtx to, from;
- int unsignedp;
- {
- enum machine_mode to_mode = GET_MODE (to);
- enum machine_mode from_mode = GET_MODE (from);
- int to_real = GET_MODE_CLASS (to_mode) == MODE_FLOAT;
- int from_real = GET_MODE_CLASS (from_mode) == MODE_FLOAT;
- int extending = (int) to_mode > (int) from_mode;
-
- to = protect_from_queue (to, 1);
- from = protect_from_queue (from, 0);
-
- if (to_real != from_real)
- abort ();
-
- if (to_mode == from_mode
- || (from_mode == VOIDmode && CONSTANT_P (from)))
- {
- emit_move_insn (to, from);
- return;
- }
-
- if (to_real)
- {
- #ifdef HAVE_extendsfdf2
- if (HAVE_extendsfdf2 && extending)
- {
- emit_unop_insn (CODE_FOR_extendsfdf2, to, from, UNKNOWN);
- return;
- }
- #endif
- #ifdef HAVE_truncdfsf2
- if (HAVE_truncdfsf2 && ! extending)
- {
- emit_unop_insn (CODE_FOR_truncdfsf2, to, from, UNKNOWN);
- return;
- }
- #endif
- emit_library_call (extending ? extendsfdf2_libfunc : truncdfsf2_libfunc,
- 0, GET_MODE (to), 1,
- from, (extending ? SFmode : DFmode));
- emit_move_insn (to, hard_libcall_value (GET_MODE (to)));
- return;
- }
-
- /* Now both modes are integers. */
-
- if (to_mode == DImode)
- {
- if (unsignedp)
- {
- #ifdef HAVE_zero_extendsidi2
- if (HAVE_zero_extendsidi2 && from_mode == SImode)
- emit_unop_insn (CODE_FOR_zero_extendsidi2, to, from, ZERO_EXTEND);
- else
- #endif
- #ifdef HAVE_zero_extendhidi2
- if (HAVE_zero_extendhidi2 && from_mode == HImode)
- emit_unop_insn (CODE_FOR_zero_extendhidi2, to, from, ZERO_EXTEND);
- else
- #endif
- #ifdef HAVE_zero_extendqidi2
- if (HAVE_zero_extendqidi2 && from_mode == QImode)
- emit_unop_insn (CODE_FOR_zero_extendqidi2, to, from, ZERO_EXTEND);
- else
- #endif
- #ifdef HAVE_zero_extendsidi2
- if (HAVE_zero_extendsidi2)
- {
- convert_move (gen_lowpart (SImode, to), from, unsignedp);
- emit_unop_insn (CODE_FOR_zero_extendsidi2, to,
- gen_lowpart (SImode, to), ZERO_EXTEND);
- }
- else
- #endif
- if (GET_MODE_SIZE (DImode) == 2 * UNITS_PER_WORD)
- {
- rtx seq;
-
- start_sequence ();
- convert_move (operand_subword (to, WORDS_BIG_ENDIAN, 1, DImode),
- from, unsignedp);
- emit_clr_insn (operand_subword (to, 1 - WORDS_BIG_ENDIAN,
- 1, DImode));
-
- seq = gen_sequence ();
- end_sequence ();
-
- emit_no_conflict_block (seq, to, from, 0,
- gen_rtx (ZERO_EXTEND, DImode, from));
- }
- else
- abort ();
- }
- #ifdef HAVE_extendsidi2
- else if (HAVE_extendsidi2 && from_mode == SImode)
- emit_unop_insn (CODE_FOR_extendsidi2, to, from, SIGN_EXTEND);
- #endif
- #ifdef HAVE_extendhidi2
- else if (HAVE_extendhidi2 && from_mode == HImode)
- emit_unop_insn (CODE_FOR_extendhidi2, to, from, SIGN_EXTEND);
- #endif
- #ifdef HAVE_extendqidi2
- else if (HAVE_extendqidi2 && from_mode == QImode)
- emit_unop_insn (CODE_FOR_extendqidi2, to, from, SIGN_EXTEND);
- #endif
- #ifdef HAVE_extendsidi2
- else if (HAVE_extendsidi2)
- {
- convert_move (gen_lowpart (SImode, to), from, unsignedp);
- emit_unop_insn (CODE_FOR_extendsidi2, to,
- gen_lowpart (SImode, to), SIGN_EXTEND);
- }
- #endif
- else if (GET_MODE_SIZE (DImode) == 2 * UNITS_PER_WORD)
- {
- rtx seq;
- rtx lowpart = operand_subword (to, WORDS_BIG_ENDIAN, 1, DImode);
- rtx highpart = operand_subword (to, 1 - WORDS_BIG_ENDIAN,
- 1, DImode);
-
- start_sequence ();
-
- convert_move (lowpart, from, unsignedp);
- #ifdef HAVE_slt
- if (HAVE_slt && insn_operand_mode[(int) CODE_FOR_slt][0] == SImode
- && STORE_FLAG_VALUE == -1)
- {
- rtx temp;
- emit_cmp_insn (lowpart, const0_rtx, NE, 0, 0, 0);
- if (!(*insn_operand_predicate[(int) CODE_FOR_slt][0]) (highpart, SImode))
- temp = gen_reg_rtx (SImode);
- else
- temp = highpart;
- emit_insn (gen_slt (temp));
- if (temp != highpart)
- emit_move_insn (highpart, temp);
- }
- else
- #endif
- {
- rtx temp;
-
- temp = expand_shift (RSHIFT_EXPR, SImode, lowpart,
- size_int (GET_MODE_BITSIZE (SImode) - 1),
- highpart, 0);
- if (temp != highpart)
- emit_move_insn (highpart, temp);
- }
-
- seq = gen_sequence ();
- end_sequence ();
-
- emit_no_conflict_block (seq, to, from, 0,
- gen_rtx (SIGN_EXTEND, DImode, from));
- }
- else
- abort ();
- return;
- }
-
- if (from_mode == DImode && GET_MODE_SIZE (DImode) > UNITS_PER_WORD)
- {
- convert_move (to, gen_lowpart (SImode, from), 0);
- return;
- }
-
- /* Handle pointer conversion */ /* SPEE 900220 */
- if (to_mode == PSImode)
- {
- if (from_mode != SImode)
- from = convert_to_mode (SImode, from, unsignedp);
-
- #ifdef HAVE_truncsipsi
- if (HAVE_truncsipsi)
- {
- emit_unop_insn (CODE_FOR_truncsipsi, to, from, UNKNOWN);
- return;
- }
- #endif /* HAVE_truncsipsi */
- abort ();
- }
-
- if (from_mode == PSImode)
- {
- if (to_mode != SImode)
- {
- from = convert_to_mode (SImode, from, unsignedp);
- from_mode = SImode;
- }
- else
- {
- #ifdef HAVE_extendpsisi
- if (HAVE_extendpsisi)
- {
- emit_unop_insn (CODE_FOR_extendpsisi, to, from, UNKNOWN);
- return;
- }
- #endif /* HAVE_extendpsisi */
- abort ();
- }
- }
-
- /* Now follow all the conversions between integers
- no more than a word long. */
-
- /* For truncation, usually we can just refer to FROM in a narrower mode. */
- if (GET_MODE_BITSIZE (to_mode) < GET_MODE_BITSIZE (from_mode)
- && TRULY_NOOP_TRUNCATION (GET_MODE_BITSIZE (to_mode),
- GET_MODE_BITSIZE (from_mode))
- && ((GET_CODE (from) == MEM
- && ! MEM_VOLATILE_P (from)
- && ! mode_dependent_address_p (XEXP (from, 0)))
- || GET_CODE (from) == REG
- || GET_CODE (from) == SUBREG))
- {
- emit_move_insn (to, gen_lowpart (to_mode, from));
- return;
- }
-
- if (to_mode == SImode && from_mode == HImode)
- {
- if (unsignedp)
- {
- #ifdef HAVE_zero_extendhisi2
- if (HAVE_zero_extendhisi2)
- emit_unop_insn (CODE_FOR_zero_extendhisi2, to, from, ZERO_EXTEND);
- else
- #endif
- abort ();
- }
- else
- {
- #ifdef HAVE_extendhisi2
- if (HAVE_extendhisi2)
- emit_unop_insn (CODE_FOR_extendhisi2, to, from, SIGN_EXTEND);
- else
- #endif
- abort ();
- }
- return;
- }
-
- if (to_mode == SImode && from_mode == QImode)
- {
- if (unsignedp)
- {
- #ifdef HAVE_zero_extendqisi2
- if (HAVE_zero_extendqisi2)
- {
- emit_unop_insn (CODE_FOR_zero_extendqisi2, to, from, ZERO_EXTEND);
- return;
- }
- #endif
- #if defined (HAVE_zero_extendqihi2) && defined (HAVE_extendhisi2)
- if (HAVE_zero_extendqihi2 && HAVE_extendhisi2)
- {
- register rtx temp = gen_reg_rtx (HImode);
- emit_unop_insn (CODE_FOR_zero_extendqihi2, temp, from, ZERO_EXTEND);
- emit_unop_insn (CODE_FOR_extendhisi2, to, temp, SIGN_EXTEND);
- return;
- }
- #endif
- }
- else
- {
- #ifdef HAVE_extendqisi2
- if (HAVE_extendqisi2)
- {
- emit_unop_insn (CODE_FOR_extendqisi2, to, from, SIGN_EXTEND);
- return;
- }
- #endif
- #if defined (HAVE_extendqihi2) && defined (HAVE_extendhisi2)
- if (HAVE_extendqihi2 && HAVE_extendhisi2)
- {
- register rtx temp = gen_reg_rtx (HImode);
- emit_unop_insn (CODE_FOR_extendqihi2, temp, from, SIGN_EXTEND);
- emit_unop_insn (CODE_FOR_extendhisi2, to, temp, SIGN_EXTEND);
- return;
- }
- #endif
- }
- abort ();
- }
-
- if (to_mode == HImode && from_mode == QImode)
- {
- if (unsignedp)
- {
- #ifdef HAVE_zero_extendqihi2
- if (HAVE_zero_extendqihi2)
- {
- emit_unop_insn (CODE_FOR_zero_extendqihi2, to, from, ZERO_EXTEND);
- return;
- }
- #endif
- }
- else
- {
- #ifdef HAVE_extendqihi2
- if (HAVE_extendqihi2)
- {
- emit_unop_insn (CODE_FOR_extendqihi2, to, from, SIGN_EXTEND);
- return;
- }
- #endif
- }
- abort ();
- }
-
- #if 0 /* This seems to be redundant with code 100 lines up. */
-
- /* Now we are truncating an integer to a smaller one.
- If the result is a temporary, we might as well just copy it,
- since only the low-order part of the result needs to be valid
- and it is valid with no change. */
-
- if (GET_CODE (to) == REG)
- {
- if (GET_CODE (from) == REG)
- {
- emit_move_insn (to, gen_lowpart (GET_MODE (to), from));
- return;
- }
- else if (GET_CODE (from) == SUBREG)
- {
- from = copy_rtx (from);
- /* This is safe since FROM is not more than one word. */
- PUT_MODE (from, GET_MODE (to));
- emit_move_insn (to, from);
- return;
- }
- #if !BYTES_BIG_ENDIAN
- else if (GET_CODE (from) == MEM)
- {
- register rtx addr = XEXP (from, 0);
- if (memory_address_p (GET_MODE (to), addr))
- {
- rtx new = gen_rtx (MEM, GET_MODE (to), addr);
- MEM_VOLATILE_P (new) = MEM_VOLATILE_P (from);
- RTX_UNCHANGING_P (new) = RTX_UNCHANGING_P (from);
- MEM_IN_STRUCT_P (new) = MEM_IN_STRUCT_P (from);
- emit_move_insn (to, new);
- return;
- }
- }
- #endif /* not BYTES_BIG_ENDIAN */
- }
- #endif /* 0 */
-
- if (from_mode == SImode && to_mode == HImode)
- {
- #ifdef HAVE_truncsihi2
- if (HAVE_truncsihi2)
- {
- emit_unop_insn (CODE_FOR_truncsihi2, to, from, UNKNOWN);
- return;
- }
- #endif
- convert_move (to, force_reg (from_mode, from), unsignedp);
- return;
- }
-
- if (from_mode == SImode && to_mode == QImode)
- {
- #ifdef HAVE_truncsiqi2
- if (HAVE_truncsiqi2)
- {
- emit_unop_insn (CODE_FOR_truncsiqi2, to, from, UNKNOWN);
- return;
- }
- #endif
- convert_move (to, force_reg (from_mode, from), unsignedp);
- return;
- }
-
- if (from_mode == HImode && to_mode == QImode)
- {
- #ifdef HAVE_trunchiqi2
- if (HAVE_trunchiqi2)
- {
- emit_unop_insn (CODE_FOR_trunchiqi2, to, from, UNKNOWN);
- return;
- }
- #endif
- convert_move (to, force_reg (from_mode, from), unsignedp);
- return;
- }
-
- /* Mode combination is not recognized. */
- abort ();
- }
-
- /* Return an rtx for a value that would result
- from converting X to mode MODE.
- Both X and MODE may be floating, or both integer.
- UNSIGNEDP is nonzero if X is an unsigned value.
- This can be done by referring to a part of X in place
- or by copying to a new temporary with conversion. */
-
- rtx
- convert_to_mode (mode, x, unsignedp)
- enum machine_mode mode;
- rtx x;
- int unsignedp;
- {
- register rtx temp;
-
- x = protect_from_queue (x);
-
- if (mode == GET_MODE (x))
- return x;
-
- /* We can do this with a gen_lowpart if both desired and current modes
- are integer, and this is either a constant integer, a register, or a
- non-volatile MEM. Except for the constant case, we must be narrowing
- the operand. */
-
- if (GET_CODE (x) == CONST_INT
- || (GET_MODE_CLASS (mode) == MODE_INT
- && GET_MODE_CLASS (GET_MODE (x)) == MODE_INT
- && (GET_CODE (x) == CONST_DOUBLE
- || (GET_MODE_SIZE (mode) <= GET_MODE_SIZE (GET_MODE (x))
- && ((GET_CODE (x) == MEM && ! MEM_VOLATILE_P (x))
- || GET_CODE (x) == REG)))))
- return gen_lowpart (mode, x);
-
- temp = gen_reg_rtx (mode);
- convert_move (temp, x, unsignedp);
- return temp;
- }
-
- /* Generate several move instructions to copy LEN bytes
- from block FROM to block TO. (These are MEM rtx's with BLKmode).
- The caller must pass FROM and TO
- through protect_from_queue before calling.
- ALIGN (in bytes) is maximum alignment we can assume. */
-
- struct move_by_pieces
- {
- rtx to;
- rtx to_addr;
- int autinc_to;
- int explicit_inc_to;
- rtx from;
- rtx from_addr;
- int autinc_from;
- int explicit_inc_from;
- int len;
- int offset;
- int reverse;
- };
-
- static void move_by_pieces_1 ();
- static int move_by_pieces_ninsns ();
-
- static void
- move_by_pieces (to, from, len, align)
- rtx to, from;
- int len, align;
- {
- struct move_by_pieces data;
- rtx to_addr = XEXP (to, 0), from_addr = XEXP (from, 0);
-
- data.offset = 0;
- data.to_addr = to_addr;
- data.from_addr = from_addr;
- data.to = to;
- data.from = from;
- data.autinc_to
- = (GET_CODE (to_addr) == PRE_INC || GET_CODE (to_addr) == PRE_DEC
- || GET_CODE (to_addr) == POST_INC || GET_CODE (to_addr) == POST_DEC);
- data.autinc_from
- = (GET_CODE (from_addr) == PRE_INC || GET_CODE (from_addr) == PRE_DEC
- || GET_CODE (from_addr) == POST_INC
- || GET_CODE (from_addr) == POST_DEC);
-
- data.explicit_inc_from = 0;
- data.explicit_inc_to = 0;
- data.reverse
- = (GET_CODE (to_addr) == PRE_DEC || GET_CODE (to_addr) == POST_DEC);
- if (data.reverse) data.offset = len;
- data.len = len;
-
- /* If copying requires more than two move insns,
- copy addresses to registers (to make displacements shorter)
- and use post-increment if available. */
- if (!(data.autinc_from && data.autinc_to)
- && move_by_pieces_ninsns (len, align) > 2)
- {
- #ifdef HAVE_PRE_DECREMENT
- if (data.reverse && ! data.autinc_from)
- {
- data.from_addr = copy_addr_to_reg (plus_constant (from_addr, len));
- data.autinc_from = 1;
- data.explicit_inc_from = -1;
- }
- #endif
- #ifdef HAVE_POST_INCREMENT
- if (! data.autinc_from)
- {
- data.from_addr = copy_addr_to_reg (from_addr);
- data.autinc_from = 1;
- data.explicit_inc_from = 1;
- }
- #endif
- if (!data.autinc_from && CONSTANT_P (from_addr))
- data.from_addr = copy_addr_to_reg (from_addr);
- #ifdef HAVE_PRE_DECREMENT
- if (data.reverse && ! data.autinc_to)
- {
- data.to_addr = copy_addr_to_reg (plus_constant (to_addr, len));
- data.autinc_to = 1;
- data.explicit_inc_to = -1;
- }
- #endif
- #ifdef HAVE_POST_INCREMENT
- if (! data.reverse && ! data.autinc_to)
- {
- data.to_addr = copy_addr_to_reg (to_addr);
- data.autinc_to = 1;
- data.explicit_inc_to = 1;
- }
- #endif
- if (!data.autinc_to && CONSTANT_P (to_addr))
- data.to_addr = copy_addr_to_reg (to_addr);
- }
-
- #if defined (STRICT_ALIGNMENT) || defined (SLOW_UNALIGNED_ACCESS)
- if (align > MOVE_MAX || align >= BIGGEST_ALIGNMENT / BITS_PER_UNIT)
- align = MOVE_MAX;
- #else
- align = MOVE_MAX;
- #endif
-
- #ifdef HAVE_movti
- if (HAVE_movti && align >= GET_MODE_SIZE (TImode))
- move_by_pieces_1 (gen_movti, TImode, &data);
- #endif
- #ifdef HAVE_movdi
- if (HAVE_movdi && align >= GET_MODE_SIZE (DImode))
- move_by_pieces_1 (gen_movdi, DImode, &data);
- #endif
- #ifdef HAVE_movsi
- if (align >= GET_MODE_SIZE (SImode))
- move_by_pieces_1 (gen_movsi, SImode, &data);
- #endif
- #ifdef HAVE_movhi
- if (HAVE_movhi && align >= GET_MODE_SIZE (HImode))
- move_by_pieces_1 (gen_movhi, HImode, &data);
- #endif
- #ifdef HAVE_movqi
- if (HAVE_movqi)
- move_by_pieces_1 (gen_movqi, QImode, &data);
- #endif
- /* The above should have handled everything. */
- if (data.len != 0)
- abort ();
- }
-
- /* Return number of insns required to move L bytes by pieces.
- ALIGN (in bytes) is maximum alignment we can assume. */
-
- static int
- move_by_pieces_ninsns (l, align)
- unsigned int l;
- int align;
- {
- register int n_insns = 0;
-
- #if defined (STRICT_ALIGNMENT) || defined (SLOW_UNALIGNED_ACCESS)
- if (align > MOVE_MAX || align >= BIGGEST_ALIGNMENT / BITS_PER_UNIT)
- align = MOVE_MAX;
- #else
- align = MOVE_MAX;
- #endif
-
- #ifdef HAVE_movti
- if (HAVE_movti && align >= GET_MODE_SIZE (TImode))
- n_insns += l / GET_MODE_SIZE (TImode), l %= GET_MODE_SIZE (TImode);
- #endif
- #ifdef HAVE_movdi
- if (HAVE_movdi && align >= GET_MODE_SIZE (DImode))
- n_insns += l / GET_MODE_SIZE (DImode), l %= GET_MODE_SIZE (DImode);
- #endif
- #ifdef HAVE_movsi
- if (HAVE_movsi && align >= GET_MODE_SIZE (SImode))
- n_insns += l / GET_MODE_SIZE (SImode), l %= GET_MODE_SIZE (SImode);
- #endif
- #ifdef HAVE_movhi
- if (HAVE_movhi && align >= GET_MODE_SIZE (HImode))
- n_insns += l / GET_MODE_SIZE (HImode), l %= GET_MODE_SIZE (HImode);
- #endif
- n_insns += l;
-
- return n_insns;
- }
-
- /* Subroutine of move_by_pieces. Move as many bytes as appropriate
- with move instructions for mode MODE. GENFUN is the gen_... function
- to make a move insn for that mode. DATA has all the other info. */
-
- static void
- move_by_pieces_1 (genfun, mode, data)
- rtx (*genfun) ();
- enum machine_mode mode;
- struct move_by_pieces *data;
- {
- register int size = GET_MODE_SIZE (mode);
- register rtx to1, from1;
-
- while (data->len >= size)
- {
- if (data->reverse) data->offset -= size;
-
- to1 = (data->autinc_to
- ? gen_rtx (MEM, mode, data->to_addr)
- : change_address (data->to, mode,
- plus_constant (data->to_addr, data->offset)));
- from1 =
- (data->autinc_from
- ? gen_rtx (MEM, mode, data->from_addr)
- : change_address (data->from, mode,
- plus_constant (data->from_addr, data->offset)));
-
- #ifdef HAVE_PRE_DECREMENT
- if (data->explicit_inc_to < 0)
- emit_insn (gen_sub2_insn (data->to_addr,
- gen_rtx (CONST_INT, VOIDmode, size)));
- if (data->explicit_inc_from < 0)
- emit_insn (gen_sub2_insn (data->from_addr,
- gen_rtx (CONST_INT, VOIDmode, size)));
- #endif
-
- emit_insn ((*genfun) (to1, from1));
- #ifdef HAVE_POST_INCREMENT
- if (data->explicit_inc_to > 0)
- emit_insn (gen_add2_insn (data->to_addr,
- gen_rtx (CONST_INT, VOIDmode, size)));
- if (data->explicit_inc_from > 0)
- emit_insn (gen_add2_insn (data->from_addr,
- gen_rtx (CONST_INT, VOIDmode, size)));
- #endif
-
- if (! data->reverse) data->offset += size;
-
- data->len -= size;
- }
- }
-
- /* Emit code to move a block Y to a block X.
- This may be done with string-move instructions,
- with multiple scalar move instructions, or with a library call.
-
- Both X and Y must be MEM rtx's (perhaps inside VOLATILE)
- with mode BLKmode.
- SIZE is an rtx that says how long they are.
- ALIGN is the maximum alignment we can assume they have,
- measured in bytes. */
-
- void
- emit_block_move (x, y, size, align)
- rtx x, y;
- rtx size;
- int align;
- {
- if (GET_MODE (x) != BLKmode)
- abort ();
-
- if (GET_MODE (y) != BLKmode)
- abort ();
-
- x = protect_from_queue (x, 1);
- y = protect_from_queue (y, 0);
-
- if (GET_CODE (x) != MEM)
- abort ();
- if (GET_CODE (y) != MEM)
- abort ();
- if (size == 0)
- abort ();
-
- if (GET_CODE (size) == CONST_INT
- && (move_by_pieces_ninsns ((unsigned) INTVAL (size), align)
- < MOVE_RATIO))
- move_by_pieces (x, y, INTVAL (size), align);
- else
- {
- /* Try the most limited insn first, because there's no point
- including more than one in the machine description unless
- the more limited one has some advantage. */
- #ifdef HAVE_movstrqi
- if (HAVE_movstrqi
- && GET_CODE (size) == CONST_INT
- && ((unsigned) INTVAL (size)
- < (1 << (GET_MODE_BITSIZE (QImode) - 1))))
- {
- rtx insn = gen_movstrqi (x, y, size,
- gen_rtx (CONST_INT, VOIDmode, align));
- if (insn)
- {
- emit_insn (insn);
- return;
- }
- }
- #endif
- #ifdef HAVE_movstrhi
- if (HAVE_movstrhi
- && GET_CODE (size) == CONST_INT
- && ((unsigned) INTVAL (size)
- < (1 << (GET_MODE_BITSIZE (HImode) - 1))))
- {
- rtx insn = gen_movstrhi (x, y, size,
- gen_rtx (CONST_INT, VOIDmode, align));
- if (insn)
- {
- emit_insn (insn);
- return;
- }
- }
- #endif
- #ifdef HAVE_movstrsi
- if (HAVE_movstrsi)
- {
- rtx insn = gen_movstrsi (x, y, size,
- gen_rtx (CONST_INT, VOIDmode, align));
- if (insn)
- {
- emit_insn (insn);
- return;
- }
- }
- #endif
-
- #ifdef TARGET_MEM_FUNCTIONS
- emit_library_call (memcpy_libfunc, 0,
- VOIDmode, 3, XEXP (x, 0), Pmode,
- XEXP (y, 0), Pmode,
- size, Pmode);
- #else
- emit_library_call (bcopy_libfunc, 0,
- VOIDmode, 3, XEXP (y, 0), Pmode,
- XEXP (x, 0), Pmode,
- size, Pmode);
- #endif
- }
- }
-
- /* Copy all or part of a value X into registers starting at REGNO.
- The number of registers to be filled is NREGS. */
-
- void
- move_block_to_reg (regno, x, nregs, mode)
- int regno;
- rtx x;
- int nregs;
- enum machine_mode mode;
- {
- enum machine_mode submode = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- int i;
- rtx pat, last;
-
- if (! LEGITIMATE_CONSTANT_P (x))
- x = validize_mem (force_const_mem (mode, x));
-
- /* See if the machine can do this with a load multiple insn. */
- #ifdef HAVE_load_multiple
- last = get_last_insn ();
- pat = gen_load_multiple (gen_rtx (REG, submode, regno), x,
- gen_rtx (CONST_INT, VOIDmode, nregs));
- if (pat)
- {
- emit_insn (pat);
- return;
- }
- else
- delete_insns_since (last);
- #endif
-
- for (i = 0; i < nregs; i++)
- emit_move_insn (gen_rtx (REG, submode, regno + i),
- operand_subword_force (x, i), mode);
- }
-
- /* Copy all or part of a BLKmode value X out of registers starting at REGNO.
- The number of registers to be filled is NREGS. */
-
- void
- move_block_from_reg (regno, x, nregs)
- int regno;
- rtx x;
- int nregs;
- {
- enum machine_mode submode = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- int i;
- rtx pat, last;
-
- /* See if the machine can do this with a store multiple insn. */
- #ifdef HAVE_store_multiple
- last = get_last_insn ();
- pat = gen_store_multiple (x, gen_rtx (REG, submode, regno),
- gen_rtx (CONST_INT, VOIDmode, nregs));
- if (pat)
- {
- emit_insn (pat);
- return;
- }
- else
- delete_insns_since (last);
- #endif
-
- for (i = 0; i < nregs; i++)
- emit_move_insn (operand_subword (x, i, 1, BLKmode),
- gen_rtx (REG, submode, regno + i));
- }
-
- /* Mark NREGS consecutive regs, starting at REGNO, as being live now. */
-
- void
- use_regs (regno, nregs)
- int regno;
- int nregs;
- {
- enum machine_mode submode = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- int i;
-
- for (i = 0; i < nregs; i++)
- emit_insn (gen_rtx (USE, VOIDmode, gen_rtx (REG, submode, regno + i)));
- }
-
- /* Write zeros through the storage of OBJECT.
- If OBJECT has BLKmode, SIZE is its length in bytes. */
-
- void
- clear_storage (object, size)
- rtx object;
- int size;
- {
- if (GET_MODE (object) == BLKmode)
- {
- #ifdef TARGET_MEM_FUNCTIONS
- emit_library_call (memset_libfunc, 0,
- VOIDmode, 3,
- XEXP (object, 0), Pmode, const0_rtx, Pmode,
- gen_rtx (CONST_INT, VOIDmode, size), Pmode);
- #else
- emit_library_call (bzero_libfunc, 0,
- VOIDmode, 2,
- XEXP (object, 0), Pmode,
- gen_rtx (CONST_INT, VOIDmode, size), Pmode);
- #endif
- }
- else
- emit_move_insn (object, const0_rtx);
- }
-
- /* Generate code to copy Y into X.
- Both Y and X must have the same mode, except that
- Y can be a constant with VOIDmode.
- This mode cannot be BLKmode; use emit_block_move for that.
-
- Return the last instruction emitted. */
-
- rtx
- emit_move_insn (x, y)
- rtx x, y;
- {
- enum machine_mode mode = GET_MODE (x);
- int i;
-
- x = protect_from_queue (x, 1);
- y = protect_from_queue (y, 0);
-
- if (mode == BLKmode || (GET_MODE (y) != mode && GET_MODE (y) != VOIDmode))
- abort ();
-
- if (CONSTANT_P (y) && ! LEGITIMATE_CONSTANT_P (y))
- y = force_const_mem (mode, y);
-
- /* If X or Y are memory references, verify that their addresses are valid
- for the machine. */
- if (GET_CODE (x) == MEM
- && ((! memory_address_p (GET_MODE (x), XEXP (x, 0))
- && ! push_operand (x, GET_MODE (x)))
- || (flag_force_addr
- && CONSTANT_ADDRESS_P (XEXP (x, 0)))))
- x = change_address (x, VOIDmode, XEXP (x, 0));
-
- if (GET_CODE (y) == MEM
- && (! memory_address_p (GET_MODE (y), XEXP (y, 0))
- || (flag_force_addr
- && CONSTANT_ADDRESS_P (XEXP (y, 0)))))
- y = change_address (y, VOIDmode, XEXP (y, 0));
-
- if (mode == BLKmode)
- abort ();
-
- if (mov_optab->handlers[(int) mode].insn_code != CODE_FOR_nothing)
- return
- emit_insn (GEN_FCN (mov_optab->handlers[(int) mode].insn_code) (x, y));
-
- /* This will handle any multi-word mode that lacks a move_insn pattern.
- However, you will get better code if you define such patterns,
- even if they must turn into multiple assembler instructions. */
- else if (GET_MODE_SIZE (mode) >= UNITS_PER_WORD)
- for (i = 0;
- i < (GET_MODE_SIZE (mode) + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD;
- i++)
- {
- rtx ypart = operand_subword (y, i, 1, mode);
-
- /* If we can't get a part of Y, put Y into memory if it is a constant.
- If we still can't get a part of Y, abort. */
- if (ypart == 0 && CONSTANT_P (y))
- {
- y = force_const_mem (mode, y);
- ypart = operand_subword (y, i, 1, mode);
- }
-
- if (ypart == 0)
- abort ();
-
- emit_move_insn (operand_subword (x, i, 1, mode), ypart);
- }
- else
- abort ();
- }
-
- /* Pushing data onto the stack. */
-
- /* Push a block of length SIZE (perhaps variable)
- and return an rtx to address the beginning of the block.
- Note that it is not possible for the value returned to be a QUEUED.
- The value may be virtual_outgoing_args_rtx.
-
- EXTRA is the number of bytes of padding to push in addition to SIZE.
- BELOW nonzero means this padding comes at low addresses;
- otherwise, the padding comes at high addresses. */
-
- rtx
- push_block (size, extra, below)
- rtx size;
- int extra, below;
- {
- register rtx temp;
- if (CONSTANT_P (size))
- anti_adjust_stack (plus_constant (size, extra));
- else if (GET_CODE (size) == REG && extra == 0)
- anti_adjust_stack (size);
- else
- {
- rtx temp = copy_to_mode_reg (Pmode, size);
- if (extra != 0)
- temp = expand_binop (Pmode, add_optab,
- temp,
- gen_rtx (CONST_INT, VOIDmode, extra),
- temp, 0, OPTAB_LIB_WIDEN);
- anti_adjust_stack (temp);
- }
-
- #ifdef STACK_GROWS_DOWNWARD
- temp = virtual_outgoing_args_rtx;
- if (extra != 0 && below)
- temp = plus_constant (temp, extra);
- #else
- if (GET_CODE (size) == CONST_INT)
- temp = plus_constant (virtual_outgoing_args_rtx,
- - INTVAL (size) - (below ? 0 : extra));
- else if (extra != 0 && !below)
- temp = gen_rtx (PLUS, Pmode, virtual_outgoing_args_rtx,
- negate_rtx (Pmode, plus_constant (size, extra)));
- else
- temp = gen_rtx (PLUS, Pmode, virtual_outgoing_args_rtx,
- negate_rtx (Pmode, size));
- #endif
-
- return memory_address (QImode, temp);
- }
-
- static rtx
- gen_push_operand ()
- {
- return gen_rtx (STACK_PUSH_CODE, Pmode, stack_pointer_rtx);
- }
-
- /* Generate code to push X onto the stack, assuming it has mode MODE and
- type TYPE.
- MODE is redundant except when X is a CONST_INT (since they don't
- carry mode info).
- SIZE is an rtx for the size of data to be copied (in bytes),
- needed only if X is BLKmode.
-
- ALIGN (in bytes) is maximum alignment we can assume.
-
- If PARTIAL is nonzero, then copy that many of the first words
- of X into registers starting with REG, and push the rest of X.
- The amount of space pushed is decreased by PARTIAL words,
- rounded *down* to a multiple of PARM_BOUNDARY.
- REG must be a hard register in this case.
-
- EXTRA is the amount in bytes of extra space to leave next to this arg.
- This is ignored if an argument block has already been allocted.
-
- On a machine that lacks real push insns, ARGS_ADDR is the address of
- the bottom of the argument block for this call. We use indexing off there
- to store the arg. On machines with push insns, ARGS_ADDR is 0 when a
- argument block has not been preallocated.
-
- ARGS_SO_FAR is the size of args previously pushed for this call. */
-
- void
- emit_push_insn (x, mode, type, size, align, partial, reg, extra,
- args_addr, args_so_far)
- register rtx x;
- enum machine_mode mode;
- tree type;
- rtx size;
- int align;
- int partial;
- rtx reg;
- int extra;
- rtx args_addr;
- rtx args_so_far;
- {
- rtx xinner;
- enum direction stack_direction
- #ifdef STACK_GROWS_DOWNWARD
- = downward;
- #else
- = upward;
- #endif
-
- /* Decide where to pad the argument: `downward' for below,
- `upward' for above, or `none' for don't pad it.
- Default is below for small data on big-endian machines; else above. */
- enum direction where_pad = FUNCTION_ARG_PADDING (mode, type);
-
- /* Invert direction if stack is post-update. */
- if (STACK_PUSH_CODE == POST_INC || STACK_PUSH_CODE == POST_DEC)
- if (where_pad != none)
- where_pad = (where_pad == downward ? upward : downward);
-
- xinner = x = protect_from_queue (x, 0);
-
- if (mode == BLKmode)
- {
- /* Copy a block into the stack, entirely or partially. */
-
- register rtx temp;
- int used = partial * UNITS_PER_WORD;
- int offset = used % (PARM_BOUNDARY / BITS_PER_UNIT);
- int skip;
-
- if (size == 0)
- abort ();
-
- used -= offset;
-
- /* USED is now the # of bytes we need not copy to the stack
- because registers will take care of them. */
-
- if (partial != 0)
- xinner = change_address (xinner, BLKmode,
- plus_constant (XEXP (xinner, 0), used));
-
- /* If the partial register-part of the arg counts in its stack size,
- skip the part of stack space corresponding to the registers.
- Otherwise, start copying to the beginning of the stack space,
- by setting SKIP to 0. */
- #ifndef REG_PARM_STACK_SPACE
- skip = 0;
- #else
- skip = used;
- #endif
-
- #ifdef PUSH_ROUNDING
- /* Do it with several push insns if that doesn't take lots of insns
- and if there is no difficulty with push insns that skip bytes
- on the stack for alignment purposes. */
- if (args_addr == 0
- && GET_CODE (size) == CONST_INT
- && skip == 0
- && (move_by_pieces_ninsns ((unsigned) INTVAL (size) - used, align)
- < MOVE_RATIO)
- #if defined (STRICT_ALIGNMENT) || defined (SLOW_UNALIGNED_ACCESS)
- /* Here we avoid the case of a structure whose weak alignment
- forces many pushes of a small amount of data,
- and such small pushes do rounding that causes trouble. */
- && (align >= BIGGEST_ALIGNMENT / BITS_PER_UNIT
- || PUSH_ROUNDING (align) == align)
- #endif
- && PUSH_ROUNDING (INTVAL (size)) == INTVAL (size))
- {
- /* Push padding now if padding above and stack grows down,
- or if padding below and stack grows up.
- But if space already allocated, this has already been done. */
- if (extra && args_addr == 0
- && where_pad != none && where_pad != stack_direction)
- anti_adjust_stack (gen_rtx (CONST_INT, VOIDmode, extra));
-
- move_by_pieces (gen_rtx (MEM, BLKmode, gen_push_operand ()), xinner,
- INTVAL (size) - used, align);
- }
- else
- #endif /* PUSH_ROUNDING */
- {
- /* Otherwise make space on the stack and copy the data
- to the address of that space. */
-
- /* Deduct words put into registers from the size we must copy. */
- if (partial != 0)
- {
- if (GET_CODE (size) == CONST_INT)
- size = gen_rtx (CONST_INT, VOIDmode, INTVAL (size) - used);
- else
- size = expand_binop (GET_MODE (size), sub_optab, size,
- gen_rtx (CONST_INT, VOIDmode, used),
- 0, 0, OPTAB_LIB_WIDEN);
- }
-
- /* Get the address of the stack space.
- In this case, we do not deal with EXTRA separately.
- A single stack adjust will do. */
- if (! args_addr)
- {
- temp = push_block (size, extra, where_pad == downward);
- extra = 0;
- }
- else if (GET_CODE (args_so_far) == CONST_INT)
- temp = memory_address (BLKmode,
- plus_constant (args_addr,
- skip + INTVAL (args_so_far)));
- else
- temp = memory_address (BLKmode,
- plus_constant (gen_rtx (PLUS, Pmode,
- args_addr, args_so_far),
- skip));
-
- /* TEMP is the address of the block. Copy the data there. */
- if (GET_CODE (size) == CONST_INT
- && (move_by_pieces_ninsns ((unsigned) INTVAL (size), align)
- < MOVE_RATIO))
- {
- move_by_pieces (gen_rtx (MEM, BLKmode, temp), xinner,
- INTVAL (size), align);
- goto ret;
- }
- /* Try the most limited insn first, because there's no point
- including more than one in the machine description unless
- the more limited one has some advantage. */
- #ifdef HAVE_movstrqi
- if (HAVE_movstrqi
- && GET_CODE (size) == CONST_INT
- && ((unsigned) INTVAL (size)
- < (1 << (GET_MODE_BITSIZE (QImode) - 1))))
- {
- emit_insn (gen_movstrqi (gen_rtx (MEM, BLKmode, temp),
- xinner, size,
- gen_rtx (CONST_INT, VOIDmode, align)));
- goto ret;
- }
- #endif
- #ifdef HAVE_movstrhi
- if (HAVE_movstrhi
- && GET_CODE (size) == CONST_INT
- && ((unsigned) INTVAL (size)
- < (1 << (GET_MODE_BITSIZE (HImode) - 1))))
- {
- emit_insn (gen_movstrhi (gen_rtx (MEM, BLKmode, temp),
- xinner, size,
- gen_rtx (CONST_INT, VOIDmode, align)));
- goto ret;
- }
- #endif
- #ifdef HAVE_movstrsi
- if (HAVE_movstrsi)
- {
- emit_insn (gen_movstrsi (gen_rtx (MEM, BLKmode, temp),
- xinner, size,
- gen_rtx (CONST_INT, VOIDmode, align)));
- goto ret;
- }
- #endif
-
- #ifndef ACCUMULATE_OUTGOING_ARGS
- /* If the source is referenced relative to the stack pointer,
- copy it to another register to stabilize it. We do not need
- to do this if we know that we won't be changing sp. */
-
- if (reg_mentioned_p (virtual_stack_dynamic_rtx, temp)
- || reg_mentioned_p (virtual_outgoing_args_rtx, temp))
- temp = copy_to_reg (temp);
- #endif
-
- /* Make inhibit_defer_pop nonzero around the library call
- to force it to pop the bcopy-arguments right away. */
- NO_DEFER_POP;
- #ifdef TARGET_MEM_FUNCTIONS
- emit_library_call (memcpy_libfunc, 0,
- VOIDmode, 3, temp, Pmode, XEXP (xinner, 0), Pmode,
- size, Pmode);
- #else
- emit_library_call (bcopy_libfunc, 0,
- VOIDmode, 3, XEXP (xinner, 0), Pmode, temp, Pmode,
- size, Pmode);
- #endif
- OK_DEFER_POP;
- }
- }
- else if (partial > 0)
- {
- /* Scalar partly in registers. */
-
- int size = GET_MODE_SIZE (mode) / UNITS_PER_WORD;
- int i;
- int not_stack;
- /* # words of start of argument
- that we must make space for but need not store. */
- int offset = partial % (PARM_BOUNDARY / BITS_PER_WORD);
- int args_offset = INTVAL (args_so_far);
- int skip;
- enum machine_mode submode = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
-
- /* Push padding now if padding above and stack grows down,
- or if padding below and stack grows up.
- But if space already allocated, this has already been done. */
- if (extra && args_addr == 0
- && where_pad != none && where_pad != stack_direction)
- anti_adjust_stack (gen_rtx (CONST_INT, VOIDmode, extra));
-
- /* If we make space by pushing it, we might as well push
- the real data. Otherwise, we can leave OFFSET nonzero
- and leave the space uninitialized. */
- if (args_addr == 0)
- offset = 0;
-
- /* Now NOT_STACK gets the number of words that we don't need to
- allocate on the stack. */
- not_stack = partial - offset;
-
- /* If the partial register-part of the arg counts in its stack size,
- skip the part of stack space corresponding to the registers.
- Otherwise, start copying to the beginning of the stack space,
- by setting SKIP to 0. */
- #ifndef REG_PARM_STACK_SPACE
- skip = 0;
- #else
- skip = not_stack;
- #endif
-
- if (! LEGITIMATE_CONSTANT_P (x))
- x = validize_mem (force_const_mem (mode, x));
-
- /* If X is a hard register in a non-integer mode, copy it into a pseudo.
- SUBREGs of such registers are not allowed. Similarly if it is a
- constant and the host and target wordsizes aren't the same. */
- if ((GET_CODE (x) == REG && REGNO (x) < FIRST_PSEUDO_REGISTER
- && GET_MODE_CLASS (GET_MODE (x)) != MODE_INT)
- #if HOST_BITS_PER_INT != BITS_PER_WORD
- || CONSTANT_P (x)
- #endif
- )
- x = copy_to_reg (x);
-
- /* Loop over all the words allocated on the stack for this arg. */
- /* We can do it by words, because any scalar bigger than a word
- has a size a multiple of a word. */
- #ifndef PUSH_ARGS_REVERSED
- for (i = not_stack; i < size; i++)
- #else
- for (i = size - 1; i >= not_stack; i--)
- #endif
- if (i >= not_stack + offset)
- emit_push_insn (operand_subword_force (x, i, mode),
- submode, 0, 0, align, 0, 0, 0, args_addr,
- gen_rtx (CONST_INT, VOIDmode,
- args_offset + ((i - not_stack + skip)
- * UNITS_PER_WORD)));
- }
- else
- {
- rtx addr;
-
- /* Push padding now if padding above and stack grows down,
- or if padding below and stack grows up.
- But if space already allocated, this has already been done. */
- if (extra && args_addr == 0
- && where_pad != none && where_pad != stack_direction)
- anti_adjust_stack (gen_rtx (CONST_INT, VOIDmode, extra));
-
- #ifdef PUSH_ROUNDING
- if (args_addr == 0)
- addr = gen_push_operand ();
- else
- #endif
- if (GET_CODE (args_so_far) == CONST_INT)
- addr
- = memory_address (mode,
- plus_constant (args_addr, INTVAL (args_so_far)));
- else
- addr = memory_address (mode, gen_rtx (PLUS, Pmode, args_addr,
- args_so_far));
-
- emit_move_insn (gen_rtx (MEM, mode, addr), x);
- }
-
- ret:
- /* If part should go in registers, copy that part
- into the appropriate registers. Do this now, at the end,
- since mem-to-mem copies above may do function calls. */
- if (partial > 0)
- move_block_to_reg (REGNO (reg), x, partial, mode);
-
- if (extra && args_addr == 0 && where_pad == stack_direction)
- anti_adjust_stack (gen_rtx (CONST_INT, VOIDmode, extra));
- }
-
- /* Output a library call to function FUN (a SYMBOL_REF rtx)
- (emitting the queue unless NO_QUEUE is nonzero),
- for a value of mode OUTMODE,
- with NARGS different arguments, passed as alternating rtx values
- and machine_modes to convert them to.
- The rtx values should have been passed through protect_from_queue already. */
-
- void
- emit_library_call (va_alist)
- va_dcl
- {
- register va_list p;
- struct args_size args_size;
- register int argnum;
- enum machine_mode outmode;
- int nargs;
- rtx fun;
- rtx orgfun;
- int inc;
- int count;
- rtx argblock = 0;
- CUMULATIVE_ARGS args_so_far;
- struct arg { rtx value; enum machine_mode mode; rtx reg; int partial;
- struct args_size offset; struct args_size size; };
- struct arg *argvec;
- int old_inhibit_defer_pop = inhibit_defer_pop;
- int no_queue = 0;
- rtx use_insns;
-
- va_start (p);
- orgfun = fun = va_arg (p, rtx);
- no_queue = va_arg (p, int);
- outmode = va_arg (p, enum machine_mode);
- nargs = va_arg (p, int);
-
- /* Copy all the libcall-arguments out of the varargs data
- and into a vector ARGVEC.
-
- Compute how to pass each argument. We only support a very small subset
- of the full argument passing conventions to limit complexity here since
- library functions shouldn't have many args. */
-
- argvec = (struct arg *) alloca (nargs * sizeof (struct arg));
-
- INIT_CUMULATIVE_ARGS (args_so_far, (tree)0, fun);
-
- args_size.constant = 0;
- args_size.var = 0;
-
- for (count = 0; count < nargs; count++)
- {
- rtx val = va_arg (p, rtx);
- enum machine_mode mode = va_arg (p, enum machine_mode);
-
- /* We cannot convert the arg value to the mode the library wants here;
- must do it earlier where we know the signedness of the arg. */
- if (mode == BLKmode
- || (GET_MODE (val) != mode && GET_MODE (val) != VOIDmode))
- abort ();
-
- /* On some machines, there's no way to pass a float to a library fcn.
- Pass it as a double instead. */
- #ifdef GNULIB_NEEDS_DOUBLE
- if (GNULIB_NEEDS_DOUBLE && mode == SFmode)
- val = convert_to_mode (DFmode, val), mode = DFmode;
- #endif
-
- /* Make sure it is a reasonable operand for a move or push insn. */
- if (GET_CODE (val) != REG && GET_CODE (val) != MEM
- && ! (CONSTANT_P (val) && LEGITIMATE_CONSTANT_P (val)))
- val = force_operand (val, 0);
-
- argvec[count].value = val;
- argvec[count].mode = mode;
-
- #ifdef FUNCTION_ARG_PASS_BY_REFERENCE
- if (FUNCTION_ARG_PASS_BY_REFERENCE (args_so_far, mode, (tree)0, 1))
- abort ();
- #endif
-
- argvec[count].reg = FUNCTION_ARG (args_so_far, mode, (tree)0, 1);
- if (argvec[count].reg && GET_CODE (argvec[count].reg) == EXPR_LIST)
- abort ();
- #ifdef FUNCTION_ARG_PARTIAL_NREGS
- argvec[count].partial
- = FUNCTION_ARG_PARTIAL_NREGS (args_so_far, mode, (tree)0, 1);
- #else
- argvec[count].partial = 0;
- #endif
-
- locate_and_pad_parm (mode, 0,
- argvec[count].reg && argvec[count].partial == 0,
- 0, &args_size, &argvec[count].offset,
- &argvec[count].size);
-
- if (argvec[count].size.var)
- abort ();
-
- #ifndef REG_PARM_STACK_SPACE
- if (argvec[count].partial)
- argvec[count].size.constant -= argvec[count].partial * UNITS_PER_WORD;
- #endif
-
- if (argvec[count].reg == 0 || argvec[count].partial != 0
- #ifdef REG_PARM_STACK_SPACE
- || 1
- #endif
- )
- args_size.constant += argvec[count].size.constant;
-
- #ifdef ACCUMULATE_OUTGOING_ARGS
- /* If this arg is actually passed on the stack, it might be
- clobbering something we already put there (this library call might
- be inside the evaluation of an argument to a function whose call
- requires the stack). This will only occur when the library call
- has sufficient args to run out of argument registers. Abort in
- this case; if this ever occurs, code must be added to save and
- restore the arg slot. */
-
- if (argvec[count].reg == 0 || argvec[count].partial != 0)
- abort ();
- #endif
-
- FUNCTION_ARG_ADVANCE (args_so_far, mode, (tree)0, 1);
- }
- va_end (p);
-
- /* If this machine requires an external definition for library
- functions, write one out. */
- assemble_external_libcall (fun);
-
- #ifdef STACK_BOUNDARY
- args_size.constant = (((args_size.constant + (STACK_BYTES - 1))
- / STACK_BYTES) * STACK_BYTES);
- #endif
-
- #ifdef REG_PARM_STACK_SPACE
- args_size.constant = MAX (args_size.constant,
- REG_PARM_STACK_SPACE ((tree) 0));
- #endif
-
- #ifdef ACCUMULATE_OUTGOING_ARGS
- if (args_size.constant > current_function_outgoing_args_size)
- current_function_outgoing_args_size = args_size.constant;
- args_size.constant = 0;
- #endif
-
- #ifndef PUSH_ROUNDING
- argblock = push_block (gen_rtx (CONST_INT, VOIDmode, args_size.constant),
- 0, 0);
- #endif
-
- #ifdef PUSH_ARGS_REVERSED
- inc = -1;
- argnum = nargs - 1;
- #else
- inc = 1;
- argnum = 0;
- #endif
-
- /* Push the args that need to be pushed. */
-
- for (count = 0; count < nargs; count++, argnum += inc)
- {
- register enum machine_mode mode = argvec[argnum].mode;
- register rtx val = argvec[argnum].value;
- rtx reg = argvec[argnum].reg;
- int partial = argvec[argnum].partial;
-
- if (! (reg != 0 && partial == 0))
- emit_push_insn (val, mode, 0, 0, 0, partial, reg, 0, argblock,
- gen_rtx (CONST_INT, VOIDmode,
- argvec[count].offset.constant));
- NO_DEFER_POP;
- }
-
- #ifdef PUSH_ARGS_REVERSED
- argnum = nargs - 1;
- #else
- argnum = 0;
- #endif
-
- /* Now load any reg parms into their regs. */
-
- for (count = 0; count < nargs; count++, argnum += inc)
- {
- register enum machine_mode mode = argvec[argnum].mode;
- register rtx val = argvec[argnum].value;
- rtx reg = argvec[argnum].reg;
- int partial = argvec[argnum].partial;
-
- if (reg != 0 && partial == 0)
- emit_move_insn (reg, val);
- NO_DEFER_POP;
- }
-
- /* For version 1.37, try deleting this entirely. */
- if (! no_queue)
- emit_queue ();
-
- /* Any regs containing parms remain in use through the call. */
- start_sequence ();
- for (count = 0; count < nargs; count++)
- if (argvec[count].reg != 0)
- emit_insn (gen_rtx (USE, VOIDmode, argvec[count].reg));
-
- use_insns = get_insns ();
- end_sequence ();
-
- fun = prepare_call_address (fun, 0, use_insns);
-
- /* Don't allow popping to be deferred, since then
- cse'ing of library calls could delete a call and leave the pop. */
- NO_DEFER_POP;
-
- /* We pass the old value of inhibit_defer_pop + 1 to emit_call_1, which
- will set inhibit_defer_pop to that value. */
-
- emit_call_1 (fun, get_identifier (XSTR (orgfun, 0)), args_size.constant,
- FUNCTION_ARG (args_so_far, VOIDmode, void_type_node, 1),
- outmode != VOIDmode ? hard_libcall_value (outmode) : 0,
- old_inhibit_defer_pop + 1, use_insns);
-
- /* Now restore inhibit_defer_pop to its actual original value. */
- OK_DEFER_POP;
- }
-
- /* Expand an assignment that stores the value of FROM into TO.
- If WANT_VALUE is nonzero, return an rtx for the value of TO.
- (This may contain a QUEUED rtx.)
- Otherwise, the returned value is not meaningful.
-
- SUGGEST_REG is no longer actually used.
- It used to mean, copy the value through a register
- and return that register, if that is possible.
- But now we do this if WANT_VALUE.
-
- If the value stored is a constant, we return the constant. */
-
- rtx
- expand_assignment (to, from, want_value, suggest_reg)
- tree to, from;
- int want_value;
- int suggest_reg;
- {
- register rtx to_rtx = 0;
- rtx result;
-
- /* Don't crash if the lhs of the assignment was erroneous. */
-
- if (TREE_CODE (to) == ERROR_MARK)
- return expand_expr (from, 0, VOIDmode, 0);
-
- /* Assignment of a structure component needs special treatment
- if the structure component's rtx is not simply a MEM.
- Assignment of an array element at a constant index
- has the same problem. */
-
- if (TREE_CODE (to) == COMPONENT_REF
- || TREE_CODE (to) == BIT_FIELD_REF
- || (TREE_CODE (to) == ARRAY_REF
- && TREE_CODE (TREE_OPERAND (to, 1)) == INTEGER_CST
- && TREE_CODE (TYPE_SIZE (TREE_TYPE (to))) == INTEGER_CST))
- {
- enum machine_mode mode1;
- int bitsize;
- int bitpos;
- int unsignedp;
- int volatilep = 0;
- tree tem = get_inner_reference (to, &bitsize, &bitpos,
- &mode1, &unsignedp, &volatilep);
-
- /* If we are going to use store_bit_field and extract_bit_field,
- make sure to_rtx will be safe for multiple use. */
-
- if (mode1 == VOIDmode && want_value)
- tem = stabilize_reference (tem);
-
- to_rtx = expand_expr (tem, 0, VOIDmode, 0);
- if (volatilep)
- {
- if (GET_CODE (to_rtx) == MEM)
- MEM_VOLATILE_P (to_rtx) = 1;
- else
- abort ();
- }
-
- result = store_field (to_rtx, bitsize, bitpos, mode1, from,
- (want_value
- /* Spurious cast makes HPUX compiler happy. */
- ? (enum machine_mode) TYPE_MODE (TREE_TYPE (to))
- : VOIDmode),
- unsignedp,
- /* Required alignment of containing datum. */
- TYPE_ALIGN (TREE_TYPE (tem)) / BITS_PER_UNIT,
- int_size_in_bytes (TREE_TYPE (tem)));
- preserve_temp_slots (result);
- free_temp_slots ();
-
- return result;
- }
-
- /* Ordinary treatment. Expand TO to get a REG or MEM rtx.
- Don't re-expand if it was expanded already (in COMPONENT_REF case). */
-
- if (to_rtx == 0)
- to_rtx = expand_expr (to, 0, VOIDmode, 0);
-
- /* In case we are returning the contents of an object which overlaps
- the place the value is being stored, use a safe function when copying
- a value through a pointer into a structure value return block. */
- if (TREE_CODE (to) == RESULT_DECL && TREE_CODE (from) == INDIRECT_REF
- && current_function_returns_struct
- && !current_function_returns_pcc_struct)
- {
- rtx from_rtx = expand_expr (from, 0, VOIDmode, 0);
- rtx size = expr_size (from);
-
- #ifdef TARGET_MEM_FUNCTIONS
- emit_library_call (memcpy_libfunc, 0,
- VOIDmode, 3, XEXP (to_rtx, 0), Pmode,
- XEXP (from_rtx, 0), Pmode,
- size, Pmode);
- #else
- emit_library_call (bcopy_libfunc, 0,
- VOIDmode, 3, XEXP (from_rtx, 0), Pmode,
- XEXP (to_rtx, 0), Pmode,
- size, Pmode);
- #endif
-
- preserve_temp_slots (to_rtx);
- free_temp_slots ();
- return to_rtx;
- }
-
- /* Compute FROM and store the value in the rtx we got. */
-
- result = store_expr (from, to_rtx, want_value);
- preserve_temp_slots (result);
- free_temp_slots ();
- return result;
- }
-
- /* Generate code for computing expression EXP,
- and storing the value into TARGET.
- Returns TARGET or an equivalent value.
- TARGET may contain a QUEUED rtx.
-
- If SUGGEST_REG is nonzero, copy the value through a register
- and return that register, if that is possible.
-
- If the value stored is a constant, we return the constant. */
-
- rtx
- store_expr (exp, target, suggest_reg)
- register tree exp;
- register rtx target;
- int suggest_reg;
- {
- register rtx temp;
- int dont_return_target = 0;
-
- if (TREE_CODE (exp) == COMPOUND_EXPR)
- {
- /* Perform first part of compound expression, then assign from second
- part. */
- expand_expr (TREE_OPERAND (exp, 0), const0_rtx, VOIDmode, 0);
- emit_queue ();
- return store_expr (TREE_OPERAND (exp, 1), target, suggest_reg);
- }
- else if (TREE_CODE (exp) == COND_EXPR && GET_MODE (target) == BLKmode)
- {
- /* For conditional expression, get safe form of the target. Then
- test the condition, doing the appropriate assignment on either
- side. This avoids the creation of unnecessary temporaries.
- For non-BLKmode, it is more efficient not to do this. */
-
- rtx lab1 = gen_label_rtx (), lab2 = gen_label_rtx ();
-
- emit_queue ();
- target = protect_from_queue (target, 1);
-
- NO_DEFER_POP;
- jumpifnot (TREE_OPERAND (exp, 0), lab1);
- store_expr (TREE_OPERAND (exp, 1), target, suggest_reg);
- emit_queue ();
- emit_jump_insn (gen_jump (lab2));
- emit_barrier ();
- emit_label (lab1);
- store_expr (TREE_OPERAND (exp, 2), target, suggest_reg);
- emit_queue ();
- emit_label (lab2);
- OK_DEFER_POP;
- return target;
- }
- else if (suggest_reg && GET_CODE (target) == MEM
- && GET_MODE (target) != BLKmode)
- /* If target is in memory and caller wants value in a register instead,
- arrange that. Pass TARGET as target for expand_expr so that,
- if EXP is another assignment, SUGGEST_REG will be nonzero for it.
- We know expand_expr will not use the target in that case. */
- {
- temp = expand_expr (exp, cse_not_expected ? 0 : target,
- GET_MODE (target), 0);
- if (GET_MODE (temp) != BLKmode && GET_MODE (temp) != VOIDmode)
- temp = copy_to_reg (temp);
- dont_return_target = 1;
- }
- else if (queued_subexp_p (target))
- /* If target contains a postincrement, it is not safe
- to use as the returned value. It would access the wrong
- place by the time the queued increment gets output.
- So copy the value through a temporary and use that temp
- as the result. */
- {
- if (GET_MODE (target) != BLKmode && GET_MODE (target) != VOIDmode)
- {
- /* Expand EXP into a new pseudo. */
- temp = gen_reg_rtx (GET_MODE (target));
- temp = expand_expr (exp, temp, GET_MODE (target), 0);
- }
- else
- temp = expand_expr (exp, 0, GET_MODE (target), 0);
- dont_return_target = 1;
- }
- else
- {
- temp = expand_expr (exp, target, GET_MODE (target), 0);
- /* DO return TARGET if it's a specified hardware register.
- expand_return relies on this. */
- if (!(target && GET_CODE (target) == REG
- && REGNO (target) < FIRST_PSEUDO_REGISTER)
- && CONSTANT_P (temp))
- dont_return_target = 1;
- }
-
- /* If value was not generated in the target, store it there.
- Convert the value to TARGET's type first if nec. */
-
- if (temp != target && TREE_CODE (exp) != ERROR_MARK)
- {
- target = protect_from_queue (target, 1);
- if (GET_MODE (temp) != GET_MODE (target)
- && GET_MODE (temp) != VOIDmode)
- {
- int unsignedp = TREE_UNSIGNED (TREE_TYPE (exp));
- if (dont_return_target)
- {
- /* In this case, we will return TEMP,
- so make sure it has the proper mode.
- But don't forget to store the value into TARGET. */
- temp = convert_to_mode (GET_MODE (target), temp, unsignedp);
- emit_move_insn (target, temp);
- }
- else
- convert_move (target, temp, unsignedp);
- }
-
- else if (GET_MODE (temp) == BLKmode)
- emit_block_move (target, temp, expr_size (exp),
- TYPE_ALIGN (TREE_TYPE (exp)) / BITS_PER_UNIT);
- else
- emit_move_insn (target, temp);
- }
- if (dont_return_target)
- return temp;
- return target;
- }
-
- /* Store the value of constructor EXP into the rtx TARGET.
- TARGET is either a REG or a MEM. */
-
- static void
- store_constructor (exp, target)
- tree exp;
- rtx target;
- {
- /* We know our target cannot conflict, since safe_from_p has been called. */
- #if 0
- /* Don't try copying piece by piece into a hard register
- since that is vulnerable to being clobbered by EXP.
- Instead, construct in a pseudo register and then copy it all. */
- if (GET_CODE (target) == REG && REGNO (target) < FIRST_PSEUDO_REGISTER)
- {
- rtx temp = gen_reg_rtx (GET_MODE (target));
- store_constructor (exp, temp);
- emit_move_insn (target, temp);
- return;
- }
- #endif
-
- if (TREE_CODE (TREE_TYPE (exp)) == RECORD_TYPE
- || TREE_CODE (TREE_TYPE (exp)) == UNION_TYPE)
- {
- register tree elt;
-
- if (TREE_CODE (TREE_TYPE (exp)) == UNION_TYPE)
- /* Inform later passes that the whole union value is dead. */
- emit_insn (gen_rtx (CLOBBER, VOIDmode, target));
- /* If the constructor has fewer fields than the structure,
- clear the whole structure first. */
- else if (list_length (CONSTRUCTOR_ELTS (exp))
- != list_length (TYPE_FIELDS (TREE_TYPE (exp))))
- clear_storage (target, int_size_in_bytes (TREE_TYPE (exp)));
- else
- /* Inform later passes that the old value is dead. */
- emit_insn (gen_rtx (CLOBBER, VOIDmode, target));
-
- /* Store each element of the constructor into
- the corresponding field of TARGET. */
-
- for (elt = CONSTRUCTOR_ELTS (exp); elt; elt = TREE_CHAIN (elt))
- {
- register tree field = TREE_PURPOSE (elt);
- register enum machine_mode mode;
- int bitsize;
- int bitpos;
- int unsignedp;
-
- bitsize = TREE_INT_CST_LOW (DECL_SIZE (field));
- unsignedp = TREE_UNSIGNED (field);
- mode = DECL_MODE (field);
- if (DECL_BIT_FIELD (field))
- mode = VOIDmode;
-
- if (TREE_CODE (DECL_FIELD_BITPOS (field)) != INTEGER_CST)
- /* ??? This case remains to be written. */
- abort ();
-
- bitpos = TREE_INT_CST_LOW (DECL_FIELD_BITPOS (field));
-
- store_field (target, bitsize, bitpos, mode, TREE_VALUE (elt),
- /* The alignment of TARGET is
- at least what its type requires. */
- VOIDmode, 0,
- TYPE_ALIGN (TREE_TYPE (exp)) / BITS_PER_UNIT,
- int_size_in_bytes (TREE_TYPE (exp)));
- }
- }
- else if (TREE_CODE (TREE_TYPE (exp)) == ARRAY_TYPE)
- {
- register tree elt;
- register int i;
- tree domain = TYPE_DOMAIN (TREE_TYPE (exp));
- int minelt = TREE_INT_CST_LOW (TYPE_MIN_VALUE (domain));
- int maxelt = TREE_INT_CST_LOW (TYPE_MAX_VALUE (domain));
- tree elttype = TREE_TYPE (TREE_TYPE (exp));
-
- /* If the constructor has fewer fields than the structure,
- clear the whole structure first. */
-
- if (list_length (CONSTRUCTOR_ELTS (exp)) < maxelt - minelt + 1)
- clear_storage (target, maxelt - minelt + 1);
- else
- /* Inform later passes that the old value is dead. */
- emit_insn (gen_rtx (CLOBBER, VOIDmode, target));
-
- /* Store each element of the constructor into
- the corresponding element of TARGET, determined
- by counting the elements. */
- for (elt = CONSTRUCTOR_ELTS (exp), i = 0;
- elt;
- elt = TREE_CHAIN (elt), i++)
- {
- register enum machine_mode mode;
- int bitsize;
- int bitpos;
- int unsignedp;
-
- mode = TYPE_MODE (elttype);
- bitsize = GET_MODE_BITSIZE (mode);
- unsignedp = TREE_UNSIGNED (elttype);
-
- bitpos = (i * TREE_INT_CST_LOW (TYPE_SIZE (elttype)));
-
- store_field (target, bitsize, bitpos, mode, TREE_VALUE (elt),
- /* The alignment of TARGET is
- at least what its type requires. */
- VOIDmode, 0,
- TYPE_ALIGN (TREE_TYPE (exp)) / BITS_PER_UNIT,
- int_size_in_bytes (TREE_TYPE (exp)));
- }
- }
-
- else
- abort ();
- }
-
- /* Store the value of EXP (an expression tree)
- into a subfield of TARGET which has mode MODE and occupies
- BITSIZE bits, starting BITPOS bits from the start of TARGET.
- If MODE is VOIDmode, it means that we are storing into a bit-field.
-
- If VALUE_MODE is VOIDmode, return nothing in particular.
- UNSIGNEDP is not used in this case.
-
- Otherwise, return an rtx for the value stored. This rtx
- has mode VALUE_MODE if that is convenient to do.
- In this case, UNSIGNEDP must be nonzero if the value is an unsigned type.
-
- ALIGN is the alignment that TARGET is known to have, measured in bytes.
- TOTAL_SIZE is the size in bytes of the structure, or -1 if varying. */
-
- static rtx
- store_field (target, bitsize, bitpos, mode, exp, value_mode,
- unsignedp, align, total_size)
- rtx target;
- int bitsize, bitpos;
- enum machine_mode mode;
- tree exp;
- enum machine_mode value_mode;
- int unsignedp;
- int align;
- int total_size;
- {
- int width_mask = 0;
-
- if (bitsize < HOST_BITS_PER_INT)
- width_mask = (1 << bitsize) - 1;
-
- /* If the structure is in a register or if the component
- is a bit field, we cannot use addressing to access it.
- Use bit-field techniques or SUBREG to store in it. */
-
- if (mode == VOIDmode || GET_CODE (target) == REG
- || GET_CODE (target) == SUBREG)
- {
- rtx temp = expand_expr (exp, 0, VOIDmode, 0);
- /* Store the value in the bitfield. */
- store_bit_field (target, bitsize, bitpos, mode, temp, align, total_size);
- if (value_mode != VOIDmode)
- {
- /* The caller wants an rtx for the value. */
- /* If possible, avoid refetching from the bitfield itself. */
- if (width_mask != 0
- && ! (GET_CODE (target) == MEM && MEM_VOLATILE_P (target)))
- return expand_and (temp,
- gen_rtx (CONST_INT, VOIDmode, width_mask), 0);
- return extract_bit_field (target, bitsize, bitpos, unsignedp,
- 0, value_mode, 0, align, total_size);
- }
- return const0_rtx;
- }
- else
- {
- rtx addr = XEXP (target, 0);
- rtx to_rtx;
-
- /* If a value is wanted, it must be the lhs;
- so make the address stable for multiple use. */
-
- if (value_mode != VOIDmode && GET_CODE (addr) != REG
- && ! CONSTANT_ADDRESS_P (addr)
- /* A frame-pointer reference is already stable. */
- && ! (GET_CODE (addr) == PLUS
- && GET_CODE (XEXP (addr, 1)) == CONST_INT
- && (XEXP (addr, 0) == virtual_incoming_args_rtx
- || XEXP (addr, 0) == virtual_stack_vars_rtx)))
- addr = copy_to_reg (addr);
-
- /* Now build a reference to just the desired component. */
-
- to_rtx = change_address (target, mode,
- plus_constant (addr, (bitpos / BITS_PER_UNIT)));
- MEM_IN_STRUCT_P (to_rtx) = 1;
-
- return store_expr (exp, to_rtx, value_mode != VOIDmode);
- }
- }
-
- /* Given an expression EXP that may be a COMPONENT_REF, a BIT_FIELD_REF,
- or an ARRAY_REF, look for nested COMPONENT_REFs, BIT_FIELD_REFs, or
- ARRAY_REFs at constant positions and find the ultimate containing object,
- which we return.
-
- We set *PBITSIZE to the size in bits that we want, *PBITPOS to the
- bit position, and *PUNSIGNEDP to the signedness of the field.
-
- If any of the extraction expressions is volatile,
- we store 1 in *PVOLATILEP. Otherwise we don't change that.
-
- If the field is a bit-field, *PMODE is set to VOIDmode. Otherwise, it
- is a mode that can be used to access the field. In that case, *PBITSIZE
- is redundant. */
-
- tree
- get_inner_reference (exp, pbitsize, pbitpos, pmode, punsignedp, pvolatilep)
- tree exp;
- int *pbitsize;
- int *pbitpos;
- enum machine_mode *pmode;
- int *punsignedp;
- int *pvolatilep;
- {
- tree size_tree = 0;
- enum machine_mode mode = VOIDmode;
-
- if (TREE_CODE (exp) == COMPONENT_REF)
- {
- size_tree = DECL_SIZE (TREE_OPERAND (exp, 1));
- if (! DECL_BIT_FIELD (TREE_OPERAND (exp, 1)))
- mode = DECL_MODE (TREE_OPERAND (exp, 1));
- *punsignedp = TREE_UNSIGNED (TREE_OPERAND (exp, 1));
- }
- else if (TREE_CODE (exp) == BIT_FIELD_REF)
- {
- size_tree = TREE_OPERAND (exp, 1);
- *punsignedp = TREE_UNSIGNED (exp);
- }
- else
- {
- mode = TYPE_MODE (TREE_TYPE (exp));
- *pbitsize = GET_MODE_BITSIZE (mode);
- *punsignedp = TREE_UNSIGNED (TREE_TYPE (exp));
- }
-
- if (size_tree)
- {
- if (TREE_CODE (size_tree) != INTEGER_CST)
- abort ();
-
- *pbitsize = TREE_INT_CST_LOW (size_tree);
- }
-
- /* Compute cumulative bit-offset for nested component-refs and array-refs,
- and find the ultimate containing object. */
-
- *pbitpos = 0;
-
- while (1)
- {
- if (TREE_CODE (exp) == COMPONENT_REF)
- {
- tree field = TREE_OPERAND (exp, 1);
-
- if (TREE_CODE (DECL_FIELD_BITPOS (field)) != INTEGER_CST)
- /* ??? This case remains to be written. */
- abort ();
-
- *pbitpos += TREE_INT_CST_LOW (DECL_FIELD_BITPOS (field));
- if (TREE_THIS_VOLATILE (exp))
- *pvolatilep = 1;
- }
- else if (TREE_CODE (exp) == BIT_FIELD_REF)
- {
- if (TREE_CODE (TREE_OPERAND (exp, 2)) != INTEGER_CST)
- /* ??? This case remains to be written. */
- abort ();
-
- *pbitpos += TREE_INT_CST_LOW (TREE_OPERAND (exp, 2));
- if (TREE_THIS_VOLATILE (exp))
- *pvolatilep = 1;
- }
- else if (TREE_CODE (exp) == ARRAY_REF
- && TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST
- && TREE_CODE (TYPE_SIZE (TREE_TYPE (exp))) == INTEGER_CST)
- {
- *pbitpos += (TREE_INT_CST_LOW (TREE_OPERAND (exp, 1))
- * TREE_INT_CST_LOW (TYPE_SIZE (TREE_TYPE (exp))));
- if (TREE_THIS_VOLATILE (exp))
- *pvolatilep = 1;
- }
- else if (TREE_CODE (exp) != NON_LVALUE_EXPR
- && ! ((TREE_CODE (exp) == NOP_EXPR
- || TREE_CODE (exp) == CONVERT_EXPR)
- && (TYPE_MODE (TREE_TYPE (exp))
- == TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))))
- break;
- exp = TREE_OPERAND (exp, 0);
- }
-
- /* If this was a bit-field, see if there is a mode that allows direct
- access in case EXP is in memory. */
- if (mode == VOIDmode && *pbitpos % *pbitsize == 0)
- for (mode = QImode; mode != VOIDmode; mode = GET_MODE_WIDER_MODE (mode))
- if (GET_MODE_BITSIZE (mode) == *pbitsize)
- break;
-
- *pmode = mode;
-
- return exp;
- }
-
- /* Given an rtx VALUE that may contain additions and multiplications,
- return an equivalent value that just refers to a register or memory.
- This is done by generating instructions to perform the arithmetic
- and returning a pseudo-register containing the value. */
-
- rtx
- force_operand (value, target)
- rtx value, target;
- {
- register optab binoptab = 0;
- /* Use a temporary to force order of execution of calls to
- `force_operand'. */
- rtx tmp;
- register rtx op2;
- /* Use subtarget as the target for operand 0 of a binary operation. */
- register rtx subtarget = (target != 0 && GET_CODE (target) == REG ? target : 0);
-
- if (GET_CODE (value) == PLUS)
- binoptab = add_optab;
- else if (GET_CODE (value) == MINUS)
- binoptab = sub_optab;
- else if (GET_CODE (value) == MULT)
- {
- op2 = XEXP (value, 1);
- if (!CONSTANT_P (op2)
- && !(GET_CODE (op2) == REG && op2 != subtarget))
- subtarget = 0;
- tmp = force_operand (XEXP (value, 0), subtarget);
- return expand_mult (GET_MODE (value), tmp,
- force_operand (op2, 0),
- target, 0);
- }
-
- if (binoptab)
- {
- op2 = XEXP (value, 1);
- if (!CONSTANT_P (op2)
- && !(GET_CODE (op2) == REG && op2 != subtarget))
- subtarget = 0;
- if (binoptab == sub_optab && GET_CODE (op2) == CONST_INT)
- {
- binoptab = add_optab;
- op2 = negate_rtx (GET_MODE (value), op2);
- }
-
- /* Check for an addition with OP2 a constant integer and our first
- operand a PLUS of a virtual register and something else. In that
- case, we want to emit the sum of the virtual register and the
- constant first and then add the other value. This allows virtual
- register instantiation to simply modify the constant rather than
- creating another one around this addition. */
- if (binoptab == add_optab && GET_CODE (op2) == CONST_INT
- && GET_CODE (XEXP (value, 0)) == PLUS
- && GET_CODE (XEXP (XEXP (value, 0), 0)) == REG
- && REGNO (XEXP (XEXP (value, 0), 0)) >= FIRST_VIRTUAL_REGISTER
- && REGNO (XEXP (XEXP (value, 0), 0)) <= LAST_VIRTUAL_REGISTER)
- {
- rtx temp = expand_binop (GET_MODE (value), binoptab,
- XEXP (XEXP (value, 0), 0), op2,
- subtarget, 0, OPTAB_LIB_WIDEN);
- return expand_binop (GET_MODE (value), binoptab, temp,
- force_operand (XEXP (XEXP (value, 0), 1), 0),
- target, 0, OPTAB_LIB_WIDEN);
- }
-
- tmp = force_operand (XEXP (value, 0), subtarget);
- return expand_binop (GET_MODE (value), binoptab, tmp,
- force_operand (op2, 0),
- target, 0, OPTAB_LIB_WIDEN);
- /* We give UNSIGNEP = 0 to expand_binop
- because the only operations we are expanding here are signed ones. */
- }
- return value;
- }
-
- /* Subroutine of expand_expr:
- save the non-copied parts (LIST) of an expr (LHS), and return a list
- which can restore these values to their previous values,
- should something modify their storage. */
-
- static tree
- save_noncopied_parts (lhs, list)
- tree lhs;
- tree list;
- {
- tree tail;
- tree parts = 0;
-
- for (tail = list; tail; tail = TREE_CHAIN (tail))
- if (TREE_CODE (TREE_VALUE (tail)) == TREE_LIST)
- parts = chainon (parts, save_noncopied_parts (lhs, TREE_VALUE (tail)));
- else
- {
- tree part = TREE_VALUE (tail);
- tree part_type = TREE_TYPE (part);
- tree to_be_saved = build (COMPONENT_REF, part_type, lhs, part, 0);
- rtx target = assign_stack_temp (TYPE_MODE (part_type),
- int_size_in_bytes (part_type), 0);
- if (! memory_address_p (TYPE_MODE (part_type), XEXP (target, 0)))
- target = change_address (target, TYPE_MODE (part_type), 0);
- parts = tree_cons (to_be_saved,
- build (RTL_EXPR, part_type, 0, (tree) target),
- parts);
- store_expr (TREE_PURPOSE (parts), RTL_EXPR_RTL (TREE_VALUE (parts)), 0);
- }
- return parts;
- }
-
- /* Subroutine of expand_expr:
- record the non-copied parts (LIST) of an expr (LHS), and return a list
- which specifies the initial values of these parts. */
-
- static tree
- init_noncopied_parts (lhs, list)
- tree lhs;
- tree list;
- {
- tree tail;
- tree parts = 0;
-
- for (tail = list; tail; tail = TREE_CHAIN (tail))
- if (TREE_CODE (TREE_VALUE (tail)) == TREE_LIST)
- parts = chainon (parts, init_noncopied_parts (lhs, TREE_VALUE (tail)));
- else
- {
- tree part = TREE_VALUE (tail);
- tree part_type = TREE_TYPE (part);
- tree to_be_initialized = build (COMPONENT_REF, part_type, lhs, part, 0);
- parts = tree_cons (TREE_PURPOSE (tail), to_be_initialized, parts);
- }
- return parts;
- }
-
- /* Subroutine of expand_expr: return nonzero iff there is no way that
- EXP can reference X, which is being modified. */
-
- static int
- safe_from_p (x, exp)
- rtx x;
- tree exp;
- {
- rtx exp_rtl = 0;
- int i, nops;
-
- if (x == 0)
- return 1;
-
- /* If this is a subreg of a hard register, declare it unsafe, otherwise,
- find the underlying pseudo. */
- if (GET_CODE (x) == SUBREG)
- {
- x = SUBREG_REG (x);
- if (GET_CODE (x) == REG && REGNO (x) < FIRST_PSEUDO_REGISTER)
- return 0;
- }
-
- /* If X is a location in the outgoing argument area, it is always safe. */
- if (GET_CODE (x) == MEM
- && (XEXP (x, 0) == virtual_outgoing_args_rtx
- || (GET_CODE (XEXP (x, 0)) == PLUS
- && XEXP (XEXP (x, 0), 0) == virtual_outgoing_args_rtx)))
- return 1;
-
- switch (TREE_CODE_CLASS (TREE_CODE (exp)))
- {
- case 'd':
- exp_rtl = DECL_RTL (exp);
- break;
-
- case 'c':
- return 1;
-
- case 'x':
- if (TREE_CODE (exp) == TREE_LIST)
- return (safe_from_p (x, TREE_VALUE (exp))
- && (TREE_CHAIN (exp) == 0
- || safe_from_p (x, TREE_CHAIN (exp))));
- else
- return 0;
-
- case '1':
- return safe_from_p (x, TREE_OPERAND (exp, 0));
-
- case '2':
- case '<':
- return (safe_from_p (x, TREE_OPERAND (exp, 0))
- && safe_from_p (x, TREE_OPERAND (exp, 1)));
-
- case 'e':
- case 'r':
- /* Now do code-specific tests. EXP_RTL is set to any rtx we find in
- the expression. If it is set, we conflict iff we are that rtx or
- both are in memory. Otherwise, we check all operands of the
- expression recursively. */
-
- switch (TREE_CODE (exp))
- {
- case ADDR_EXPR:
- return staticp (TREE_OPERAND (exp, 0));
-
- case INDIRECT_REF:
- if (GET_CODE (x) == MEM)
- return 0;
- break;
-
- case CALL_EXPR:
- exp_rtl = CALL_EXPR_RTL (exp);
- if (exp_rtl == 0)
- {
- /* Assume that the call will clobber all hard registers and
- all of memory. */
- if ((GET_CODE (x) == REG && REGNO (x) < FIRST_PSEUDO_REGISTER)
- || GET_CODE (x) == MEM)
- return 0;
- }
-
- break;
-
- case RTL_EXPR:
- exp_rtl = RTL_EXPR_RTL (exp);
- if (exp_rtl == 0)
- /* We don't know what this can modify. */
- return 0;
-
- break;
-
- case WITH_CLEANUP_EXPR:
- exp_rtl = RTL_EXPR_RTL (exp);
- break;
-
- case SAVE_EXPR:
- exp_rtl = SAVE_EXPR_RTL (exp);
- break;
-
- case METHOD_CALL_EXPR:
- /* This takes a rtx argument, but shouldn't appear here. */
- abort ();
- }
-
- /* If we have an rtx, we do not need to scan our operands. */
- if (exp_rtl)
- break;
-
- nops = tree_code_length[(int) TREE_CODE (exp)];
- for (i = 0; i < nops; i++)
- if (TREE_OPERAND (exp, i) != 0
- && ! safe_from_p (x, TREE_OPERAND (exp, i)))
- return 0;
- }
-
- /* If we have an rtl, find any enclosed object. Then see if we conflict
- with it. */
- if (exp_rtl)
- {
- if (GET_CODE (exp_rtl) == SUBREG)
- {
- exp_rtl = SUBREG_REG (exp_rtl);
- if (GET_CODE (exp_rtl) == REG
- && REGNO (exp_rtl) < FIRST_PSEUDO_REGISTER)
- return 0;
- }
-
- /* If the rtl is X, then it is not safe. Otherwise, it is unless both
- are memory and EXP is not readonly. */
- return ! (rtx_equal_p (x, exp_rtl)
- || (GET_CODE (x) == MEM && GET_CODE (exp_rtl) == MEM
- && ! TREE_READONLY (exp)));
- }
-
- /* If we reach here, it is safe. */
- return 1;
- }
-
- /* Subroutine of expand_expr: return nonzero iff EXP is an
- expression whose type is statically determinable. */
-
- static int
- fixed_type_p (exp)
- tree exp;
- {
- if (TREE_CODE (exp) == PARM_DECL
- || TREE_CODE (exp) == VAR_DECL
- || TREE_CODE (exp) == CALL_EXPR || TREE_CODE (exp) == TARGET_EXPR
- || TREE_CODE (exp) == COMPONENT_REF
- || TREE_CODE (exp) == ARRAY_REF)
- return 1;
- return 0;
- }
-
- /* expand_expr: generate code for computing expression EXP.
- An rtx for the computed value is returned. The value is never null.
- In the case of a void EXP, const0_rtx is returned.
-
- The value may be stored in TARGET if TARGET is nonzero.
- TARGET is just a suggestion; callers must assume that
- the rtx returned may not be the same as TARGET.
-
- If TARGET is CONST0_RTX, it means that the value will be ignored.
-
- If TMODE is not VOIDmode, it suggests generating the
- result in mode TMODE. But this is done only when convenient.
- Otherwise, TMODE is ignored and the value generated in its natural mode.
- TMODE is just a suggestion; callers must assume that
- the rtx returned may not have mode TMODE.
-
- If MODIFIER is EXPAND_SUM then when EXP is an addition
- we can return an rtx of the form (MULT (REG ...) (CONST_INT ...))
- or a nest of (PLUS ...) and (MINUS ...) where the terms are
- products as above, or REG or MEM, or constant.
- Ordinarily in such cases we would output mul or add instructions
- and then return a pseudo reg containing the sum.
-
- If MODIFIER is EXPAND_CONST_ADDRESS, EXPAND_SUM, or EXPAND_INTO_STACK
- then it is ok to return a MEM rtx whose address is a constant that isn't
- a legitimate address. It is the caller's responsibility to call
- memory_address or equivalent before using the returned address in an
- insn. */
-
- rtx
- expand_expr (exp, target, tmode, modifier)
- register tree exp;
- rtx target;
- enum machine_mode tmode;
- enum expand_modifier modifier;
- {
- register rtx op0, op1, temp;
- tree type = TREE_TYPE (exp);
- int unsignedp = TREE_UNSIGNED (type);
- register enum machine_mode mode = TYPE_MODE (type);
- register enum tree_code code = TREE_CODE (exp);
- optab this_optab;
- /* Use subtarget as the target for operand 0 of a binary operation. */
- rtx subtarget = (target != 0 && GET_CODE (target) == REG ? target : 0);
- rtx original_target = target;
- int ignore = target == const0_rtx;
- tree context;
-
- extern rtx (*lang_expand_expr)();
-
- /* Don't use hard regs as subtargets, because the combiner
- can only handle pseudo regs. */
- if (subtarget && REGNO (subtarget) < FIRST_PSEUDO_REGISTER)
- subtarget = 0;
- /* Avoid subtargets inside loops,
- since they hide some invariant expressions. */
- if (preserve_subexpressions_p ())
- subtarget = 0;
-
- if (ignore) target = 0, original_target = 0;
-
- /* If will do cse, generate all results into pseudo registers
- since 1) that allows cse to find more things
- and 2) otherwise cse could produce an insn the machine
- cannot support. */
-
- if (! cse_not_expected && mode != BLKmode && target
- && (GET_CODE (target) != REG || REGNO (target) < FIRST_PSEUDO_REGISTER))
- target = subtarget;
-
- /* Ensure we reference a volatile object even if value is ignored. */
- if (ignore && TREE_THIS_VOLATILE (exp)
- && mode != VOIDmode && mode != BLKmode)
- {
- target = gen_reg_rtx (mode);
- temp = expand_expr (exp, target, VOIDmode, modifier);
- if (temp != target)
- emit_move_insn (target, temp);
- return target;
- }
-
- switch (code)
- {
- case LABEL_DECL:
- return gen_rtx (MEM, Pmode, gen_rtx (LABEL_REF, VOIDmode,
- label_rtx (exp)));
-
- case PARM_DECL:
- if (DECL_RTL (exp) == 0)
- {
- error_with_decl (exp, "prior parameter's size depends on `%s'");
- return const0_rtx;
- }
-
- case FUNCTION_DECL:
- case VAR_DECL:
- case RESULT_DECL:
- if (DECL_RTL (exp) == 0)
- abort ();
- /* Ensure variable marked as used
- even if it doesn't go through a parser. */
- TREE_USED (exp) = 1;
- /* Handle variables inherited from containing functions. */
- context = decl_function_context (exp);
- if (context != 0 && context != current_function_decl
- /* If var is static, we don't need a static chain to access it. */
- && ! (GET_CODE (DECL_RTL (exp)) == MEM
- && CONSTANT_P (XEXP (DECL_RTL (exp), 0))))
- {
- rtx addr;
-
- /* Mark as non-local and addressable. */
- TREE_NONLOCAL (exp) = 1;
- mark_addressable (exp);
- if (GET_CODE (DECL_RTL (exp)) != MEM)
- abort ();
- addr = XEXP (DECL_RTL (exp), 0);
- if (GET_CODE (addr) == MEM)
- addr = gen_rtx (MEM, Pmode, fix_lexical_addr (XEXP (addr, 0), exp));
- else
- addr = fix_lexical_addr (addr, exp);
- return change_address (DECL_RTL (exp), mode, addr);
- }
- /* This is the case of an array whose size is to be determined
- from its initializer, while the initializer is still being parsed.
- See expand_decl. */
- if (GET_CODE (DECL_RTL (exp)) == MEM
- && GET_CODE (XEXP (DECL_RTL (exp), 0)) == REG)
- return change_address (DECL_RTL (exp), GET_MODE (DECL_RTL (exp)),
- XEXP (DECL_RTL (exp), 0));
- if (GET_CODE (DECL_RTL (exp)) == MEM
- && modifier != EXPAND_CONST_ADDRESS
- && modifier != EXPAND_SUM
- && modifier != EXPAND_INTO_STACK)
- {
- /* DECL_RTL probably contains a constant address.
- On RISC machines where a constant address isn't valid,
- make some insns to get that address into a register. */
- if (!memory_address_p (DECL_MODE (exp), XEXP (DECL_RTL (exp), 0))
- || (flag_force_addr
- && CONSTANT_ADDRESS_P (XEXP (DECL_RTL (exp), 0))))
- return change_address (DECL_RTL (exp), VOIDmode,
- copy_rtx (XEXP (DECL_RTL (exp), 0)));
- }
- return DECL_RTL (exp);
-
- case INTEGER_CST:
- return immed_double_const (TREE_INT_CST_LOW (exp),
- TREE_INT_CST_HIGH (exp),
- mode);
-
- case CONST_DECL:
- return expand_expr (DECL_INITIAL (exp), target, VOIDmode, 0);
-
- case REAL_CST:
- /* If optimized, generate immediate CONST_DOUBLE
- which will be turned into memory by reload if necessary.
-
- We used to force a register so that loop.c could see it. But
- this does not allow gen_* patterns to perform optimizations with
- the constants. It also produces two insns in cases like "x = 1.0;".
- On most machines, floating-point constants are not permitted in
- many insns, so we'd end up copying it to a register in any case.
-
- Now, we do the copying in expand_binop, if appropriate. */
- return immed_real_const (exp);
-
- case COMPLEX_CST:
- case STRING_CST:
- if (! TREE_CST_RTL (exp))
- output_constant_def (exp);
-
- /* TREE_CST_RTL probably contains a constant address.
- On RISC machines where a constant address isn't valid,
- make some insns to get that address into a register. */
- if (GET_CODE (TREE_CST_RTL (exp)) == MEM
- && modifier != EXPAND_CONST_ADDRESS
- && !memory_address_p (mode, XEXP (TREE_CST_RTL (exp), 0)))
- return change_address (TREE_CST_RTL (exp), VOIDmode,
- copy_rtx (XEXP (TREE_CST_RTL (exp), 0)));
- return TREE_CST_RTL (exp);
-
- case SAVE_EXPR:
- context = decl_function_context (exp);
- if (context == current_function_decl)
- context = 0;
-
- /* If this is non-local, handle it. */
- if (context)
- {
- temp = SAVE_EXPR_RTL (exp);
- if (temp && GET_CODE (temp) == REG)
- {
- put_var_into_stack (exp);
- temp = SAVE_EXPR_RTL (exp);
- }
- if (temp == 0 || GET_CODE (temp) != MEM)
- abort ();
- return change_address (temp, mode,
- fix_lexical_addr (XEXP (temp, 0), exp));
- }
- if (SAVE_EXPR_RTL (exp) == 0)
- {
- if (mode == BLKmode)
- temp
- = assign_stack_temp (mode,
- int_size_in_bytes (TREE_TYPE (exp)), 0);
- else
- temp = gen_reg_rtx (mode);
- SAVE_EXPR_RTL (exp) = temp;
- store_expr (TREE_OPERAND (exp, 0), temp, 0);
- if (!optimize && GET_CODE (temp) == REG)
- save_expr_regs = gen_rtx (EXPR_LIST, VOIDmode, temp,
- save_expr_regs);
- }
- return SAVE_EXPR_RTL (exp);
-
- case EXIT_EXPR:
- /* Exit the current loop if the body-expression is true. */
- {
- rtx label = gen_label_rtx ();
- do_jump (TREE_OPERAND (exp, 0), label, 0);
- expand_exit_loop (0);
- emit_label (label);
- }
- return const0_rtx;
-
- case LOOP_EXPR:
- expand_start_loop (1);
- expand_expr_stmt (TREE_OPERAND (exp, 0));
- expand_end_loop ();
-
- return const0_rtx;
-
- case BIND_EXPR:
- {
- tree vars = TREE_OPERAND (exp, 0);
- int vars_need_expansion = 0;
-
- /* Need to open a binding contour here because
- if there are any cleanups they most be contained here. */
- expand_start_bindings (0);
-
- /* Mark the corresponding BLOCK for output. */
- if (TREE_OPERAND (exp, 2) != 0)
- TREE_USED (TREE_OPERAND (exp, 2)) = 1;
-
- /* If VARS have not yet been expanded, expand them now. */
- while (vars)
- {
- if (DECL_RTL (vars) == 0)
- {
- vars_need_expansion = 1;
- expand_decl (vars);
- }
- expand_decl_init (vars);
- vars = TREE_CHAIN (vars);
- }
-
- temp = expand_expr (TREE_OPERAND (exp, 1), target, tmode, modifier);
-
- expand_end_bindings (vars, 0, 0);
-
- return temp;
- }
-
- case RTL_EXPR:
- if (RTL_EXPR_SEQUENCE (exp) == const0_rtx)
- abort ();
- emit_insns (RTL_EXPR_SEQUENCE (exp));
- RTL_EXPR_SEQUENCE (exp) = const0_rtx;
- return RTL_EXPR_RTL (exp);
-
- case CONSTRUCTOR:
- /* All elts simple constants => refer to a constant in memory. */
- if (TREE_STATIC (exp))
- /* For aggregate types with non-BLKmode modes,
- this should ideally construct a CONST_INT. */
- {
- rtx constructor = output_constant_def (exp);
- if (! memory_address_p (GET_MODE (constructor),
- XEXP (constructor, 0)))
- constructor = change_address (constructor, VOIDmode,
- XEXP (constructor, 0));
- return constructor;
- }
-
- if (ignore)
- {
- tree elt;
- for (elt = CONSTRUCTOR_ELTS (exp); elt; elt = TREE_CHAIN (elt))
- expand_expr (TREE_VALUE (elt), const0_rtx, VOIDmode, 0);
- return const0_rtx;
- }
- else
- {
- if (target == 0 || ! safe_from_p (target, exp))
- {
- if (mode != BLKmode)
- target = gen_reg_rtx (mode);
- else
- target = assign_stack_temp (mode, int_size_in_bytes (type), 0);
- }
- store_constructor (exp, target);
- return target;
- }
-
- case INDIRECT_REF:
- {
- tree exp1 = TREE_OPERAND (exp, 0);
- tree exp2;
-
- /* A SAVE_EXPR as the address in an INDIRECT_EXPR is generated
- for *PTR += ANYTHING where PTR is put inside the SAVE_EXPR.
- This code has the same general effect as simply doing
- expand_expr on the save expr, except that the expression PTR
- is computed for use as a memory address. This means different
- code, suitable for indexing, may be generated. */
- if (TREE_CODE (exp1) == SAVE_EXPR
- && SAVE_EXPR_RTL (exp1) == 0
- && TREE_CODE (exp2 = TREE_OPERAND (exp1, 0)) != ERROR_MARK
- && TYPE_MODE (TREE_TYPE (exp1)) == Pmode
- && TYPE_MODE (TREE_TYPE (exp2)) == Pmode)
- {
- temp = expand_expr (TREE_OPERAND (exp1, 0), 0, VOIDmode, EXPAND_SUM);
- op0 = memory_address (mode, temp);
- op0 = copy_all_regs (op0);
- SAVE_EXPR_RTL (exp1) = op0;
- }
- else
- {
- /* ??? Tiemann: please explain why these precise
- conditions are desirable. */
- if (modifier == EXPAND_INTO_STACK
- && original_target
- && GET_CODE (original_target) == MEM)
- op0 = expand_expr (exp1, XEXP (original_target, 0),
- VOIDmode, EXPAND_INTO_STACK);
- else
- op0 = expand_expr (exp1, 0, VOIDmode, EXPAND_SUM);
- op0 = memory_address (mode, op0);
- }
- }
- temp = gen_rtx (MEM, mode, op0);
- /* If address was computed by addition,
- mark this as an element of an aggregate. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == PLUS_EXPR
- || (TREE_CODE (TREE_OPERAND (exp, 0)) == SAVE_EXPR
- && TREE_CODE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)) == PLUS_EXPR))
- MEM_IN_STRUCT_P (temp) = 1;
- MEM_VOLATILE_P (temp) = TREE_THIS_VOLATILE (exp) || flag_volatile;
- #if 0 /* It is incorrect to set RTX_UNCHANGING_P here, because the fact that
- a location is accessed through a pointer to const does not mean
- that the value there can never change. */
- RTX_UNCHANGING_P (temp) = TREE_READONLY (exp);
- #endif
- return temp;
-
- case ARRAY_REF:
- if (TREE_CODE (TREE_OPERAND (exp, 1)) != INTEGER_CST
- || TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
- {
- /* Nonconstant array index or nonconstant element size.
- Generate the tree for *(&array+index) and expand that,
- except do it in a language-independent way
- and don't complain about non-lvalue arrays.
- `mark_addressable' should already have been called
- for any array for which this case will be reached. */
-
- /* Don't forget the const or volatile flag from the array element. */
- tree variant_type = build_type_variant (type,
- TREE_READONLY (exp),
- TREE_THIS_VOLATILE (exp));
- tree array_adr = build1 (ADDR_EXPR, build_pointer_type (variant_type),
- TREE_OPERAND (exp, 0));
- tree index = TREE_OPERAND (exp, 1);
- tree elt;
-
- /* Convert the integer argument to a type the same size as a pointer
- so the multiply won't overflow spuriously. */
- if (TYPE_PRECISION (TREE_TYPE (index)) != POINTER_SIZE)
- index = convert (type_for_size (POINTER_SIZE, 0), index);
-
- /* Don't think the address has side effects
- just because the array does.
- (In some cases the address might have side effects,
- and we fail to record that fact here. However, it should not
- matter, since expand_expr should not care.) */
- TREE_SIDE_EFFECTS (array_adr) = 0;
-
- elt = build1 (INDIRECT_REF, type,
- fold (build (PLUS_EXPR, TYPE_POINTER_TO (variant_type),
- array_adr,
- fold (build (MULT_EXPR,
- TYPE_POINTER_TO (variant_type),
- index, size_in_bytes (type))))));
-
- /* Volatility, etc., of new expression is same as old expression. */
- TREE_SIDE_EFFECTS (elt) = TREE_SIDE_EFFECTS (exp);
- TREE_THIS_VOLATILE (elt) = TREE_THIS_VOLATILE (exp);
- TREE_READONLY (elt) = TREE_READONLY (exp);
-
- return expand_expr (elt, target, tmode, modifier);
- }
-
- /* Fold an expression like: "foo"[2].
- This is not done in fold so it won't happen inside &. */
- {
- int i;
- tree arg0 = TREE_OPERAND (exp, 0);
- tree arg1 = TREE_OPERAND (exp, 1);
-
- if (TREE_CODE (arg0) == STRING_CST
- && TREE_CODE (arg1) == INTEGER_CST
- && !TREE_INT_CST_HIGH (arg1)
- && (i = TREE_INT_CST_LOW (arg1)) < TREE_STRING_LENGTH (arg0))
- {
- if (TREE_TYPE (TREE_TYPE (arg0)) == integer_type_node)
- {
- exp = build_int_2 (((int *)TREE_STRING_POINTER (arg0))[i], 0);
- TREE_TYPE (exp) = integer_type_node;
- return expand_expr (exp, target, tmode, modifier);
- }
- if (TREE_TYPE (TREE_TYPE (arg0)) == char_type_node)
- {
- exp = build_int_2 (TREE_STRING_POINTER (arg0)[i], 0);
- TREE_TYPE (exp) = integer_type_node;
- return expand_expr (convert (TREE_TYPE (TREE_TYPE (arg0)), exp), target, tmode, modifier);
- }
- }
- }
-
- /* If this is a constant index into a constant array,
- just get the value from the array. */
- if (TREE_READONLY (TREE_OPERAND (exp, 0))
- && ! TREE_SIDE_EFFECTS (TREE_OPERAND (exp, 0))
- && TREE_CODE (TREE_TYPE (TREE_OPERAND (exp, 0))) == ARRAY_TYPE
- && TREE_CODE (TREE_OPERAND (exp, 0)) == VAR_DECL
- && DECL_INITIAL (TREE_OPERAND (exp, 0))
- && TREE_CODE (DECL_INITIAL (TREE_OPERAND (exp, 0))) != ERROR_MARK)
- {
- tree index = fold (TREE_OPERAND (exp, 1));
- if (TREE_CODE (index) == INTEGER_CST)
- {
- int i = TREE_INT_CST_LOW (index);
- tree init = CONSTRUCTOR_ELTS (DECL_INITIAL (TREE_OPERAND (exp, 0)));
-
- while (init && i--)
- init = TREE_CHAIN (init);
- if (init)
- return expand_expr (fold (TREE_VALUE (init)), target, tmode, modifier);
- }
- }
- /* Treat array-ref with constant index as a component-ref. */
-
- case COMPONENT_REF:
- case BIT_FIELD_REF:
- {
- enum machine_mode mode1;
- int bitsize;
- int bitpos;
- int volatilep = 0;
- tree tem = get_inner_reference (exp, &bitsize, &bitpos,
- &mode1, &unsignedp, &volatilep);
-
- /* In some cases, we will be offsetting OP0's address by a constant.
- So get it as a sum, if possible. If we will be using it
- directly in an insn, we validate it. */
- op0 = expand_expr (tem, 0, VOIDmode, EXPAND_SUM);
-
- /* Don't forget about volatility even if this is a bitfield. */
- if (GET_CODE (op0) == MEM && volatilep && ! MEM_VOLATILE_P (op0))
- {
- op0 = copy_rtx (op0);
- MEM_VOLATILE_P (op0) = 1;
- }
-
- if (mode1 == VOIDmode
- || GET_CODE (op0) == REG || GET_CODE (op0) == SUBREG)
- return extract_bit_field (validize_mem (op0), bitsize, bitpos,
- unsignedp, target, mode, tmode,
- TYPE_ALIGN (TREE_TYPE (tem)) / BITS_PER_UNIT,
- int_size_in_bytes (TREE_TYPE (tem)));
- /* Get a reference to just this component. */
- if (modifier == EXPAND_CONST_ADDRESS)
- op0 = gen_rtx (MEM, mode1, plus_constant (XEXP (op0, 0),
- (bitpos / BITS_PER_UNIT)));
- else
- op0 = change_address (op0, mode1,
- plus_constant (XEXP (op0, 0),
- (bitpos / BITS_PER_UNIT)));
- MEM_IN_STRUCT_P (op0) = 1;
- MEM_VOLATILE_P (op0) |= volatilep;
- if (mode == mode1 || mode1 == BLKmode || mode1 == tmode)
- return op0;
- if (target == 0)
- target = gen_reg_rtx (tmode != VOIDmode ? tmode : mode);
- convert_move (target, op0, unsignedp);
- return target;
- }
-
- /* Intended for a reference to a buffer of a file-object in Pascal.
- But it's not certain that a special tree code will really be
- necessary for these. INDIRECT_REF might work for them. */
- case BUFFER_REF:
- abort ();
-
- case WITH_CLEANUP_EXPR:
- if (RTL_EXPR_RTL (exp) == 0)
- {
- RTL_EXPR_RTL (exp)
- = expand_expr (TREE_OPERAND (exp, 0), target, tmode, modifier);
- cleanups_this_call = tree_cons (0, TREE_OPERAND (exp, 2), cleanups_this_call);
- /* That's it for this cleanup. */
- TREE_OPERAND (exp, 2) = 0;
- }
- return RTL_EXPR_RTL (exp);
-
- case CALL_EXPR:
- /* Check for a built-in function. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == ADDR_EXPR
- && TREE_CODE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)) == FUNCTION_DECL
- && DECL_BUILT_IN (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))
- return expand_builtin (exp, target, subtarget, tmode, ignore);
- /* If this call was expanded already by preexpand_calls,
- just return the result we got. */
- if (CALL_EXPR_RTL (exp) != 0)
- return CALL_EXPR_RTL (exp);
- return expand_call (exp,
- (modifier == EXPAND_INTO_STACK) ? original_target : target,
- ignore, modifier);
-
- case NON_LVALUE_EXPR:
- case NOP_EXPR:
- case CONVERT_EXPR:
- case REFERENCE_EXPR:
- if (TREE_CODE (type) == VOID_TYPE || ignore)
- {
- expand_expr (TREE_OPERAND (exp, 0), const0_rtx, VOIDmode, modifier);
- return const0_rtx;
- }
- if (mode == TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- return expand_expr (TREE_OPERAND (exp, 0), target, VOIDmode, modifier);
- if (TREE_CODE (type) == UNION_TYPE)
- {
- tree valtype = TREE_TYPE (TREE_OPERAND (exp, 0));
- if (target == 0)
- {
- if (mode == BLKmode)
- {
- if (TYPE_SIZE (type) == 0
- || TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
- abort ();
- target = assign_stack_temp (BLKmode,
- (TREE_INT_CST_LOW (TYPE_SIZE (type))
- + BITS_PER_UNIT - 1)
- / BITS_PER_UNIT, 0);
- }
- else
- target = gen_reg_rtx (mode);
- }
- if (GET_CODE (target) == MEM)
- /* Store data into beginning of memory target. */
- store_expr (TREE_OPERAND (exp, 0),
- change_address (target, TYPE_MODE (valtype), 0), 0);
- else if (GET_CODE (target) == REG)
- /* Store this field into a union of the proper type. */
- store_field (target, GET_MODE_BITSIZE (TYPE_MODE (valtype)), 0,
- TYPE_MODE (valtype), TREE_OPERAND (exp, 0),
- VOIDmode, 0, 1,
- int_size_in_bytes (TREE_TYPE (TREE_OPERAND (exp, 0))));
- else
- abort ();
-
- /* Return the entire union. */
- return target;
- }
- op0 = expand_expr (TREE_OPERAND (exp, 0), 0, mode, 0);
- if (GET_MODE (op0) == mode || GET_MODE (op0) == VOIDmode)
- return op0;
- if (flag_force_mem && GET_CODE (op0) == MEM)
- op0 = copy_to_reg (op0);
-
- if (target == 0)
- return convert_to_mode (mode, op0, TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 0))));
- else
- convert_move (target, op0, TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 0))));
- return target;
-
- case PLUS_EXPR:
- /* We come here from MINUS_EXPR when the second operand is a constant. */
- plus_expr:
- this_optab = add_optab;
-
- /* If the result is to be Pmode and we are adding an integer to
- something, we might be forming a constant. So try to use
- plus_constant. If it produces a sum and we can't accept it,
- use force_operand. This allows P = &ARR[const] to generate
- efficient code on machines where a SYMBOL_REF is not a valid
- address.
-
- If this is an EXPAND_SUM call, always return the sum. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == INTEGER_CST
- && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_INT
- && (modifier == EXPAND_SUM || mode == Pmode))
- {
- op1 = expand_expr (TREE_OPERAND (exp, 1), subtarget, VOIDmode,
- EXPAND_SUM);
- op1 = plus_constant (op1, TREE_INT_CST_LOW (TREE_OPERAND (exp, 0)));
- if (modifier != EXPAND_SUM)
- op1 = force_operand (op1, target);
- return op1;
- }
-
- else if (TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST
- && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_INT
- && (modifier == EXPAND_SUM || mode == Pmode))
- {
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode,
- EXPAND_SUM);
- op0 = plus_constant (op0, TREE_INT_CST_LOW (TREE_OPERAND (exp, 1)));
- if (modifier != EXPAND_SUM)
- op0 = force_operand (op0, target);
- return op0;
- }
-
- /* No sense saving up arithmetic to be done
- if it's all in the wrong mode to form part of an address.
- And force_operand won't know whether to sign-extend or
- zero-extend. */
- if (modifier != EXPAND_SUM || mode != Pmode) goto binop;
-
- preexpand_calls (exp);
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
-
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, EXPAND_SUM);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, EXPAND_SUM);
- /* Put a sum last, to simplify what follows. */
- #ifdef OLD_INDEXING
- if (GET_CODE (op1) == MULT)
- {
- temp = op0;
- op0 = op1;
- op1 = temp;
- }
- #endif
- #ifndef OLD_INDEXING
- /* Make sure any term that's a sum with a constant comes last. */
- if (GET_CODE (op0) == PLUS
- && CONSTANT_P (XEXP (op0, 1)))
- {
- temp = op0;
- op0 = op1;
- op1 = temp;
- }
- /* If adding to a sum including a constant,
- associate it to put the constant outside. */
- if (GET_CODE (op1) == PLUS
- && CONSTANT_P (XEXP (op1, 1)))
- {
- rtx tem;
- int constant_term = 0;
-
- op0 = gen_rtx (PLUS, mode, XEXP (op1, 0), op0);
- /* Let's also eliminate constants from op0 if possible. */
- tem = eliminate_constant_term (op0, &constant_term);
- if (GET_CODE (XEXP (op1, 1)) == CONST_INT)
- {
- if (constant_term != 0)
- return plus_constant (tem, INTVAL (XEXP (op1, 1)) + constant_term);
- else
- return plus_constant (op0, INTVAL (XEXP (op1, 1)));
- }
- else
- return gen_rtx (PLUS, mode, op0, XEXP (op1, 1));
- }
- #endif
- /* Put a constant term last. */
- if (CONSTANT_P (op0))
- return gen_rtx (PLUS, mode, op1, op0);
- else
- return gen_rtx (PLUS, mode, op0, op1);
-
- case MINUS_EXPR:
- /* Handle difference of two symbolic constants,
- for the sake of an initializer. */
- if (modifier == EXPAND_SUM
- && really_constant_p (TREE_OPERAND (exp, 0))
- && really_constant_p (TREE_OPERAND (exp, 1)))
- {
- rtx op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode, modifier);
- rtx op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, modifier);
- return gen_rtx (MINUS, mode, op0, op1);
- }
- /* Convert A - const to A + (-const). */
- if (TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST)
- {
- exp = build (PLUS_EXPR, type, TREE_OPERAND (exp, 0),
- fold (build1 (NEGATE_EXPR, type,
- TREE_OPERAND (exp, 1))));
- goto plus_expr;
- }
- this_optab = sub_optab;
- goto binop;
-
- case MULT_EXPR:
- preexpand_calls (exp);
- /* If first operand is constant, swap them.
- Thus the following special case checks need only
- check the second operand. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == INTEGER_CST)
- {
- register tree t1 = TREE_OPERAND (exp, 0);
- TREE_OPERAND (exp, 0) = TREE_OPERAND (exp, 1);
- TREE_OPERAND (exp, 1) = t1;
- }
-
- /* Attempt to return something suitable for generating an
- indexed address, for machines that support that. */
-
- if (modifier == EXPAND_SUM && mode == Pmode
- && TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST
- && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_INT)
- {
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, EXPAND_SUM);
-
- /* Apply distributive law if OP0 is x+c. */
- if (GET_CODE (op0) == PLUS
- && GET_CODE (XEXP (op0, 1)) == CONST_INT)
- return gen_rtx (PLUS, mode,
- gen_rtx (MULT, mode, XEXP (op0, 0),
- gen_rtx (CONST_INT, VOIDmode,
- TREE_INT_CST_LOW (TREE_OPERAND (exp, 1)))),
- gen_rtx (CONST_INT, VOIDmode,
- (TREE_INT_CST_LOW (TREE_OPERAND (exp, 1))
- * INTVAL (XEXP (op0, 1)))));
-
- if (GET_CODE (op0) != REG)
- op0 = force_operand (op0, 0);
- if (GET_CODE (op0) != REG)
- op0 = copy_to_mode_reg (mode, op0);
-
- return gen_rtx (MULT, mode, op0,
- gen_rtx (CONST_INT, VOIDmode,
- TREE_INT_CST_LOW (TREE_OPERAND (exp, 1))));
- }
-
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
-
- /* Check for multiplying things that have been extended
- from a narrower type. If this machine supports multiplying
- in that narrower type with a result in the desired type,
- do it that way, and avoid the explicit type-conversion. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == NOP_EXPR
- && TREE_CODE (type) == INTEGER_TYPE
- && (TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))
- < TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (exp, 0))))
- && ((TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST
- && int_fits_type_p (TREE_OPERAND (exp, 1),
- TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))
- /* Don't use a widening multiply if a shift will do. */
- && ((GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 1))))
- > HOST_BITS_PER_INT)
- || exact_log2 (TREE_INT_CST_LOW (TREE_OPERAND (exp, 1))) < 0))
- ||
- (TREE_CODE (TREE_OPERAND (exp, 1)) == NOP_EXPR
- && (TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 1), 0)))
- ==
- TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0))))
- /* If both operands are extended, they must either both
- be zero-extended or both be sign-extended. */
- && (TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 1), 0)))
- ==
- TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))))))
- {
- enum machine_mode innermode
- = TYPE_MODE (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)));
- this_optab = (TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))
- ? umul_widen_optab : smul_widen_optab);
- if (mode == GET_MODE_WIDER_MODE (innermode)
- && this_optab->handlers[(int) mode].insn_code != CODE_FOR_nothing)
- {
- op0 = expand_expr (TREE_OPERAND (TREE_OPERAND (exp, 0), 0),
- 0, VOIDmode, 0);
- if (TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST)
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- else
- op1 = expand_expr (TREE_OPERAND (TREE_OPERAND (exp, 1), 0),
- 0, VOIDmode, 0);
- goto binop2;
- }
- }
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- return expand_mult (mode, op0, op1, target, unsignedp);
-
- case TRUNC_DIV_EXPR:
- case FLOOR_DIV_EXPR:
- case CEIL_DIV_EXPR:
- case ROUND_DIV_EXPR:
- case EXACT_DIV_EXPR:
- preexpand_calls (exp);
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
- /* Possible optimization: compute the dividend with EXPAND_SUM
- then if the divisor is constant can optimize the case
- where some terms of the dividend have coeffs divisible by it. */
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- return expand_divmod (0, code, mode, op0, op1, target, unsignedp);
-
- case RDIV_EXPR:
- this_optab = flodiv_optab;
- goto binop;
-
- case TRUNC_MOD_EXPR:
- case FLOOR_MOD_EXPR:
- case CEIL_MOD_EXPR:
- case ROUND_MOD_EXPR:
- preexpand_calls (exp);
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- return expand_divmod (1, code, mode, op0, op1, target, unsignedp);
- #if 0
- #ifdef HAVE_divmoddisi4
- if (GET_MODE (op0) != DImode)
- {
- temp = gen_reg_rtx (DImode);
- convert_move (temp, op0, 0);
- op0 = temp;
- if (GET_MODE (op1) != SImode && GET_CODE (op1) != CONST_INT)
- {
- temp = gen_reg_rtx (SImode);
- convert_move (temp, op1, 0);
- op1 = temp;
- }
- temp = gen_reg_rtx (SImode);
- if (target == 0)
- target = gen_reg_rtx (SImode);
- emit_insn (gen_divmoddisi4 (temp, protect_from_queue (op0, 0),
- protect_from_queue (op1, 0),
- protect_from_queue (target, 1)));
- return target;
- }
- #endif
- #endif
-
- case FIX_ROUND_EXPR:
- case FIX_FLOOR_EXPR:
- case FIX_CEIL_EXPR:
- abort (); /* Not used for C. */
-
- case FIX_TRUNC_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode, 0);
- if (target == 0)
- target = gen_reg_rtx (mode);
- expand_fix (target, op0, unsignedp);
- return target;
-
- case FLOAT_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode, 0);
- if (target == 0)
- target = gen_reg_rtx (mode);
- #ifdef NeXT
- /* expand_float can't figure out what to do if FROM has VOIDmode.
- So give it the correct mode. With -O, cse will optimize this. */
- if (GET_MODE (op0) == VOIDmode)
- op0 = copy_to_mode_reg (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))),
- op0);
- #endif /* NeXT */
- expand_float (target, op0,
- TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 0))));
- return target;
-
- case NEGATE_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), target, VOIDmode, 0);
- temp = expand_unop (mode, neg_optab, op0, target, 0);
- if (temp == 0)
- abort ();
- return temp;
-
- case ABS_EXPR:
- /* First try to do it with a special abs instruction.
- If that does not win, use conditional jump and negate. */
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- temp = expand_unop (mode, abs_optab, op0, target, 0);
- if (temp != 0)
- return temp;
-
- target = original_target;
- temp = gen_label_rtx ();
- if (target == 0 || ! safe_from_p (target, TREE_OPERAND (exp, 0)))
- target = gen_reg_rtx (mode);
- emit_move_insn (target, op0);
- emit_cmp_insn (target,
- expand_expr (convert (type, integer_zero_node),
- 0, VOIDmode, 0),
- GE, 0, 0, 0);
- NO_DEFER_POP;
- emit_jump_insn (gen_bge (temp));
- op0 = expand_unop (mode, neg_optab, target, target, 0);
- if (op0 != target)
- emit_move_insn (target, op0);
- emit_label (temp);
- OK_DEFER_POP;
- return target;
-
- case MAX_EXPR:
- case MIN_EXPR:
- target = original_target;
- mode = TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 1)));
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- if (target == 0 || ! safe_from_p (target, exp))
- target = gen_reg_rtx (mode);
- op0 = expand_expr (TREE_OPERAND (exp, 0), target, VOIDmode, 0);
- if (target != op0)
- emit_move_insn (target, op0);
- op0 = gen_label_rtx ();
- if (code == MAX_EXPR)
- temp = (TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 1)))
- ? compare_from_rtx (target, op1, GEU, 1, mode, 0, 0)
- : compare_from_rtx (target, op1, GE, 0, mode, 0, 0));
- else
- temp = (TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 1)))
- ? compare_from_rtx (target, op1, LEU, 1, mode, 0, 0)
- : compare_from_rtx (target, op1, LE, 0, mode, 0, 0));
- if (temp == const0_rtx)
- emit_move_insn (target, op1);
- else if (temp != const1_rtx)
- {
- if (bcc_gen_fctn[(int) GET_CODE (temp)] != 0)
- emit_jump_insn ((*bcc_gen_fctn[(int) GET_CODE (temp)]) (op0));
- else
- abort ();
- emit_move_insn (target, op1);
- }
- emit_label (op0);
- return target;
-
- /* ??? Can optimize when the operand of this is a bitwise operation,
- by using a different bitwise operation. */
- case BIT_NOT_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- temp = expand_unop (mode, one_cmpl_optab, op0, target, 1);
- if (temp == 0)
- abort ();
- return temp;
-
- case FFS_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- temp = expand_unop (mode, ffs_optab, op0, target, 1);
- if (temp == 0)
- abort ();
- return temp;
-
- /* ??? Can optimize bitwise operations with one arg constant.
- Can optimize (a bitwise1 n) bitwise2 (a bitwise3 b)
- and (a bitwise1 b) bitwise2 b (etc)
- but that is probably not worth while. */
-
- /* BIT_AND_EXPR is for bitwise anding.
- TRUTH_AND_EXPR is for anding two boolean values
- when we want in all cases to compute both of them.
- In general it is fastest to do TRUTH_AND_EXPR by
- computing both operands as actual zero-or-1 values
- and then bitwise anding. In cases where there cannot
- be any side effects, better code would be made by
- treating TRUTH_AND_EXPR like TRUTH_ANDIF_EXPR;
- but the question is how to recognize those cases. */
-
- case TRUTH_AND_EXPR:
- case BIT_AND_EXPR:
- this_optab = and_optab;
- goto binop;
-
- /* See comment above about TRUTH_AND_EXPR; it applies here too. */
- case TRUTH_OR_EXPR:
- case BIT_IOR_EXPR:
- this_optab = ior_optab;
- goto binop;
-
- case BIT_XOR_EXPR:
- this_optab = xor_optab;
- goto binop;
-
- case LSHIFT_EXPR:
- case RSHIFT_EXPR:
- case LROTATE_EXPR:
- case RROTATE_EXPR:
- preexpand_calls (exp);
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- return expand_shift (code, mode, op0, TREE_OPERAND (exp, 1), target,
- unsignedp);
-
- /* Could determine the answer when only additive constants differ.
- Also, the addition of one can be handled by changing the condition. */
- case LT_EXPR:
- case LE_EXPR:
- case GT_EXPR:
- case GE_EXPR:
- case EQ_EXPR:
- case NE_EXPR:
- preexpand_calls (exp);
- temp = do_store_flag (exp, target, tmode != VOIDmode ? tmode : mode, 0);
- if (temp != 0)
- return temp;
- /* For foo != 0, load foo, and if it is nonzero load 1 instead. */
- if (code == NE_EXPR && integer_zerop (TREE_OPERAND (exp, 1))
- && original_target
- && GET_CODE (original_target) == REG
- && (GET_MODE (original_target)
- == TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- {
- temp = expand_expr (TREE_OPERAND (exp, 0), original_target, VOIDmode, 0);
- if (temp != original_target)
- temp = copy_to_reg (temp);
- op1 = gen_label_rtx ();
- emit_cmp_insn (temp, const0_rtx, EQ, 0, unsignedp, 0);
- emit_jump_insn (gen_beq (op1));
- emit_move_insn (temp, const1_rtx);
- emit_label (op1);
- return temp;
- }
- /* If no set-flag instruction, must generate a conditional
- store into a temporary variable. Drop through
- and handle this like && and ||. */
-
- case TRUTH_ANDIF_EXPR:
- case TRUTH_ORIF_EXPR:
- if (target == 0 || ! safe_from_p (target, exp)
- /* Make sure we don't have a hard reg (such as function's return
- value) live across basic blocks, if not optimizing. */
- || (!optimize && GET_CODE (target) == REG
- && REGNO (target) < FIRST_PSEUDO_REGISTER))
- target = gen_reg_rtx (tmode != VOIDmode ? tmode : mode);
- emit_clr_insn (target);
- op1 = gen_label_rtx ();
- jumpifnot (exp, op1);
- emit_0_to_1_insn (target);
- emit_label (op1);
- return target;
-
- case TRUTH_NOT_EXPR:
- op0 = expand_expr (TREE_OPERAND (exp, 0), target, VOIDmode, 0);
- /* The parser is careful to generate TRUTH_NOT_EXPR
- only with operands that are always zero or one. */
- temp = expand_binop (mode, xor_optab, op0,
- gen_rtx (CONST_INT, mode, 1),
- target, 1, OPTAB_LIB_WIDEN);
- if (temp == 0)
- abort ();
- return temp;
-
- case COMPOUND_EXPR:
- expand_expr (TREE_OPERAND (exp, 0), const0_rtx, VOIDmode, 0);
- emit_queue ();
- return expand_expr (TREE_OPERAND (exp, 1),
- (ignore ? const0_rtx : target),
- VOIDmode, 0);
-
- case COND_EXPR:
- {
- /* Note that COND_EXPRs whose type is a structure or union
- are required to be constructed to contain assignments of
- a temporary variable, so that we can evaluate them here
- for side effect only. If type is void, we must do likewise. */
-
- /* If an arm of the branch requires a cleanup,
- only that cleanup is performed. */
-
- tree singleton = 0;
- tree binary_op = 0, unary_op = 0;
- tree old_cleanups = cleanups_this_call;
- cleanups_this_call = 0;
-
- /* If this is (A ? 1 : 0) and A is a condition, just evaluate it and
- convert it to our mode, if necessary. */
- if (integer_onep (TREE_OPERAND (exp, 1))
- && integer_zerop (TREE_OPERAND (exp, 2))
- && TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 0))) == '<')
- {
- op0 = expand_expr (TREE_OPERAND (exp, 0), target, mode, modifier);
- if (GET_MODE (op0) == mode)
- return op0;
- if (target == 0)
- target = gen_reg_rtx (mode);
- convert_move (target, op0, unsignedp);
- return target;
- }
-
- /* If we are not to produce a result, we have no target. Otherwise,
- if a target was specified use it; it will not be used as an
- intermediate target unless it is safe. If no target, use a
- temporary. */
-
- if (mode == VOIDmode || ignore)
- temp = 0;
- else if (original_target
- && safe_from_p (original_target, TREE_OPERAND (exp, 0)))
- temp = original_target;
- else if (mode == BLKmode)
- {
- if (TYPE_SIZE (type) == 0
- || TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
- abort ();
- temp = assign_stack_temp (BLKmode,
- (TREE_INT_CST_LOW (TYPE_SIZE (type))
- + BITS_PER_UNIT - 1)
- / BITS_PER_UNIT, 0);
- }
- else
- temp = gen_reg_rtx (mode);
-
- /* Check for X ? A + B : A. If we have this, we can copy
- A to the output and conditionally add B. Similarly for unary
- operations. */
- if (TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 1))) == '2'
- && operand_equal_p (TREE_OPERAND (exp, 2),
- TREE_OPERAND (TREE_OPERAND (exp, 1), 0), 0))
- singleton = TREE_OPERAND (exp, 2), binary_op = TREE_OPERAND (exp, 1);
- else if (TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 2))) == '2'
- && operand_equal_p (TREE_OPERAND (exp, 1),
- TREE_OPERAND (TREE_OPERAND (exp, 2), 0), 0))
- singleton = TREE_OPERAND (exp, 1), binary_op = TREE_OPERAND (exp, 2);
- else if (TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 1))) == '1'
- && operand_equal_p (TREE_OPERAND (exp, 2),
- TREE_OPERAND (TREE_OPERAND (exp, 1), 0), 0))
- singleton = TREE_OPERAND (exp, 2), unary_op = TREE_OPERAND (exp, 1);
- else if (TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 2))) == '1'
- && operand_equal_p (TREE_OPERAND (exp, 1),
- TREE_OPERAND (TREE_OPERAND (exp, 2), 0), 0))
- singleton = TREE_OPERAND (exp, 1), unary_op = TREE_OPERAND (exp, 2);
-
- /* If we had X ? A + 1 : A and we can do the test of X as a store-flag
- operation, do this as A + (X != 0). */
- if (singleton && binary_op && TREE_CODE (binary_op) == PLUS_EXPR
- && integer_onep (TREE_OPERAND (binary_op, 1))
- && TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 0))) == '<')
- {
- rtx result
- = do_store_flag (TREE_OPERAND (exp, 0),
- safe_from_p (temp, singleton) ? temp : 0,
- mode, 1);
-
- if (result)
- {
- op1 = expand_expr (singleton, 0, VOIDmode, 0);
- return expand_binop (mode, add_optab, result, op1, temp,
- unsignedp, OPTAB_LIB_WIDEN);
- }
- }
-
- NO_DEFER_POP;
- op0 = gen_label_rtx ();
-
- if (singleton)
- {
- if (temp != 0)
- {
- /* Don't allow a hard-register because evaluating the
- condition might involve a CALL which can clobber it. */
- if ((GET_CODE (temp) == REG
- && REGNO (temp) < FIRST_PSEUDO_REGISTER)
- /* If the target conflicts with the other operand of the
- binary op, we can't use it. */
- || (binary_op
- && ! safe_from_p (temp, TREE_OPERAND (binary_op, 1))))
- temp = gen_reg_rtx (mode);
- store_expr (singleton, temp, 0);
- }
- else
- expand_expr (singleton, ignore ? const1_rtx : 0, VOIDmode, 0);
- if (cleanups_this_call)
- {
- sorry ("aggregate value in COND_EXPR");
- cleanups_this_call = 0;
- }
- if (singleton == TREE_OPERAND (exp, 1))
- jumpif (TREE_OPERAND (exp, 0), op0);
- else
- jumpifnot (TREE_OPERAND (exp, 0), op0);
-
- if (binary_op && temp == 0)
- /* Just touch the other operand. */
- expand_expr (TREE_OPERAND (binary_op, 1),
- ignore ? const0_rtx : 0, VOIDmode, 0);
- else if (binary_op)
- store_expr (build (TREE_CODE (binary_op), type,
- make_tree (type, temp),
- TREE_OPERAND (binary_op, 1)),
- temp, 0);
- else
- store_expr (build1 (TREE_CODE (unary_op), type,
- make_tree (type, temp)),
- temp, 0);
- op1 = op0;
- }
- #if 0
- /* This is now done in jump.c and is better done there because it
- produces shorter register lifetimes. */
-
- /* Check for both possibilities either constants or variables
- in registers (but not the same as the target!). If so, can
- save branches by assigning one, branching, and assigning the
- other. */
- else if (temp && GET_MODE (temp) != BLKmode
- && (TREE_CONSTANT (TREE_OPERAND (exp, 1))
- || ((TREE_CODE (TREE_OPERAND (exp, 1)) == PARM_DECL
- || TREE_CODE (TREE_OPERAND (exp, 1)) == VAR_DECL)
- && DECL_RTL (TREE_OPERAND (exp, 1))
- && GET_CODE (DECL_RTL (TREE_OPERAND (exp, 1))) == REG
- && DECL_RTL (TREE_OPERAND (exp, 1)) != temp))
- && (TREE_CONSTANT (TREE_OPERAND (exp, 2))
- || ((TREE_CODE (TREE_OPERAND (exp, 2)) == PARM_DECL
- || TREE_CODE (TREE_OPERAND (exp, 2)) == VAR_DECL)
- && DECL_RTL (TREE_OPERAND (exp, 2))
- && GET_CODE (DECL_RTL (TREE_OPERAND (exp, 2))) == REG
- && DECL_RTL (TREE_OPERAND (exp, 2)) != temp)))
- {
- if (GET_CODE (temp) == REG && REGNO (temp) < FIRST_PSEUDO_REGISTER)
- temp = gen_reg_rtx (mode);
- store_expr (TREE_OPERAND (exp, 2), temp, 0);
- jumpifnot (TREE_OPERAND (exp, 0), op0);
- store_expr (TREE_OPERAND (exp, 1), temp, 0);
- op1 = op0;
- }
- #endif
- /* Check for A op 0 ? A : FOO and A op 0 ? FOO : A where OP is any
- comparison operator. If we have one of these cases, set the
- output to A, branch on A (cse will merge these two references),
- then set the output to FOO. */
- else if (temp
- && TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 0))) == '<'
- && integer_zerop (TREE_OPERAND (TREE_OPERAND (exp, 0), 1))
- && operand_equal_p (TREE_OPERAND (TREE_OPERAND (exp, 0), 0),
- TREE_OPERAND (exp, 1), 0)
- && safe_from_p (temp, TREE_OPERAND (exp, 2)))
- {
- if (GET_CODE (temp) == REG && REGNO (temp) < FIRST_PSEUDO_REGISTER)
- temp = gen_reg_rtx (mode);
- store_expr (TREE_OPERAND (exp, 1), temp, 0);
- jumpif (TREE_OPERAND (exp, 0), op0);
- store_expr (TREE_OPERAND (exp, 2), temp, 0);
- op1 = op0;
- }
- else if (temp
- && TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, 0))) == '<'
- && integer_zerop (TREE_OPERAND (TREE_OPERAND (exp, 0), 1))
- && operand_equal_p (TREE_OPERAND (TREE_OPERAND (exp, 0), 0),
- TREE_OPERAND (exp, 2), 0)
- && safe_from_p (temp, TREE_OPERAND (exp, 1)))
- {
- if (GET_CODE (temp) == REG && REGNO (temp) < FIRST_PSEUDO_REGISTER)
- temp = gen_reg_rtx (mode);
- store_expr (TREE_OPERAND (exp, 2), temp, 0);
- jumpifnot (TREE_OPERAND (exp, 0), op0);
- store_expr (TREE_OPERAND (exp, 1), temp, 0);
- op1 = op0;
- }
- else
- {
- op1 = gen_label_rtx ();
- jumpifnot (TREE_OPERAND (exp, 0), op0);
- if (temp != 0)
- store_expr (TREE_OPERAND (exp, 1), temp, 0);
- else
- expand_expr (TREE_OPERAND (exp, 1), ignore ? const0_rtx : 0,
- VOIDmode, 0);
- if (cleanups_this_call)
- {
- sorry ("aggregate value in COND_EXPR");
- cleanups_this_call = 0;
- }
-
- emit_queue ();
- emit_jump_insn (gen_jump (op1));
- emit_barrier ();
- emit_label (op0);
- if (temp != 0)
- store_expr (TREE_OPERAND (exp, 2), temp, 0);
- else
- expand_expr (TREE_OPERAND (exp, 2), ignore ? const0_rtx : 0,
- VOIDmode, 0);
- }
-
- if (cleanups_this_call)
- {
- sorry ("aggregate value in COND_EXPR");
- cleanups_this_call = 0;
- }
-
- emit_queue ();
- emit_label (op1);
- OK_DEFER_POP;
- cleanups_this_call = old_cleanups;
- return temp;
- }
-
- case TARGET_EXPR:
- {
- /* Something needs to be initialized, but we didn't know
- where that thing was when building the tree. For example,
- it could be the return value of a function, or a parameter
- to a function which lays down in the stack, or a temporary
- variable which must be passed by reference.
-
- We guarantee that the expression will either be constructed
- or copied into our original target. */
-
- tree slot = TREE_OPERAND (exp, 0);
-
- if (TREE_CODE (slot) != VAR_DECL)
- abort ();
-
- if (target == 0)
- {
- if (DECL_RTL (slot) != 0)
- target = DECL_RTL (slot);
- else
- {
- target = assign_stack_temp (mode, int_size_in_bytes (type), 0);
- /* All temp slots at this level must not conflict. */
- preserve_temp_slots (target);
- DECL_RTL (slot) = target;
- }
-
- #if 0
- /* Since SLOT is not known to the called function
- to belong to its stack frame, we must build an explicit
- cleanup. This case occurs when we must build up a reference
- to pass the reference as an argument. In this case,
- it is very likely that such a reference need not be
- built here. */
-
- if (TREE_OPERAND (exp, 2) == 0)
- TREE_OPERAND (exp, 2) = maybe_build_cleanup (slot);
- if (TREE_OPERAND (exp, 2))
- cleanups_this_call = tree_cons (0, TREE_OPERAND (exp, 2),
- cleanups_this_call);
- #endif
- }
- else
- {
- /* This case does occur, when expanding a parameter which
- needs to be constructed on the stack. The target
- is the actual stack address that we want to initialize.
- The function we call will perform the cleanup in this case. */
-
- DECL_RTL (slot) = target;
- }
-
- return expand_expr (TREE_OPERAND (exp, 1), target, tmode, modifier);
- }
-
- case INIT_EXPR:
- {
- tree lhs = TREE_OPERAND (exp, 0);
- tree rhs = TREE_OPERAND (exp, 1);
- tree noncopied_parts = 0;
- tree lhs_type = TREE_TYPE (lhs);
-
- temp = expand_assignment (lhs, rhs, ! ignore, original_target != 0);
- if (TYPE_NONCOPIED_PARTS (lhs_type) != 0 && !fixed_type_p (rhs))
- noncopied_parts = init_noncopied_parts (stabilize_reference (lhs),
- TYPE_NONCOPIED_PARTS (lhs_type));
- while (noncopied_parts != 0)
- {
- expand_assignment (TREE_VALUE (noncopied_parts),
- TREE_PURPOSE (noncopied_parts), 0, 0);
- noncopied_parts = TREE_CHAIN (noncopied_parts);
- }
- return temp;
- }
-
- case MODIFY_EXPR:
- {
- /* If lhs is complex, expand calls in rhs before computing it.
- That's so we don't compute a pointer and save it over a call.
- If lhs is simple, compute it first so we can give it as a
- target if the rhs is just a call. This avoids an extra temp and copy
- and that prevents a partial-subsumption which makes bad code.
- Actually we could treat component_ref's of vars like vars. */
-
- tree lhs = TREE_OPERAND (exp, 0);
- tree rhs = TREE_OPERAND (exp, 1);
- tree noncopied_parts = 0;
- tree lhs_type = TREE_TYPE (lhs);
-
- temp = 0;
-
- if (TREE_CODE (lhs) != VAR_DECL
- && TREE_CODE (lhs) != RESULT_DECL
- && TREE_CODE (lhs) != PARM_DECL)
- preexpand_calls (exp);
-
- /* Check for |= or &= of a bitfield of size one into another bitfield
- of size 1. In this case, (unless we need the result of the
- assignment) we can do this more efficiently with a
- test followed by an assignment, if necessary.
-
- ??? At this point, we can't get a BIT_FIELD_REF here. But if
- things change so we do, this code should be enhanced to
- support it. */
- if (ignore
- && TREE_CODE (lhs) == COMPONENT_REF
- && (TREE_CODE (rhs) == BIT_IOR_EXPR
- || TREE_CODE (rhs) == BIT_AND_EXPR)
- && TREE_OPERAND (rhs, 0) == lhs
- && TREE_CODE (TREE_OPERAND (rhs, 1)) == COMPONENT_REF
- && TREE_INT_CST_LOW (DECL_SIZE (TREE_OPERAND (lhs, 1))) == 1
- && TREE_INT_CST_LOW (DECL_SIZE (TREE_OPERAND (TREE_OPERAND (rhs, 1), 1))) == 1)
- {
- rtx label = gen_label_rtx ();
-
- do_jump (TREE_OPERAND (rhs, 1),
- TREE_CODE (rhs) == BIT_IOR_EXPR ? label : 0,
- TREE_CODE (rhs) == BIT_AND_EXPR ? label : 0);
- expand_assignment (lhs, convert (TREE_TYPE (rhs),
- (TREE_CODE (rhs) == BIT_IOR_EXPR
- ? integer_one_node
- : integer_zero_node)),
- 0, 0);
- emit_label (label);
- return const0_rtx;
- }
-
- if (TYPE_NONCOPIED_PARTS (lhs_type) != 0
- && ! (fixed_type_p (lhs) && fixed_type_p (rhs)))
- noncopied_parts = save_noncopied_parts (stabilize_reference (lhs),
- TYPE_NONCOPIED_PARTS (lhs_type));
-
- temp = expand_assignment (lhs, rhs, ! ignore, original_target != 0);
- while (noncopied_parts != 0)
- {
- expand_assignment (TREE_PURPOSE (noncopied_parts),
- TREE_VALUE (noncopied_parts), 0, 0);
- noncopied_parts = TREE_CHAIN (noncopied_parts);
- }
- return temp;
- }
-
- case PREINCREMENT_EXPR:
- case PREDECREMENT_EXPR:
- return expand_increment (exp, 0);
-
- case POSTINCREMENT_EXPR:
- case POSTDECREMENT_EXPR:
- /* Faster to treat as pre-increment if result is not used. */
- return expand_increment (exp, ! ignore);
-
- case ADDR_EXPR:
- /* Are we taking the address of a nested function? */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == FUNCTION_DECL
- && decl_function_context (TREE_OPERAND (exp, 0)) != 0)
- return trampoline_address (TREE_OPERAND (exp, 0));
- op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode,
- (modifier != EXPAND_INTO_STACK
- ? EXPAND_CONST_ADDRESS
- : EXPAND_INTO_STACK));
- if (GET_CODE (op0) != MEM)
- abort ();
-
- if (modifier == EXPAND_SUM)
- return XEXP (op0, 0);
- op0 = force_operand (XEXP (op0, 0), target);
- if (flag_force_addr && GET_CODE (op0) != REG)
- return force_reg (Pmode, op0);
- return op0;
-
- case ENTRY_VALUE_EXPR:
- abort ();
-
- case ERROR_MARK:
- return const0_rtx;
-
- default:
- return (*lang_expand_expr) (exp, target, tmode, modifier);
- }
-
- /* Here to do an ordinary binary operator, generating an instruction
- from the optab already placed in `this_optab'. */
- binop:
- preexpand_calls (exp);
- if (! safe_from_p (subtarget, TREE_OPERAND (exp, 1)))
- subtarget = 0;
- op0 = expand_expr (TREE_OPERAND (exp, 0), subtarget, VOIDmode, 0);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- binop2:
- temp = expand_binop (mode, this_optab, op0, op1, target,
- unsignedp, OPTAB_LIB_WIDEN);
- if (temp == 0)
- abort ();
- return temp;
- }
-
- /* Return the alignment of EXP, a pointer valued expression for the mem*
- builtin functions. Alignments greater than MAX_ALIGN are not significant.
- The alignment returned is, by default, the alignment of the thing that
- EXP points to (if it is not a POINTER_TYPE, 0 is returned).
-
- Otherwise, look at the expression to see if we can do better, i.e., if the
- expression is actually pointing at an object whose alignment is tighter. */
-
- static int
- get_pointer_alignment (exp, max_align)
- tree exp;
- int max_align;
- {
- int align, inner;
-
- if (TREE_CODE (TREE_TYPE (exp)) != POINTER_TYPE)
- return 0;
-
- align = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (exp)));
- align = MIN (align, max_align);
-
- while (1)
- {
- switch (TREE_CODE (exp))
- {
- case NOP_EXPR:
- case CONVERT_EXPR:
- case NON_LVALUE_EXPR:
- exp = TREE_OPERAND (exp, 0);
- if (TREE_CODE (TREE_TYPE (exp)) != POINTER_TYPE)
- return align;
- inner = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (exp)));
- inner = MIN (inner, max_align);
- align = MAX (align, inner);
- break;
-
- case PLUS_EXPR:
- /* If sum of pointer + int, restrict our maximum alignment to that
- imposed by the integer. If not, we can't do any better than
- ALIGN. */
- if (TREE_CODE (TREE_OPERAND (exp, 1)) != INTEGER_CST)
- return align;
-
- while ((TREE_INT_CST_LOW (TREE_OPERAND (exp, 1))
- & (max_align - 1)) != 0)
- max_align >>= 1;
-
- exp = TREE_OPERAND (exp, 0);
- break;
-
- case ADDR_EXPR:
- /* See what we are pointing at and look at its alignment. */
- exp = TREE_OPERAND (exp, 0);
- if (TREE_CODE_CLASS (TREE_CODE (exp)) == 'd')
- align = MAX (align, DECL_ALIGN (exp));
- #ifdef CONSTANT_ALIGNMENT
- else if (TREE_CODE_CLASS (TREE_CODE (exp)) == 'c')
- align = CONSTANT_ALIGNMENT (exp, align);
- #endif
- return MIN (align, max_align);
-
- default:
- return align;
- }
- }
- }
-
- /* Return the tree node and offset if a given argument corresponds to
- a string constant. */
-
- static tree
- string_constant (arg, ptr_offset)
- tree arg;
- tree *ptr_offset;
- {
- while (TREE_CODE (arg) == NOP_EXPR
- || TREE_CODE (arg) == CONVERT_EXPR
- || TREE_CODE (arg) == NON_LVALUE_EXPR)
- arg = TREE_OPERAND (arg, 0);
-
- if (TREE_CODE (arg) == ADDR_EXPR
- && TREE_CODE (TREE_OPERAND (arg, 0)) == STRING_CST)
- {
- *ptr_offset = integer_zero_node;
- return TREE_OPERAND (arg, 0);
- }
- else if (TREE_CODE (arg) == PLUS_EXPR)
- {
- if (TREE_CODE (TREE_OPERAND (arg, 0)) == ADDR_EXPR
- && TREE_CODE (TREE_OPERAND (TREE_OPERAND (arg, 0), 0)) == STRING_CST)
- {
- *ptr_offset = TREE_OPERAND (arg, 1);
- return TREE_OPERAND (TREE_OPERAND (arg, 0), 0);
- }
- else if (TREE_CODE (TREE_OPERAND (arg, 1)) == ADDR_EXPR
- && TREE_CODE (TREE_OPERAND (TREE_OPERAND (arg, 1), 0)) == STRING_CST)
- {
- *ptr_offset = TREE_OPERAND (arg, 0);
- return TREE_OPERAND (TREE_OPERAND (arg, 1), 0);
- }
- }
-
- return 0;
- }
-
- /* Expand an expression EXP that calls a built-in function,
- with result going to TARGET if that's convenient
- (and in mode MODE if that's convenient).
- SUBTARGET may be used as the target for computing one of EXP's operands.
- IGNORE is nonzero if the value is to be ignored. */
-
- static rtx
- expand_builtin (exp, target, subtarget, mode, ignore)
- tree exp;
- rtx target;
- rtx subtarget;
- enum machine_mode mode;
- int ignore;
- {
- tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
- tree arglist = TREE_OPERAND (exp, 1);
- rtx op0;
-
- switch (DECL_FUNCTION_CODE (fndecl))
- {
- case BUILT_IN_ABS:
- case BUILT_IN_LABS:
- case BUILT_IN_FABS:
- /* build_function_call changes these into ABS_EXPR. */
- abort ();
-
- case BUILT_IN_SAVEREGS:
- /* Don't do __builtin_saveregs more than once in a function.
- Save the result of the first call and reuse it. */
- if (saveregs_value != 0)
- return saveregs_value;
- {
- /* When this function is called, it means that registers must be
- saved on entry to this function. So we migrate the
- call to the first insn of this function. */
- rtx temp;
- rtx seq;
- /* Now really call the function. `expand_call' does not call
- expand_builtin, so there is no danger of infinite recursion here. */
- start_sequence ();
- #ifdef EXPAND_BUILTIN_SAVEREGS
- /* Do whatever the machine needs done in this case. */
- temp = EXPAND_BUILTIN_SAVEREGS (arglist);
- #else
- temp = expand_call (exp, target, ignore);
- #endif
- seq = get_insns ();
- end_sequence ();
-
- saveregs_value = temp;
-
- /* This won't work inside a SEQUENCE--it really has to be
- at the start of the function. */
- if (in_sequence_p ())
- {
- /* Better to do this than to crash. */
- error ("`va_start' used within `({...})'");
- return temp;
- }
-
- /* Put the sequence after the NOTE that starts the function. */
- emit_insns_before (seq, NEXT_INSN (get_insns ()));
- return temp;
- }
-
- case BUILT_IN_NEXT_ARG:
- {
- tree fntype = TREE_TYPE (current_function_decl);
- if (!(TYPE_ARG_TYPES (fntype) != 0
- && (TREE_VALUE (tree_last (TYPE_ARG_TYPES (fntype)))
- != void_type_node)))
- {
- error ("`stdarg.h' facilities used, but function has fixed args");
- return const0_rtx;
- }
- }
-
- return expand_binop (Pmode, add_optab,
- virtual_incoming_args_rtx,
- current_function_arg_offset_rtx,
- 0, 0, OPTAB_LIB_WIDEN);
-
- case BUILT_IN_CLASSIFY_TYPE:
- if (arglist != 0)
- {
- tree type = TREE_TYPE (TREE_VALUE (arglist));
- enum tree_code code = TREE_CODE (type);
- if (code == VOID_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, void_type_class);
- if (code == INTEGER_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, integer_type_class);
- if (code == CHAR_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, char_type_class);
- if (code == ENUMERAL_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, enumeral_type_class);
- if (code == BOOLEAN_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, boolean_type_class);
- if (code == POINTER_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, pointer_type_class);
- if (code == REFERENCE_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, reference_type_class);
- if (code == OFFSET_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, offset_type_class);
- if (code == REAL_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, real_type_class);
- if (code == COMPLEX_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, complex_type_class);
- if (code == FUNCTION_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, function_type_class);
- if (code == METHOD_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, method_type_class);
- if (code == RECORD_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, record_type_class);
- if (code == UNION_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, union_type_class);
- if (code == ARRAY_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, array_type_class);
- if (code == STRING_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, string_type_class);
- if (code == SET_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, set_type_class);
- if (code == FILE_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, file_type_class);
- if (code == LANG_TYPE)
- return gen_rtx (CONST_INT, VOIDmode, lang_type_class);
- }
- return gen_rtx (CONST_INT, VOIDmode, no_type_class);
-
- case BUILT_IN_CONSTANT_P:
- return (TREE_CODE_CLASS (TREE_VALUE (arglist)) == 'c'
- ? const1_rtx : const0_rtx);
-
- case BUILT_IN_ALLOCA:
- if (arglist == 0
- /* Arg could be non-integer if user redeclared this fcn wrong. */
- || TREE_CODE (TREE_TYPE (TREE_VALUE (arglist))) != INTEGER_TYPE)
- return const0_rtx;
- current_function_calls_alloca = 1;
- /* Compute the argument. */
- op0 = expand_expr (TREE_VALUE (arglist), 0, VOIDmode, 0);
-
- /* Allocate the desired space. */
- target = allocate_dynamic_stack_space (op0, target);
-
- /* Record the new stack level for nonlocal gotos. */
- if (nonlocal_goto_stack_level != 0)
- emit_move_insn (nonlocal_goto_stack_level, stack_pointer_rtx);
- return target;
-
- case BUILT_IN_FFS:
- if (arglist == 0
- /* Arg could be non-integer if user redeclared this fcn wrong. */
- || TREE_CODE (TREE_TYPE (TREE_VALUE (arglist))) != INTEGER_TYPE)
- return const0_rtx;
-
- /* Compute the argument. */
- op0 = expand_expr (TREE_VALUE (arglist), subtarget, VOIDmode, 0);
- /* Compute ffs, into TARGET if possible.
- Set TARGET to wherever the result comes back. */
- target = expand_unop (TYPE_MODE (TREE_TYPE (TREE_VALUE (arglist))),
- ffs_optab, op0, target, 1);
- if (target == 0)
- abort ();
- return target;
-
- case BUILT_IN_STRCPY:
- {
- tree src = TREE_VALUE (TREE_CHAIN (arglist));
- tree offset;
- tree len;
-
- src = string_constant (src, &offset);
- if (src == 0)
- break;
-
- len = size_binop (MINUS_EXPR,
- build_int_2 (TREE_STRING_LENGTH (src), 0),
- offset);
-
- chainon (arglist, build_tree_list (0, len));
- }
-
- /* Drops in. */
- case BUILT_IN_MEMCPY:
- {
- tree dest = TREE_VALUE (arglist);
- tree src = TREE_VALUE (TREE_CHAIN (arglist));
- tree len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
-
- int src_align
- = get_pointer_alignment (src, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
- int dest_align
- = get_pointer_alignment (dest, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
- rtx dest_rtx;
-
- /* If either SRC or DEST is not a pointer type, don't do
- this operation in-line. */
- if (src_align == 0 || dest_align == 0)
- break;
-
- dest_rtx = expand_expr (dest, 0, Pmode, EXPAND_NORMAL);
-
- /* Copy word part most expediently. */
- emit_block_move (gen_rtx (MEM, BLKmode, dest_rtx),
- gen_rtx (MEM, BLKmode,
- expand_expr (src, 0, Pmode, EXPAND_NORMAL)),
- expand_expr (len, 0, VOIDmode, 0),
- MIN (src_align, dest_align));
- return dest_rtx;
- }
-
- /* These comparison functions need an instruction that returns an actual
- index. An ordinary compare that just sets the condition codes
- is not enough. */
- #ifdef HAVE_cmpstrsi
- case BUILT_IN_STRCMP:
- {
- tree arg1 = TREE_VALUE (arglist);
- tree arg2 = TREE_VALUE (TREE_CHAIN (arglist));
- tree offset;
- tree len = 0;
-
- arg1 = string_constant (arg1, &offset);
- if (arg1)
- len = size_binop (MINUS_EXPR,
- build_int_2 (TREE_STRING_LENGTH (arg1), 0),
- offset);
- else
- {
- arg2 = string_constant (arg2, &offset);
- if (arg2)
- len = size_binop (MINUS_EXPR,
- build_int_2 (TREE_STRING_LENGTH (arg2), 0),
- offset);
- }
-
- if (len == 0)
- break;
-
- chainon (arglist, build_tree_list (0, len));
- }
-
- /* Drops in. */
- case BUILT_IN_MEMCMP:
- {
- tree arg1 = TREE_VALUE (arglist);
- tree arg2 = TREE_VALUE (TREE_CHAIN (arglist));
- tree len = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist)));
-
- int arg1_align
- = get_pointer_alignment (arg1, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
- int arg2_align
- = get_pointer_alignment (arg2, BIGGEST_ALIGNMENT) / BITS_PER_UNIT;
- rtx result;
-
- /* If we don't have POINTER_TYPE, call the function. */
- if (arg1_align == 0 || arg2_align == 0)
- break;
-
- result = gen_reg_rtx (mode != VOIDmode ? mode
- : mode_for_size (BITS_PER_WORD, MODE_INT, 0));
- emit_insn (gen_cmpstrsi (result,
- gen_rtx (MEM, BLKmode,
- expand_expr (arg1, 0, Pmode, EXPAND_NORMAL)),
- gen_rtx (MEM, BLKmode,
- expand_expr (arg2, 0, Pmode, EXPAND_NORMAL)),
- expand_expr (len, 0, VOIDmode, 0),
- gen_rtx (CONST_INT, VOIDmode,
- MIN (arg1_align, arg2_align))));
- return result;
- }
- #else
- case BUILT_IN_STRCMP:
- case BUILT_IN_MEMCMP:
- break;
- #endif
-
- default: /* just do library call, if unknown builtin */
- error ("built-in function %s not currently supported",
- IDENTIFIER_POINTER (DECL_NAME (fndecl)));
- }
-
- /* The switch statement above can drop through to cause the function
- to be called normally. */
-
- return expand_call (exp, target, ignore);
- }
-
- /* Expand code for a post- or pre- increment or decrement
- and return the RTX for the result.
- POST is 1 for postinc/decrements and 0 for preinc/decrements. */
-
- static rtx
- expand_increment (exp, post)
- register tree exp;
- int post;
- {
- register rtx op0, op1;
- register rtx temp, value;
- register tree incremented = TREE_OPERAND (exp, 0);
- optab this_optab = add_optab;
- int icode;
- enum machine_mode mode = TYPE_MODE (TREE_TYPE (exp));
- int op0_is_copy = 0;
-
- /* Stabilize any component ref that might need to be
- evaluated more than once below. */
- if (TREE_CODE (incremented) == BIT_FIELD_REF
- || (TREE_CODE (incremented) == COMPONENT_REF
- && (TREE_CODE (TREE_OPERAND (incremented, 0)) != INDIRECT_REF
- || DECL_BIT_FIELD (TREE_OPERAND (incremented, 1)))))
- incremented = stabilize_reference (incremented);
-
- /* Compute the operands as RTX.
- Note whether OP0 is the actual lvalue or a copy of it:
- I believe it is a copy iff it is a register and insns were
- generated in computing it. */
- temp = get_last_insn ();
- op0 = expand_expr (incremented, 0, VOIDmode, 0);
- if (temp != get_last_insn ())
- op0_is_copy = (GET_CODE (op0) == REG || GET_CODE (op0) == SUBREG);
- op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
-
- /* Decide whether incrementing or decrementing. */
- if (TREE_CODE (exp) == POSTDECREMENT_EXPR
- || TREE_CODE (exp) == PREDECREMENT_EXPR)
- this_optab = sub_optab;
-
- /* If OP0 is not the actual lvalue, but rather a copy in a register,
- then we cannot just increment OP0. We must
- therefore contrive to increment the original value.
- Then we can return OP0 since it is a copy of the old value. */
- if (op0_is_copy)
- {
- /* This is the easiest way to increment the value wherever it is.
- Problems with multiple evaluation of INCREMENTED
- are prevented because either (1) it is a component_ref,
- in which case it was stabilized above, or (2) it is an array_ref
- with constant index in an array in a register, which is
- safe to reevaluate. */
- tree newexp = build ((this_optab == add_optab
- ? PLUS_EXPR : MINUS_EXPR),
- TREE_TYPE (exp),
- incremented,
- TREE_OPERAND (exp, 1));
- temp = expand_assignment (incremented, newexp, ! post, 0);
- return post ? op0 : temp;
- }
-
- /* Convert decrement by a constant into a negative increment. */
- if (this_optab == sub_optab
- && GET_CODE (op1) == CONST_INT)
- {
- op1 = gen_rtx (CONST_INT, VOIDmode, - INTVAL (op1));
- this_optab = add_optab;
- }
-
- if (post)
- {
- /* We have a true reference to the value in OP0.
- If there is an insn to add or subtract in this mode, queue it. */
-
- #if 0 /* Turned off to avoid making extra insn for indexed memref. */
- op0 = stabilize (op0);
- #endif
-
- icode = (int) this_optab->handlers[(int) mode].insn_code;
- if (icode != (int) CODE_FOR_nothing
- /* Make sure that OP0 is valid for operands 0 and 1
- of the insn we want to queue. */
- && (*insn_operand_predicate[icode][0]) (op0, mode)
- && (*insn_operand_predicate[icode][1]) (op0, mode))
- {
- if (! (*insn_operand_predicate[icode][2]) (op1, mode))
- op1 = force_reg (mode, op1);
-
- return enqueue_insn (op0, GEN_FCN (icode) (op0, op0, op1));
- }
- }
-
- /* Preincrement, or we can't increment with one simple insn. */
- if (post)
- /* Save a copy of the value before inc or dec, to return it later. */
- temp = value = copy_to_reg (op0);
- else
- /* Arrange to return the incremented value. */
- /* Copy the rtx because expand_binop will protect from the queue,
- and the results of that would be invalid for us to return
- if our caller does emit_queue before using our result. */
- temp = copy_rtx (value = op0);
-
- /* Increment however we can. */
- op1 = expand_binop (mode, this_optab, value, op1, op0,
- TREE_UNSIGNED (TREE_TYPE (exp)), OPTAB_LIB_WIDEN);
- /* Make sure the value is stored into OP0. */
- if (op1 != op0)
- emit_move_insn (op0, op1);
-
- return temp;
- }
-
- /* Expand all function calls contained within EXP, innermost ones first.
- But don't look within expressions that have sequence points.
- For each CALL_EXPR, record the rtx for its value
- in the CALL_EXPR_RTL field. */
-
- static void
- preexpand_calls (exp)
- tree exp;
- {
- register int nops, i;
- int type = TREE_CODE_CLASS (TREE_CODE (exp));
-
- if (! do_preexpand_calls)
- return;
-
- /* Only expressions and references can contain calls. */
-
- if (type != 'e' && type != '<' && type != '1' && type != '2' && type != 'r')
- return;
-
- switch (TREE_CODE (exp))
- {
- case CALL_EXPR:
- /* Do nothing if already expanded. */
- if (CALL_EXPR_RTL (exp) != 0)
- return;
-
- /* Do nothing to built-in functions. */
- if (TREE_CODE (TREE_OPERAND (exp, 0)) != ADDR_EXPR
- || TREE_CODE (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)) != FUNCTION_DECL
- || ! DECL_BUILT_IN (TREE_OPERAND (TREE_OPERAND (exp, 0), 0)))
- CALL_EXPR_RTL (exp) = expand_call (exp, 0, 0, 0);
- return;
-
- case COMPOUND_EXPR:
- case COND_EXPR:
- case TRUTH_ANDIF_EXPR:
- case TRUTH_ORIF_EXPR:
- /* If we find one of these, then we can be sure
- the adjust will be done for it (since it makes jumps).
- Do it now, so that if this is inside an argument
- of a function, we don't get the stack adjustment
- after some other args have already been pushed. */
- do_pending_stack_adjust ();
- return;
-
- case BLOCK:
- case RTL_EXPR:
- return;
-
- case SAVE_EXPR:
- if (SAVE_EXPR_RTL (exp) != 0)
- return;
- }
-
- nops = tree_code_length[(int) TREE_CODE (exp)];
- for (i = 0; i < nops; i++)
- if (TREE_OPERAND (exp, i) != 0)
- {
- type = TREE_CODE_CLASS (TREE_CODE (TREE_OPERAND (exp, i)));
- if (type == 'e' || type == '<' || type == '1' || type == '2'
- || type == 'r')
- preexpand_calls (TREE_OPERAND (exp, i));
- }
- }
-
- /* At the start of a function, record that we have no previously-pushed
- arguments waiting to be popped. */
-
- void
- init_pending_stack_adjust ()
- {
- pending_stack_adjust = 0;
- }
-
- /* When exiting from function, if safe, clear out any pending stack adjust
- so the adjustment won't get done. */
-
- void
- clear_pending_stack_adjust ()
- {
- #ifdef EXIT_IGNORE_STACK
- if (!flag_omit_frame_pointer && EXIT_IGNORE_STACK
- && ! (TREE_INLINE (current_function_decl) && optimize)
- && ! flag_inline_functions)
- pending_stack_adjust = 0;
- #endif
- }
-
- /* Pop any previously-pushed arguments that have not been popped yet. */
-
- void
- do_pending_stack_adjust ()
- {
- if (inhibit_defer_pop == 0)
- {
- if (pending_stack_adjust != 0)
- adjust_stack (gen_rtx (CONST_INT, VOIDmode, pending_stack_adjust));
- pending_stack_adjust = 0;
- }
- }
-
- /* Expand all cleanups up to OLD_CLEANUPS.
- Needed here, and also for language-dependent calls. */
-
- void
- expand_cleanups_to (old_cleanups)
- tree old_cleanups;
- {
- while (cleanups_this_call != old_cleanups)
- {
- expand_expr (TREE_VALUE (cleanups_this_call), 0, VOIDmode, 0);
- cleanups_this_call = TREE_CHAIN (cleanups_this_call);
- }
- }
-
- /* Expand conditional expressions. */
-
- /* Generate code to evaluate EXP and jump to LABEL if the value is zero.
- LABEL is an rtx of code CODE_LABEL, in this function and all the
- functions here. */
-
- void
- jumpifnot (exp, label)
- tree exp;
- rtx label;
- {
- do_jump (exp, label, 0);
- }
-
- /* Generate code to evaluate EXP and jump to LABEL if the value is nonzero. */
-
- void
- jumpif (exp, label)
- tree exp;
- rtx label;
- {
- do_jump (exp, 0, label);
- }
-
- /* Generate code to evaluate EXP and jump to IF_FALSE_LABEL if
- the result is zero, or IF_TRUE_LABEL if the result is one.
- Either of IF_FALSE_LABEL and IF_TRUE_LABEL may be zero,
- meaning fall through in that case.
-
- This function is responsible for optimizing cases such as
- &&, || and comparison operators in EXP. */
-
- void
- do_jump (exp, if_false_label, if_true_label)
- tree exp;
- rtx if_false_label, if_true_label;
- {
- register enum tree_code code = TREE_CODE (exp);
- /* Some cases need to create a label to jump to
- in order to properly fall through.
- These cases set DROP_THROUGH_LABEL nonzero. */
- rtx drop_through_label = 0;
- rtx temp;
- rtx comparison = 0;
- int i;
- tree type;
-
- emit_queue ();
-
- switch (code)
- {
- case ERROR_MARK:
- break;
-
- case INTEGER_CST:
- temp = integer_zerop (exp) ? if_false_label : if_true_label;
- if (temp)
- emit_jump (temp);
- break;
-
- #if 0
- /* This is not true with #pragma weak */
- case ADDR_EXPR:
- /* The address of something can never be zero. */
- if (if_true_label)
- emit_jump (if_true_label);
- break;
- #endif
-
- case NOP_EXPR:
- if (TREE_CODE (TREE_OPERAND (exp, 0)) == COMPONENT_REF
- || TREE_CODE (TREE_OPERAND (exp, 0)) == BIT_FIELD_REF
- || TREE_CODE (TREE_OPERAND (exp, 0)) == ARRAY_REF)
- goto normal;
- case CONVERT_EXPR:
- /* If we are narrowing the operand, we have to do the compare in the
- narrower mode. */
- if ((TYPE_PRECISION (TREE_TYPE (exp))
- < TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- goto normal;
- case NON_LVALUE_EXPR:
- case REFERENCE_EXPR:
- case ABS_EXPR:
- case NEGATE_EXPR:
- case LROTATE_EXPR:
- case RROTATE_EXPR:
- /* These cannot change zero->non-zero or vice versa. */
- do_jump (TREE_OPERAND (exp, 0), if_false_label, if_true_label);
- break;
-
- #if 0
- /* This is never less insns than evaluating the PLUS_EXPR followed by
- a test and can be longer if the test is eliminated. */
- case PLUS_EXPR:
- /* Reduce to minus. */
- exp = build (MINUS_EXPR, TREE_TYPE (exp),
- TREE_OPERAND (exp, 0),
- fold (build1 (NEGATE_EXPR, TREE_TYPE (TREE_OPERAND (exp, 1)),
- TREE_OPERAND (exp, 1))));
- /* Process as MINUS. */
- #endif
-
- case MINUS_EXPR:
- /* Non-zero iff operands of minus differ. */
- comparison = compare (build (NE_EXPR, TREE_TYPE (exp),
- TREE_OPERAND (exp, 0),
- TREE_OPERAND (exp, 1)),
- NE, NE);
- break;
-
- case BIT_AND_EXPR:
- /* If we are AND'ing with a small constant, do this comparison in the
- smallest type that fits. If the machine doesn't have comparisons
- that small, it will be converted back to the wider comparison.
- This helps if we are testing the sign bit of a narrower object.
- combine can't do this for us because it can't know whether a
- ZERO_EXTRACT or a compare in a smaller mode exists, but we do. */
-
- if (TREE_CODE (TREE_OPERAND (exp, 1)) == INTEGER_CST
- && TYPE_PRECISION (TREE_TYPE (exp)) <= HOST_BITS_PER_INT
- && (i = floor_log2 (TREE_INT_CST_LOW (TREE_OPERAND (exp, 1)))) >= 0
- && (type = type_for_size (i + 1, 1)) != 0
- && TYPE_PRECISION (type) < TYPE_PRECISION (TREE_TYPE (exp)))
- {
- do_jump (convert (type, exp), if_false_label, if_true_label);
- break;
- }
- goto normal;
-
- case TRUTH_NOT_EXPR:
- do_jump (TREE_OPERAND (exp, 0), if_true_label, if_false_label);
- break;
-
- case TRUTH_ANDIF_EXPR:
- if (if_false_label == 0)
- if_false_label = drop_through_label = gen_label_rtx ();
- do_jump (TREE_OPERAND (exp, 0), if_false_label, 0);
- do_jump (TREE_OPERAND (exp, 1), if_false_label, if_true_label);
- break;
-
- case TRUTH_ORIF_EXPR:
- if (if_true_label == 0)
- if_true_label = drop_through_label = gen_label_rtx ();
- do_jump (TREE_OPERAND (exp, 0), 0, if_true_label);
- do_jump (TREE_OPERAND (exp, 1), if_false_label, if_true_label);
- break;
-
- case COMPOUND_EXPR:
- expand_expr (TREE_OPERAND (exp, 0), const0_rtx, VOIDmode, 0);
- free_temp_slots ();
- emit_queue ();
- do_jump (TREE_OPERAND (exp, 1), if_false_label, if_true_label);
- break;
-
- case COMPONENT_REF:
- case BIT_FIELD_REF:
- case ARRAY_REF:
- {
- int bitsize, bitpos, unsignedp;
- enum machine_mode mode;
- tree type;
- int volatilep = 0;
-
- /* Get description of this reference. We don't actually care
- about the underlying object here. */
- get_inner_reference (exp, &bitsize, &bitpos, &mode, &unsignedp,
- &volatilep);
-
- type = type_for_size (bitsize, unsignedp);
- if (type != 0
- && TYPE_PRECISION (type) < TYPE_PRECISION (TREE_TYPE (exp)))
- {
- do_jump (convert (type, exp), if_false_label, if_true_label);
- break;
- }
- goto normal;
- }
-
- case COND_EXPR:
- /* Do (a ? 1 : 0) and (a ? 0 : 1) as special cases. */
- if (integer_onep (TREE_OPERAND (exp, 1))
- && integer_zerop (TREE_OPERAND (exp, 2)))
- do_jump (TREE_OPERAND (exp, 0), if_false_label, if_true_label);
-
- else if (integer_zerop (TREE_OPERAND (exp, 1))
- && integer_onep (TREE_OPERAND (exp, 2)))
- do_jump (TREE_OPERAND (exp, 0), if_true_label, if_false_label);
-
- else
- {
- register rtx label1 = gen_label_rtx ();
- drop_through_label = gen_label_rtx ();
- do_jump (TREE_OPERAND (exp, 0), label1, 0);
- /* Now the THEN-expression. */
- do_jump (TREE_OPERAND (exp, 1),
- if_false_label ? if_false_label : drop_through_label,
- if_true_label ? if_true_label : drop_through_label);
- emit_label (label1);
- /* Now the ELSE-expression. */
- do_jump (TREE_OPERAND (exp, 2),
- if_false_label ? if_false_label : drop_through_label,
- if_true_label ? if_true_label : drop_through_label);
- }
- break;
-
- case EQ_EXPR:
- if (integer_zerop (TREE_OPERAND (exp, 1)))
- do_jump (TREE_OPERAND (exp, 0), if_true_label, if_false_label);
- else if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- &&
- !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_equality (exp, if_false_label, if_true_label);
- else
- comparison = compare (exp, EQ, EQ);
- break;
-
- case NE_EXPR:
- if (integer_zerop (TREE_OPERAND (exp, 1)))
- do_jump (TREE_OPERAND (exp, 0), if_false_label, if_true_label);
- else if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- &&
- !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_equality (exp, if_true_label, if_false_label);
- else
- comparison = compare (exp, NE, NE);
- break;
-
- case LT_EXPR:
- if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- && !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_greater (exp, 1, if_false_label, if_true_label);
- else
- comparison = compare (exp, LT, LTU);
- break;
-
- case LE_EXPR:
- if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- && !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_greater (exp, 0, if_true_label, if_false_label);
- else
- comparison = compare (exp, LE, LEU);
- break;
-
- case GT_EXPR:
- if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- && !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_greater (exp, 0, if_false_label, if_true_label);
- else
- comparison = compare (exp, GT, GTU);
- break;
-
- case GE_EXPR:
- if ((GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0))))
- == MODE_INT)
- && !can_compare_p (TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)))))
- do_jump_by_parts_greater (exp, 1, if_true_label, if_false_label);
- else
- comparison = compare (exp, GE, GEU);
- break;
-
- default:
- normal:
- temp = expand_expr (exp, 0, VOIDmode, 0);
- #if 0
- /* This is not needed any more and causes poor code since it causes
- comparisons and tests from non-SI objects to have different code
- sequences. */
- /* Copy to register to avoid generating bad insns by cse
- from (set (mem ...) (arithop)) (set (cc0) (mem ...)). */
- if (!cse_not_expected && GET_CODE (temp) == MEM)
- temp = copy_to_reg (temp);
- #endif
- do_pending_stack_adjust ();
- if (GET_CODE (temp) == CONST_INT)
- comparison = (temp == const0_rtx ? const0_rtx : const_true_rtx);
- else if (GET_MODE_CLASS (GET_MODE (temp)) == MODE_INT
- && !can_compare_p (GET_MODE (temp)))
- /* Note swapping the labels gives us not-equal. */
- do_jump_by_parts_equality_rtx (temp, if_true_label, if_false_label);
- else if (GET_MODE (temp) != VOIDmode)
- comparison = compare_from_rtx (temp, CONST0_RTX (GET_MODE (temp)),
- NE, 1, GET_MODE (temp), 0, 0);
- else
- abort ();
- }
-
- /* Do any postincrements in the expression that was tested. */
- emit_queue ();
-
- /* If COMPARISON is nonzero here, it is an rtx that can be substituted
- straight into a conditional jump instruction as the jump condition.
- Otherwise, all the work has been done already. */
-
- if (comparison == const_true_rtx)
- {
- if (if_true_label)
- emit_jump (if_true_label);
- }
- else if (comparison == const0_rtx)
- {
- if (if_false_label)
- emit_jump (if_false_label);
- }
- else if (comparison)
- do_jump_for_compare (comparison, if_false_label, if_true_label);
-
- free_temp_slots ();
-
- if (drop_through_label)
- emit_label (drop_through_label);
- }
-
- /* Given a comparison expression EXP for values too wide to be compared
- with one insn, test the comparison and jump to the appropriate label.
- The code of EXP is ignored; we always test GT if SWAP is 0,
- and LT if SWAP is 1. */
-
- static void
- do_jump_by_parts_greater (exp, swap, if_false_label, if_true_label)
- tree exp;
- int swap;
- rtx if_false_label, if_true_label;
- {
- rtx op0 = expand_expr (TREE_OPERAND (exp, swap), 0, VOIDmode, 0);
- rtx op1 = expand_expr (TREE_OPERAND (exp, !swap), 0, VOIDmode, 0);
- enum machine_mode mode = TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)));
- int nwords = (GET_MODE_SIZE (mode) / UNITS_PER_WORD);
- enum machine_mode submode
- = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- rtx drop_through_label = 0;
- int unsignedp = TREE_UNSIGNED (TREE_TYPE (TREE_OPERAND (exp, 0)));
- int i;
-
- if (! if_true_label || ! if_false_label)
- drop_through_label = gen_label_rtx ();
- if (! if_true_label)
- if_true_label = drop_through_label;
- if (! if_false_label)
- if_false_label = drop_through_label;
-
- /* Compare a word at a time, high order first. */
- for (i = 0; i < nwords; i++)
- {
- rtx comp;
- rtx op0_word, op1_word;
-
- if (WORDS_BIG_ENDIAN)
- {
- op0_word = operand_subword_force (op0, i, mode);
- op1_word = operand_subword_force (op1, i, mode);
- }
- else
- {
- op0_word = operand_subword_force (op0, nwords - 1 - i, mode);
- op1_word = operand_subword_force (op1, nwords - 1 - i, mode);
- }
-
- /* All but high-order word must be compared as unsigned. */
- comp = compare_from_rtx (op0_word, op1_word,
- (unsignedp || i > 0) ? GTU : GT,
- unsignedp, submode, 0, 0);
- do_jump_for_compare (comp, 0, if_true_label);
- /* Consider lower words only if these are equal. */
- comp = compare_from_rtx (op0_word, op1_word, NE, unsignedp, submode,
- 0, 0);
- do_jump_for_compare (comp, 0, if_false_label);
- }
-
- if (if_false_label)
- emit_jump (if_false_label);
- if (drop_through_label)
- emit_label (drop_through_label);
- }
-
- /* Given an EQ_EXPR expression EXP for values too wide to be compared
- with one insn, test the comparison and jump to the appropriate label. */
-
- static void
- do_jump_by_parts_equality (exp, if_false_label, if_true_label)
- tree exp;
- rtx if_false_label, if_true_label;
- {
- rtx op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode, 0);
- rtx op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- enum machine_mode mode = TYPE_MODE (TREE_TYPE (TREE_OPERAND (exp, 0)));
- int nwords = (GET_MODE_SIZE (mode) / UNITS_PER_WORD);
- enum machine_mode submode
- = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- int i;
- rtx drop_through_label = 0;
-
- if (! if_false_label)
- drop_through_label = if_false_label = gen_label_rtx ();
-
- for (i = 0; i < nwords; i++)
- {
- rtx comp = compare_from_rtx (operand_subword_force (op0, i, mode),
- operand_subword_force (op1, i, mode),
- EQ, 0, submode, 0, 0);
- do_jump_for_compare (comp, if_false_label, 0);
- }
-
- if (if_true_label)
- emit_jump (if_true_label);
- if (drop_through_label)
- emit_label (drop_through_label);
- }
-
- /* Jump according to whether OP0 is 0.
- We assume that OP0 has an integer mode that is too wide
- for the available compare insns. */
-
- static void
- do_jump_by_parts_equality_rtx (op0, if_false_label, if_true_label)
- rtx op0;
- rtx if_false_label, if_true_label;
- {
- int nwords = GET_MODE_SIZE (GET_MODE (op0)) / UNITS_PER_WORD;
- enum machine_mode submode
- = mode_for_size (BITS_PER_WORD, MODE_INT, 0);
- int i;
- rtx drop_through_label = 0;
-
- if (! if_false_label)
- drop_through_label = if_false_label = gen_label_rtx ();
-
- for (i = 0; i < nwords; i++)
- {
- rtx comp = compare_from_rtx (operand_subword_force (op0, i,
- GET_MODE (op0)),
- const0_rtx, EQ, 0, submode, 0, 0);
- do_jump_for_compare (comp, if_false_label, 0);
- }
-
- if (if_true_label)
- emit_jump (if_true_label);
- if (drop_through_label)
- emit_label (drop_through_label);
- }
-
- /* Given a comparison expression in rtl form, output conditional branches to
- IF_TRUE_LABEL, IF_FALSE_LABEL, or both. */
-
- static void
- do_jump_for_compare (comparison, if_false_label, if_true_label)
- rtx comparison, if_false_label, if_true_label;
- {
- if (if_true_label)
- {
- if (bcc_gen_fctn[(int) GET_CODE (comparison)] != 0)
- emit_jump_insn ((*bcc_gen_fctn[(int) GET_CODE (comparison)]) (if_true_label));
- else
- abort ();
-
- if (if_false_label)
- emit_jump (if_false_label);
- }
- else if (if_false_label)
- {
- rtx insn = get_last_insn ();
- rtx branch = 0;
-
- /* Output the branch with the opposite condition. Then try to invert
- what is generated. If more than one insn is a branch, or if the
- branch is not the last insn written, abort. If we can't invert
- the branch, emit make a true label, redirect this jump to that,
- emit a jump to the false label and define the true label. */
-
- if (bcc_gen_fctn[(int) GET_CODE (comparison)] != 0)
- emit_jump_insn ((*bcc_gen_fctn[(int) GET_CODE (comparison)]) (if_false_label));
- else
- abort ();
-
- for (insn = NEXT_INSN (insn); insn; insn = NEXT_INSN (insn))
- if (GET_CODE (insn) == JUMP_INSN)
- {
- if (branch)
- abort ();
- branch = insn;
- }
-
- if (branch != get_last_insn ())
- abort ();
-
- if (! invert_jump (branch, if_false_label))
- {
- if_true_label = gen_label_rtx ();
- redirect_jump (branch, if_true_label);
- emit_jump (if_false_label);
- emit_label (if_true_label);
- }
- }
- }
-
- /* Compare two integer constant rtx's, OP0 and OP1.
- The comparison operation is OPERATION.
- Return an rtx representing the value 1 or 0.
- WIDTH is the width in bits that is significant. */
-
- static rtx
- compare_constants (operation, unsignedp, op0, op1, width)
- enum rtx_code operation;
- int unsignedp;
- int op0, op1;
- int width;
- {
- int val;
-
- /* Sign-extend or zero-extend the operands to a full word
- from an initial width of WIDTH bits. */
- if (width < HOST_BITS_PER_INT)
- {
- op0 &= (1 << width) - 1;
- op1 &= (1 << width) - 1;
-
- if (! unsignedp)
- {
- if (op0 & (1 << (width - 1)))
- op0 |= ((-1) << width);
- if (op1 & (1 << (width - 1)))
- op1 |= ((-1) << width);
- }
- }
-
- switch (operation)
- {
- case EQ:
- val = op0 == op1;
- break;
-
- case NE:
- val = op0 != op1;
- break;
-
- case GT:
- val = op0 > op1;
- break;
-
- case GTU:
- val = (unsigned) op0 > (unsigned) op1;
- break;
-
- case LT:
- val = op0 < op1;
- break;
-
- case LTU:
- val = (unsigned) op0 < (unsigned) op1;
- break;
-
- case GE:
- val = op0 >= op1;
- break;
-
- case GEU:
- val = (unsigned) op0 >= (unsigned) op1;
- break;
-
- case LE:
- val = op0 <= op1;
- break;
-
- case LEU:
- val = (unsigned) op0 <= (unsigned) op1;
- break;
- }
-
- return val ? const_true_rtx : const0_rtx;
- }
-
- /* Generate code for a comparison expression EXP
- (including code to compute the values to be compared)
- and set (CC0) according to the result.
- SIGNED_CODE should be the rtx operation for this comparison for
- signed data; UNSIGNED_CODE, likewise for use if data is unsigned.
-
- We force a stack adjustment unless there are currently
- things pushed on the stack that aren't yet used. */
-
- static rtx
- compare (exp, signed_code, unsigned_code)
- register tree exp;
- enum rtx_code signed_code, unsigned_code;
- {
- register rtx op0 = expand_expr (TREE_OPERAND (exp, 0), 0, VOIDmode, 0);
- register rtx op1 = expand_expr (TREE_OPERAND (exp, 1), 0, VOIDmode, 0);
- register tree type = TREE_TYPE (TREE_OPERAND (exp, 0));
- register enum machine_mode mode = TYPE_MODE (type);
- int unsignedp = TREE_UNSIGNED (type);
- enum rtx_code code = unsignedp ? unsigned_code : signed_code;
-
- return compare_from_rtx (op0, op1, code, unsignedp, mode,
- ((mode == BLKmode)
- ? expr_size (TREE_OPERAND (exp, 0)) : 0),
- TYPE_ALIGN (TREE_TYPE (exp)) / BITS_PER_UNIT);
- }
-
- /* Like compare but expects the values to compare as two rtx's.
- The decision as to signed or unsigned comparison must be made by the caller.
-
- If MODE is BLKmode, SIZE is an RTX giving the size of the objects being
- compared.
-
- If ALIGN is non-zero, it is the alignment of this type; if zero, the
- size of MODE should be used. */
-
- rtx
- compare_from_rtx (op0, op1, code, unsignedp, mode, size, align)
- register rtx op0, op1;
- enum rtx_code code;
- int unsignedp;
- enum machine_mode mode;
- rtx size;
- int align;
- {
- /* If one operand is constant, make it the second one. */
-
- if (GET_CODE (op0) == CONST_INT || GET_CODE (op0) == CONST_DOUBLE)
- {
- rtx tem = op0;
- op0 = op1;
- op1 = tem;
- code = swap_condition (code);
- }
-
- if (flag_force_mem)
- {
- op0 = force_not_mem (op0);
- op1 = force_not_mem (op1);
- }
-
- do_pending_stack_adjust ();
-
- if (GET_CODE (op0) == CONST_INT && GET_CODE (op1) == CONST_INT)
- {
- if (mode == VOIDmode)
- abort ();
- return compare_constants (code, unsignedp, INTVAL (op0), INTVAL (op1),
- GET_MODE_BITSIZE (mode));
- }
-
- /* If this is a signed equality comparison, we can do it as an
- unsigned comparison since zero-extension is cheaper than sign
- extension and comparisons with zero are done as unsigned. If we
- are comparing against a constant, we must convert it to what it
- would look like unsigned. */
- if ((code == EQ || code == NE) && ! unsignedp)
- {
- if (GET_CODE (op1) == CONST_INT
- && (INTVAL (op1) & GET_MODE_MASK (GET_MODE (op0))) != INTVAL (op1))
- op1 = gen_rtx (CONST_INT, VOIDmode,
- INTVAL (op1) & GET_MODE_MASK (GET_MODE (op0)));
- unsignedp = 1;
- }
-
- emit_cmp_insn (op0, op1, code, size, unsignedp, align);
-
- return gen_rtx (code, VOIDmode, cc0_rtx, const0_rtx);
- }
-
- /* Generate code to calculate EXP using a store-flag instruction
- and return an rtx for the result.
- If TARGET is nonzero, store the result there if convenient.
-
- If ONLY_CHEAP is non-zero, only do this if it is likely to be very
- cheap.
-
- Return zero if there is no suitable set-flag instruction
- available on this machine.
-
- Once expand_expr has been called on the arguments of the comparison,
- we are committed to doing the store flag, since it is not safe to
- re-evaluate the expression. We emit the store-flag insn by calling
- emit_store_flag, but only expand the arguments if we have a reason
- to believe that emit_store_flag will be successful. If we think that
- it will, but it isn't, we have to simulate the store-flag with a
- set/jump/set sequence. */
-
- static rtx
- do_store_flag (exp, target, mode, only_cheap)
- tree exp;
- rtx target;
- enum machine_mode mode;
- int only_cheap;
- {
- enum rtx_code code;
- tree arg0 = TREE_OPERAND (exp, 0);
- tree arg1 = TREE_OPERAND (exp, 1);
- tree tem;
- tree type = TREE_TYPE (arg0);
- enum machine_mode operand_mode = TYPE_MODE (type);
- int unsignedp = TREE_UNSIGNED (type);
- rtx op0, op1;
- enum insn_code icode;
- rtx subtarget = target;
- rtx result, label, pattern, jump_pat;
-
- /* We won't bother with BLKmode store-flag operations because it would mean
- passing a lot of information to emit_store_flag. */
- if (operand_mode == BLKmode)
- return 0;
-
- while (TREE_CODE (arg0) == NON_LVALUE_EXPR)
- arg0 = TREE_OPERAND (arg0, 0);
-
- while (TREE_CODE (arg1) == NON_LVALUE_EXPR)
- arg1 = TREE_OPERAND (arg1, 0);
-
- /* Get the rtx comparison code to use. We know that EXP is a comparison
- operation of some type. */
- switch (TREE_CODE (exp))
- {
- case EQ_EXPR:
- code = EQ;
- break;
- case NE_EXPR:
- code = NE;
- break;
- case LT_EXPR:
- code = unsignedp ? LTU : LT;
- break;
- case LE_EXPR:
- code = unsignedp ? LEU : LE;
- break;
- case GT_EXPR:
- code = unsignedp ? GTU : GT;
- break;
- case GE_EXPR:
- code = unsignedp ? GEU : GE;
- break;
- default:
- abort ();
- }
-
- /* Put a constant second. */
- if (TREE_CODE (arg0) == REAL_CST || TREE_CODE (arg0) == INTEGER_CST)
- {
- tem = arg0; arg0 = arg1; arg1 = tem;
- code = swap_condition (code);
- }
-
- /* If this is an equality or inequality test of a single bit, we can
- do this by shifting the bit being tested to the low-order bit and
- masking the result with the constant 1. If the condition was EQ,
- we xor it with 1. This does not require an scc insn and is faster
- than an scc insn even if we have it. */
-
- if ((code == NE || code == EQ)
- && TREE_CODE (arg0) == BIT_AND_EXPR && integer_zerop (arg1)
- && integer_pow2p (TREE_OPERAND (arg0, 1))
- && TYPE_PRECISION (type) <= HOST_BITS_PER_INT)
- {
- int bitnum = exact_log2 (INTVAL (expand_expr (TREE_OPERAND (arg0, 1),
- 0, VOIDmode, 0)));
-
- if (subtarget == 0 || GET_CODE (subtarget) != REG
- || GET_MODE (subtarget) != operand_mode
- || ! safe_from_p (subtarget, TREE_OPERAND (arg0, 0)))
- subtarget = 0;
-
- op0 = expand_expr (TREE_OPERAND (arg0, 0), subtarget, VOIDmode, 0);
-
- if (bitnum != 0)
- op0 = expand_shift (RSHIFT_EXPR, GET_MODE (op0), op0,
- size_int (bitnum), target, 1);
-
- if (GET_MODE (op0) != mode)
- op0 = convert_to_mode (mode, op0, 1);
-
- if (bitnum != TYPE_PRECISION (type) - 1)
- op0 = expand_and (op0, const1_rtx, target);
-
- if (code == EQ)
- op0 = expand_binop (mode, xor_optab, op0, const1_rtx, target, 0,
- OPTAB_LIB_WIDEN);
-
- return op0;
- }
-
- /* Now see if we are likely to be able to do this. Return if not. */
- if (! can_compare_p (operand_mode))
- return 0;
- icode = setcc_gen_code[(int) code];
- if (icode == CODE_FOR_nothing
- || (only_cheap && insn_operand_mode[(int) icode][0] != mode))
- {
- /* We can only do this if it is one of the special cases that
- can be handled without an scc insn. */
- if ((code == LT && integer_zerop (arg1))
- || (! only_cheap && code == GE && integer_zerop (arg1)))
- ;
- else if (! only_cheap && (code == NE || code == EQ)
- && TREE_CODE (type) != REAL_TYPE
- && (abs_optab->handlers[(int) operand_mode].insn_code
- != CODE_FOR_nothing))
- ;
- else
- return 0;
- }
-
- preexpand_calls (exp);
- if (subtarget == 0 || GET_CODE (subtarget) != REG
- || GET_MODE (subtarget) != operand_mode
- || ! safe_from_p (subtarget, arg1))
- subtarget = 0;
-
- op0 = expand_expr (arg0, subtarget, VOIDmode, 0);
- op1 = expand_expr (arg1, 0, VOIDmode, 0);
-
- if (target == 0)
- target = gen_reg_rtx (mode);
-
- result = emit_store_flag (target, code, op0, op1, operand_mode, unsignedp, 1);
-
- if (result)
- return result;
-
- /* If this failed, we have to do this with set/compare/jump/set code. */
- if (target == 0 || GET_CODE (target) != REG
- || reg_mentioned_p (target, op0) || reg_mentioned_p (target, op1))
- target = gen_reg_rtx (GET_MODE (target));
-
- emit_move_insn (target, const1_rtx);
- result = compare_from_rtx (op0, op1, code, unsignedp, operand_mode, 0, 0);
- if (GET_CODE (result) == CONST_INT)
- return result == const0_rtx ? const0_rtx : const1_rtx;
-
- label = gen_label_rtx ();
- if (bcc_gen_fctn[(int) code] == 0)
- abort ();
-
- emit_jump_insn ((*bcc_gen_fctn[(int) code]) (label));
- emit_move_insn (target, const0_rtx);
- emit_label (label);
-
- return target;
- }
-
- /* Generate a tablejump instruction (used for switch statements). */
-
- #ifdef HAVE_tablejump
-
- /* INDEX is the value being switched on, with the lowest value
- in the table already subtracted.
- RANGE is the length of the jump table.
- TABLE_LABEL is a CODE_LABEL rtx for the table itself.
-
- DEFAULT_LABEL is a CODE_LABEL rtx to jump to if the
- index value is out of range. */
-
- void
- do_tablejump (index, range, table_label, default_label)
- rtx index, range, table_label, default_label;
- {
- register rtx temp, vector;
-
- emit_cmp_insn (range, index, LTU, 0, 0, 0);
- emit_jump_insn (gen_bltu (default_label));
- /* If flag_force_addr were to affect this address
- it could interfere with the tricky assumptions made
- about addresses that contain label-refs,
- which may be valid only very near the tablejump itself. */
- index = memory_address_noforce
- (CASE_VECTOR_MODE,
- gen_rtx (PLUS, Pmode,
- gen_rtx (MULT, Pmode, index,
- gen_rtx (CONST_INT, VOIDmode,
- GET_MODE_SIZE (CASE_VECTOR_MODE))),
- gen_rtx (LABEL_REF, Pmode, table_label)));
- temp = gen_reg_rtx (CASE_VECTOR_MODE);
- vector = gen_rtx (MEM, CASE_VECTOR_MODE, index);
- RTX_UNCHANGING_P (vector) = 1;
- convert_move (temp, vector, 0);
-
- emit_jump_insn (gen_tablejump (temp, table_label));
- emit_barrier ();
- }
-
- #endif /* HAVE_tablejump */
-