1 /* This file is part of the program psim.
3 Copyright (C) 1994-1995, Andrew Cagney <cagney@highland.com.au>
5 This program is free software; you can redistribute it and/or modify
6 it under the terms of the GNU General Public License as published by
7 the Free Software Foundation; either version 2 of the License, or
8 (at your option) any later version.
10 This program is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 GNU General Public License for more details.
15 You should have received a copy of the GNU General Public License
16 along with this program; if not, write to the Free Software
17 Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
26 /* bit manipulation routines:
28 Bit numbering: The bits are numbered according to the PowerPC
29 convention - the left most (or most significant) is bit 0 while the
30 right most (least significant) is bit 1.
32 Size convention: Each macro is in three forms - <MACRO>32 which
33 operates in 32bit quantity (bits are numbered 0..31); <MACRO>64
34 which operates using 64bit quantites (and bits are numbered 0..64);
35 and <MACRO> which operates using the bit size of the target
36 architecture (bits are still numbered 0..63), with 32bit
37 architectures ignoring the first 32bits having bit 32 as the most
40 BIT*(POS): Quantity with just 1 bit set.
42 MASK*(FIRST, LAST): Create a constant bit mask of the specified
43 size with bits [FIRST .. LAST] set.
45 MASKED*(VALUE, FIRST, LAST): Masks out all but bits [FIRST
48 LSMASKED*(VALUE, FIRST, LAST): Like MASKED - LS bit is zero.
50 EXTRACTED*(VALUE, FIRST, LAST): Masks out bits [FIRST .. LAST] but
51 also right shifts the masked value so that bit LAST becomes the
52 least significant (right most).
54 LSEXTRACTED*(VALUE, FIRST, LAST): Same as extracted - LS bit is
57 SHUFFLED**(VALUE, OLD, NEW): Mask then move a single bit from OLD
60 MOVED**(VALUE, OLD_FIRST, OLD_LAST, NEW_FIRST, NEW_LAST): Moves
61 things around so that bits OLD_FIRST..OLD_LAST are masked then
62 moved to NEW_FIRST..NEW_LAST.
64 INSERTED*(VALUE, FIRST, LAST): Takes VALUE and `inserts' the (LAST
65 - FIRST + 1) least significant bits into bit positions [ FIRST
66 .. LAST ]. This is almost the complement to EXTRACTED.
68 IEA_MASKED(SHOULD_MASK, ADDR): Convert the address to the targets
69 natural size. If in 32bit mode, discard the high 32bits.
71 EXTENDED(VALUE): Convert VALUE (32bits of it) to the targets
72 natural size. If in 64bit mode, sign extend the value.
74 ALIGN_*(VALUE): Round upwards the value so that it is aligned.
76 FLOOR_*(VALUE): Truncate the value so that it is aligned.
78 ROTL*(VALUE, NR_BITS): Return the value rotated by NR_BITS
82 #define _MAKE_SHIFT(WIDTH, pos) ((WIDTH) - 1 - (pos))
85 #if (WITH_TARGET_WORD_MSB == 0)
86 #define _LSB_POS(WIDTH, SHIFT) (WIDTH - 1 - SHIFT)
88 #define _LSB_POS(WIDTH, SHIFT) (SHIFT)
93 #define _BITn(WIDTH, pos) (((natural##WIDTH)(1)) \
94 << _MAKE_SHIFT(WIDTH, pos))
96 #define BIT4(POS) (1 << _MAKE_SHIFT(4, POS))
97 #define BIT5(POS) (1 << _MAKE_SHIFT(5, POS))
98 #define BIT8(POS) (1 << _MAKE_SHIFT(8, POS))
99 #define BIT10(POS) (1 << _MAKE_SHIFT(10, POS))
100 #define BIT32(POS) _BITn(32, POS)
101 #define BIT64(POS) _BITn(64, POS)
103 #if (WITH_TARGET_WORD_BITSIZE == 64)
104 #define BIT(POS) BIT64(POS)
106 #define BIT(POS) (((POS) < 32) ? 0 : _BITn(32, (POS)-32))
111 #define _MASKn(WIDTH, START, STOP) \
112 (((((unsigned##WIDTH)0) - 1) \
113 >> (WIDTH - ((STOP) - (START) + 1))) \
114 << (WIDTH - 1 - (STOP)))
116 #define MASK32(START, STOP) _MASKn(32, START, STOP)
117 #define MASK64(START, STOP) _MASKn(64, START, STOP)
119 /* Multi-bit mask on least significant bits */
121 #define _LSMASKn(WIDTH, FIRST, LAST) _MASKn (WIDTH, \
122 _LSB_POS (WIDTH, FIRST), \
123 _LSB_POS (WIDTH, LAST))
125 #define LSMASK64(FIRST, LAST) _LSMASKn (64, (FIRST), (LAST))
127 #if (WITH_TARGET_WORD_BITSIZE == 64)
128 #define MASK(START, STOP) \
129 (((START) <= (STOP)) \
130 ? _MASKn(64, START, STOP) \
131 : (_MASKn(64, 0, STOP) \
132 | _MASKn(64, START, 63)))
134 #define MASK(START, STOP) \
135 (((START) <= (STOP)) \
139 (START) < 32 ? 0 : (START) - 32, \
142 (START) < 32 ? 0 : (START) - 32, \
152 /* mask the required bits, leaving them in place */
155 (unsigned32
) MASKED32
161 (unsigned64
) MASKED64
167 (unsigned_word
) MASKED
173 (unsigned64
) LSMASKED64
179 /* extract the required bits aligning them with the lsb */
180 #define _EXTRACTEDn(WIDTH, WORD, START, STOP) \
181 ((((natural##WIDTH)(WORD)) >> (WIDTH - (STOP) - 1)) \
182 & _MASKn(WIDTH, WIDTH-1+(START)-(STOP), WIDTH-1))
184 /* #define EXTRACTED10(WORD, START, STOP) _EXTRACTEDn(10, WORD, START, STOP) */
185 #define EXTRACTED32(WORD, START, STOP) _EXTRACTEDn(32, WORD, START, STOP)
186 #define EXTRACTED64(WORD, START, STOP) _EXTRACTEDn(64, WORD, START, STOP)
189 (unsigned_word
) EXTRACTED
195 (unsigned64
) LSEXTRACTED64
200 /* move a single bit around */
201 /* NB: the wierdness (N>O?N-O:0) is to stop a warning from GCC */
202 #define _SHUFFLEDn(N, WORD, OLD, NEW) \
204 ? (((unsigned##N)(WORD) \
205 >> (((NEW) > (OLD)) ? ((NEW) - (OLD)) : 0)) \
206 & MASK32((NEW), (NEW))) \
207 : (((unsigned##N)(WORD) \
208 << (((OLD) > (NEW)) ? ((OLD) - (NEW)) : 0)) \
209 & MASK32((NEW), (NEW))))
211 #define SHUFFLED32(WORD, OLD, NEW) _SHUFFLEDn(32, WORD, OLD, NEW)
212 #define SHUFFLED64(WORD, OLD, NEW) _SHUFFLEDn(64, WORD, OLD, NEW)
214 #define SHUFFLED(WORD, OLD, NEW) _SHUFFLEDn(_word, WORD, OLD, NEW)
217 /* move a group of bits around */
218 #define _INSERTEDn(N, WORD, START, STOP) \
219 (((natural##N)(WORD) << _MAKE_SHIFT(N, STOP)) & _MASKn(N, START, STOP))
221 #define INSERTED32(WORD, START, STOP) _INSERTEDn(32, WORD, START, STOP)
222 #define INSERTED64(WORD, START, STOP) _INSERTEDn(64, WORD, START, STOP)
225 (unsigned_word
) INSERTED
231 /* depending on MODE return a 64bit or 32bit (sign extended) value */
232 #if (WITH_TARGET_WORD_BITSIZE == 64)
233 #define EXTENDED(X) ((signed64)(signed32)(X))
235 #define EXTENDED(X) (X)
239 /* memory alignment macro's */
240 #define _ALIGNa(A,X) (((X) + ((A) - 1)) & ~((A) - 1))
241 #define _FLOORa(A,X) ((X) & ~((A) - 1))
243 #define ALIGN_8(X) _ALIGNa(8, X)
244 #define ALIGN_16(X) _ALIGNa(16, X)
246 #define ALIGN_PAGE(X) _ALIGNa(0x1000, X)
247 #define FLOOR_PAGE(X) ((X) & ~(0x1000 - 1))
250 /* bit bliting macro's */
251 #define BLIT32(V, POS, BIT) \
258 #define MBLIT32(V, LO, HI, VAL) \
260 (V) = (((V) & ~MASK32((LO), (HI))) \
261 | INSERTED32(VAL, LO, HI)); \
265 /* some rotate functions to make things easier
267 NOTE: These are functions not macro's as the latter tickles bugs in
270 #define _ROTLn(N, VAL, SHIFT) \
271 (((VAL) << (SHIFT)) | ((VAL) >> ((N)-(SHIFT))))
284 #if (BITS_INLINE & INCLUDE_MODULE)
288 #endif /* _BITS_H_ */