5 // ; Peephole 1 removed pop %1 push %1 (not push pop)
13 // ; Peephole 2 removed pop %1 push %1 (not push pop)
18 // added by Jean Louis VERN for
24 ; Peephole 2.a removed redundant xch xch
32 ; Peephole 3.a changed mov to clr
42 ; Peephole 3.b changed mov to clr
48 // saving 1 byte, loosing 1 cycle but maybe allowing peephole 3.b to start
53 ; Peephole 3.c changed mov to clr
66 ; Peephole 100 removed redundant mov
79 ; Peephole 101 removed redundant mov
97 ; Peephole 102 removed redundant mov
119 ; Peephole 103 removed redundant mov
136 ; Peephole 104 optimized increment (acc not set to r%1, flags undefined)
145 ; Peephole 105 removed redundant mov
154 ; Peephole 106 removed redundant mov
163 ; Peephole 107 removed redundant ljmp
172 ; Peephole 108 removed ljmp by inverse jump logic
182 ; Peephole 109 removed ljmp by inverse jump logic
192 ; Peephole 110 removed ljmp by inverse jump logic
202 ; Peephole 111 removed ljmp by inverse jump logic
212 ; Peephole 112.a removed ljmp by inverse jump logic
221 ; Peephole 112.b changed ljmp to sjmp
235 ; Peephole 113 optimized misc sequence
240 } if labelRefCount %3 1
251 ; Peephole 114 optimized misc sequence
257 } if labelRefCount %3 2
266 ; Peephole 115 jump optimization
270 } if labelRefCount %3 1
280 ; Peephole 116 jump optimization
285 } if labelRefCount %3 2
296 ; Peephole 117 jump optimization
302 } if labelRefCount %3 3
314 ; Peephole 118 jump optimization
321 } if labelRefCount %3 4
330 ; Peephole 119 jump optimization
333 } if labelRefCount %3 1
343 ; Peephole 120 jump optimization
347 } if labelRefCount %3 2
358 ; Peephole 121 jump optimization
363 } if labelRefCount %3 3
375 ; Peephole 122 jump optimization
381 } if labelRefCount %3 4
390 ; Peephole 123 jump optimization
394 } if labelRefCount %3 1
404 ; Peephole 124 jump optimization
409 } if labelRefCount %3 2
420 ; Peephole 125 jump optimization
426 } if labelRefCount %3 3
438 ; Peephole 126 jump optimization
445 } if labelRefCount %3 4
458 ; Peephole 127 removed misc sequence
460 } if labelRefCount %3 0
467 ; Peephole 128 jump optimization
476 ; Peephole 129 jump optimization
483 ; Peephole 130 changed target address mode r%1 to ar%1
493 ; Peephole 131 optimized decrement (not caring for c)
505 ; Peephole 133 removed redundant moves
518 ; Peephole 134 removed redundant moves
529 ; Peephole 135 removed redundant mov
540 ; Peephole 136 removed redundant moves
546 // WTF? Doesn't look sensible to me...
556 // ; Peephole 137 optimized misc jump sequence
560 //} if labelRefCount %4 1
571 // ; Peephole 138 optimized misc jump sequence
576 //} if labelRefCount %4 1
583 ; Peephole 139 removed redundant mov
593 ; Peephole 140 removed redundant mov
602 ; Peephole 141 removed redundant mov
612 ; Peephole 142 removed redundant moves
621 ; Peephole 143 converted rlc to rl
629 ; Peephole 144 converted rrc to rc
637 ; Peephole 145 changed to add without carry
646 ; Peephole 146 changed to add without carry
654 ; Peephole 147 changed target address mode r%1 to ar%1
661 ; Peephole 148 changed target address mode r%1 to ar%1
668 ; Peephole 149 changed target address mode r%1 to ar%1
678 ; Peephole 150 removed misc moves via dpl before return
691 ; Peephole 151 removed misc moves via dph, dpl before return
703 ; Peephole 152 removed misc moves via dph, dpl before return
718 ; Peephole 153 removed misc moves via dph, dpl, b before return
731 ; Peephole 154 removed misc moves via dph, dpl, b before return
745 ; Peephole 155 removed misc moves via dph, dpl, b before return
762 ; Peephole 156 removed misc moves via dph, dpl, b, a before return
777 ; Peephole 157 removed misc moves via dph, dpl, b, a before return
791 ; Peephole 158 removed misc moves via dph, dpl, b, a before return
800 ; Peephole 159 avoided xrl during execution
809 ; Peephole 160 removed sjmp by inverse jump logic
819 ; Peephole 161 removed sjmp by inverse jump logic
829 ; Peephole 162 removed sjmp by inverse jump logic
839 ; Peephole 163 removed sjmp by inverse jump logic
849 ; Peephole 164 removed sjmp by inverse jump logic
859 ; Peephole 165 removed sjmp by inverse jump logic
869 ; Peephole 166 removed redundant mov
872 } if notVolatile %1 %2
879 ; Peephole 167 removed redundant bit moves (c not set to %1)
888 ; Peephole 168 jump optimization
898 ; Peephole 169 jump optimization
910 ; Peephole 170 jump optimization
913 } if labelRefCount %3 1
923 ; Peephole 171 jump optimization
927 } if labelRefCount %3 2
938 ; Peephole 172 jump optimization
943 } if labelRefCount %3 3
955 ; Peephole 173 jump optimization
961 } if labelRefCount %3 4
970 ; Peephole 174 optimized decrement (acc not set to %2, flags undefined)
981 ; Peephole 175 optimized increment (acc not set to %2, flags undefined)
991 ; Peephole 176 optimized increment, removed redundant mov
996 // this one will screw assignes to volatile/sfr's
1001 ; Peephole 177.a removed redundant mov
1003 } if notVolatile %1 %2
1005 // applies to f.e. scott-add.asm (--model-large)
1010 ; Peephole 177.b removed redundant mov
1014 // applies to f.e. bug-408972.c
1019 ; Peephole 177.c removed redundant move
1021 } if notVolatile %1 %2
1023 // applies to f.e. bug-408972.c
1024 // not before peephole 177.c
1030 ; Peephole 177.d removed redundant move
1033 } if notVolatile %1 %2
1035 // applies to f.e. bug-607243.c
1036 // also check notVolatile %3, as it will return FALSE if it's @r%1
1041 ; peephole 177.e removed redundant move
1043 } if notVolatile %2 %3
1049 ; peephole 177.f removed redundant move
1058 ; Peephole 178 removed redundant mov
1063 // rules 179-182 provided by : Frieder <fe@lake.iup.uni-heidelberg.de>
1064 // saving 2 byte, 1 cycle
1069 ; Peephole 179 changed mov to clr
1075 // volatile xdata char t; t=0x01; t=0x03;
1082 ; Peephole 180.a removed redundant mov to dptr
1088 // volatile xdata char t; t=0x01; t=0x03; t=0x01;
1097 ; Peephole 180.b removed redundant mov to dptr
1105 // saving 1 byte, 0 cycles
1109 ; Peephole 181 changed mov to clr
1113 // saving 3 bytes, 2 cycles
1114 // provided by Bernhard Held <bernhard.held@de.westinghouse.com>
1119 ; Peephole 182.a used 16 bit load of DPTR
1123 // saving 3 byte, 2 cycles, return(NULL) profits here
1128 ; Peephole 182.b used 16 bit load of dptr
1132 // saving 3 byte, 2 cycles. Probably obsoleted by 182.b
1137 ; Peephole 182.c used 16 bit load of dptr
1138 mov dptr,#(((%2)<<8) + %1)
1141 // applies to return 0.0; in f.e. sincosf.c
1147 ; Peephole 182.d used 16 bit load of dptr
1148 mov dptr,#(%1&0x00ff)
1156 ; Peephole 183 avoided anl during execution
1165 ; Peephole 184 removed redundant mov
1171 // acc being incremented might cause problems
1175 ; Peephole 185 changed order of increment (acc incremented also!)
1200 ; Peephole 186.a optimized movc sequence
1233 ; Peephole 186.b optimized movc sequence
1258 ; Peephole 186.c optimized movc sequence
1267 // char indexed access to: char code table[] = {4,3,2,1};
1277 ; Peephole 186.d optimized movc sequence
1282 // char indexed access to: int code table[] = {4,3,2,1};
1297 ; Peephole 186.e optimized movc sequence (b, dptr differ)
1315 ; Peephole 187 used a instead of ar%1 for anl
1327 ; Peephole 188 removed redundant mov
1339 ; Peephole 189 removed redundant mov and anl
1344 // rules 190 & 191 need to be in order
1350 ; Peephole 190 removed redundant mov
1362 ; Peephole 191 removed redundant mov
1373 ; Peephole 192 used a instead of ar%1 as source
1390 ; Peephole 193.a optimized misc jump sequence
1400 } if labelRefCount %3 4
1414 ; Peephole 193.b optimized misc jump sequence
1424 } if labelRefCount %3 4
1438 ; Peephole 193.c optimized misc jump sequence
1448 } if labelRefCount %3 4
1459 ; Peephole 194 optimized misc jump sequence
1466 } if labelRefCount %3 4
1478 ; Peephole 195.a optimized misc jump sequence
1486 } if labelRefCount %3 3
1498 ; Peephole 195.b optimized misc jump sequence
1506 } if labelRefCount %3 3
1518 ; Peephole 195.c optimized misc jump sequence
1526 } if labelRefCount %3 3
1536 ; Peephole 196 optimized misc jump sequence
1542 } if labelRefCount %3 3
1552 ; Peephole 197.a optimized misc jump sequence
1558 } if labelRefCount %3 2
1568 ; Peephole 197.b optimized misc jump sequence
1574 } if labelRefCount %3 2
1584 ; Peephole 197.c optimized misc jump sequence
1590 } if labelRefCount %3 2
1599 ; Peephole 198 optimized misc jump sequence
1604 } if labelRefCount %3 2
1612 ; Peephole 199 optimized misc jump sequence
1616 } if labelRefCount %3 1
1622 ; Peephole 200 removed redundant sjmp
1631 ; Peephole 201 removed redundant sjmp
1641 ; Peephole 202 removed redundant push pop
1650 ; Peephole 203 removed mov r%1,_spx
1659 ; Peephole 204 removed redundant mov
1671 ; Peephole 205 optimized misc jump sequence
1675 } if labelRefCount %2 1
1680 ; Peephole 206 removed redundant mov %1,%1
1688 ; Peephole 207 removed zero add (acc not set to %1, flags undefined)
1697 ; Peephole 208 removed redundant push pop
1707 ; Peephole 209 optimized increment (acc not set to %1, flags undefined)
1713 mov dptr,#((((%1 >> 8)) <<8) + %1)
1715 ; Peephole 210 simplified expression
1723 ; Peephole 211 removed redundant push %1 pop %1
1731 ; Peephole 212 reduced add sequence to inc
1737 mov %1,#(( %2 >> 8 ) ^ 0x80)
1739 ; Peephole 213.a inserted fix
1745 mov %1,#(( %2 + %3 >> 8 ) ^ 0x80)
1747 ; Peephole 213.b inserted fix
1748 mov %1,#((%2 + %3) >> 8)
1758 ; Peephole 214 reduced some extra moves
1761 } if operandsNotSame
1768 ; Peephole 215 removed some moves
1771 } if operandsNotSame
1781 ; Peephole 216 simplified clear (2bytes)
1800 ; Peephole 217 simplified clear (3bytes)
1824 ; Peephole 218 simplified clear (4bytes)
1843 ; Peephole 219 removed redundant clear
1859 ; Peephole 219.a removed redundant clear
1872 ; Peephole 220.a removed bogus DPS set
1880 ; Peephole 220.b removed bogus DPS set
1885 mov %1 + %2,(%2 + %1)
1887 ; Peephole 221.a remove redundant move
1891 mov (%1 + %2 + %3),((%2 + %1) + %3)
1893 ; Peephole 221.b remove redundant move
1900 ; Peephole 222 removed dec/inc pair
1909 ; Peephole 223 removed redundant dph/dpl moves
1912 } if notVolatile %1 %2
1920 ; Peephole 224 removed redundant dph/dpl moves
1933 ; Peephole 225 removed redundant move to acc
1949 ; Peephole 226 removed unnecessary clr
1971 ; Peephole 227 replaced inefficient 32 bit clear
1999 ; Peephole 228 replaced inefficient 32 constant
2021 ; Peephole 229 replaced inefficient 16 bit clear
2039 ; Peephole 230 replaced inefficient 16 bit constant
2049 // this last peephole often removes the last mov from 227-230
2054 ; Peephole 231 removed redundant mov to dptr
2061 ; Peephole 232 using movc to read xdata (--xram-movc)
2069 ; Peephole 233 using _gptrgetc instead of _gptrget (--xram-movc)
2079 ; Peephole 234 loading dpl directly from a(ccumulator), r%1 not set
2092 ; Peephole 235 loading dph directly from a(ccumulator), r%1 not set
2099 // 14 rules by Fiorenzo D. Ramaglia <fd.ramaglia@tin.it>
2104 ; Peephole 236.a used r%1 instead of ar%1
2111 ; Peephole 236.b used r%1 instead of ar%1
2118 ; Peephole 236.c used r%1 instead of ar%1
2125 ; Peephole 236.d used r%1 instead of ar%1
2132 ; Peephole 236.e used r%1 instead of ar%1
2139 ; Peephole 236.f used r%1 instead of ar%1
2146 ; Peephole 236.g used r%1 instead of ar%1
2153 ; Peephole 236.h used r%1 instead of ar%1
2160 ; Peephole 236.i used r%1 instead of ar%1
2167 ; Peephole 236.j used r%1 instead of ar%1
2174 ; Peephole 236.k used r%1 instead of ar%1
2181 ; Peephole 236.l used r%1 instead of ar%1
2188 ; Peephole 236.m used r%1 instead of ar%1
2195 ; Peephole 236.n used r%1 instead of ar%1
2206 ; Peephole 237.a removed sjmp to ret
2223 ; Peephole 237.b removed sjmp to ret
2233 // applies to f.e. device/lib/log10f.c
2259 ; Peephole 238.a removed 4 redundant moves
2260 } if operandsNotSame8 %1 %2 %3 %4 %5 %6 %7 %8
2262 // applies to device/lib/log10f.c
2277 ; Peephole 238.b removed 3 redundant moves
2278 } if operandsNotSame7 %1 %2 %3 %4 %5 %6 %7
2280 // applies to f.e. device/lib/time.c
2296 ; Peephole 238.c removed 2 redundant moves
2297 } if operandsNotSame4 %1 %2 %3 %4
2299 // applies to f.e. support/regression/tests/bug-524209.c
2312 ; Peephole 238.d removed 3 redundant moves
2313 } if operandsNotSame6 %1 %2 %3 %4 %5 %6
2315 // applies to f.e. ser_ir.asm
2319 ; Peephole 239 used a instead of acc
2327 ; Peephole 240 use clr instead of addc a,#0
2332 // peepholes 241.a to 241.c and 241.d to 241.f need to be in order
2344 ; Peephole 241.a optimized compare
2355 // applies to f.e. time.c
2365 ; Peephole 241.b optimized compare
2374 // applies to f.e. malloc.c
2383 ; Peephole 241.c optimized compare
2391 // applies to f.e. j = (k!=0x1000);
2392 // with volatile idata long k;
2407 ; Peephole 241.d optimized compare
2421 // applies to f.e. j = (k!=0x1000);
2422 // with volatile idata int k;
2433 ; Peephole 241.e optimized compare
2443 // applies to f.e. vprintf.asm (--stack-auto)
2452 ; Peephole 241.f optimized compare
2460 // applies to f.e. scott-bool1.c
2467 ; Peephole 242.a avoided branch jnz to jz
2472 } if labelRefCount %1 1
2474 // applies to f.e. scott-bool1.c
2482 ; Peephole 242.b avoided branch jnz to jz
2488 } if labelRefCount %1 1
2490 // applies to f.e. logic.c
2500 ; Peephole 242.c avoided branch jnz to jz
2508 } if labelRefCount %1 1
2510 // applies to f.e. vprintf.c
2511 // this is a rare case, usually the "tail increment" is noticed earlier
2518 ; Peephole 243 avoided branch to sjmp
2525 // applies to f.e. simplefloat.c (saving 1 cycle)
2530 ; Peephole 244.a moving first to a instead of r%1
2535 // applies to f.e. _itoa.c (saving 1 cycle)
2540 ; Peephole 244.b moving first to a instead of r%1
2546 // applies to f.e. bug-460010.c (saving 1 cycle)
2551 ; Peephole 244.c loading dpl from a instead of r%1
2560 ; Peephole 244.d loading dph from a instead of r%1
2565 // this one is safe but disables 245.a 245.b
2566 // please remove 245 if 245.a 245.b are found to be safe
2567 // applies to f.e. scott-compare.c
2578 ; Peephole 245 optimized complement (r%1 and acc set needed?)
2583 } if labelRefCount %2 1
2585 // this one will not be triggered if 245 is present
2586 // please remove 245 if 245.a 245.b are found to be safe
2587 // applies to f.e. vprintf.c
2599 ; Peephole 245.a optimized conditional jump (r%1 and acc not set!)
2601 } if labelRefCount %2 1
2603 // this one will not be triggered if 245 is present
2604 // please remove 245 if 245.a 245.b are found to be safe
2605 // applies to f.e. scott-compare.c
2617 ; Peephole 245.b optimized conditional jump (r%1 and acc not set!)
2619 } if labelRefCount %2 1
2622 // rules 246.x apply to f.e. bitfields.c
2633 ; Peephole 246.a combined clr/clr
2650 ; Peephole 246.b combined set/set
2667 ; Peephole 246.c combined set/clr
2685 ; Peephole 246.d combined clr/set
2704 ; Peephole 246.e combined set/clr/clr
2723 ; Peephole 246.f combined set/clr/set
2743 ; Peephole 246.g combined clr/set/clr
2763 ; Peephole 246.h combined clr/set/set
2774 // rules 247.x apply to f.e. bitfields.c
2785 ; Peephole 247.a combined clr/clr
2802 ; Peephole 247.b combined set/set
2819 ; Peephole 247.c combined set/clr
2837 ; Peephole 247.d combined clr/set
2856 ; Peephole 247.e combined set/clr/clr
2875 ; Peephole 247.f combined set/clr/set
2895 ; Peephole 247.g combined clr/set/clr
2915 ; Peephole 247.h combined clr/set/set
2924 // Peepholes 248.x have to be compatible with the keyword volatile.
2925 // They optimize typical accesses to memory mapped I/O devices:
2926 // volatile xdata char t; t|=0x01;
2936 ; Peephole 248.a optimized or to xdata
2944 // volatile xdata char t; t&=0x01;
2954 ; Peephole 248.b optimized and to xdata
2962 // volatile xdata char t; t^=0x01;
2972 ; Peephole 248.c optimized xor to xdata
2980 // volatile xdata char t; t|=0x01; t&=~0x01; t|=0x01;
3000 ; Peephole 248.d optimized or/and/or to volatile xdata
3014 // volatile xdata char t; t&=~0x01; t|=0x01; t&=~0x01;
3034 ; Peephole 248.e optimized and/or/and to volatile xdata
3048 // volatile xdata char t; t|=0x01; t&=~0x01;
3062 ; Peephole 248.f optimized or/and to volatile xdata
3073 // volatile xdata char t; t&=~0x01; t|=0x01;
3087 ; Peephole 248.g optimized and/or to volatile xdata
3098 // volatile xdata char t; t^=0x01; t^=0x01;
3112 ; Peephole 248.h optimized xor/xor to volatile xdata
3123 // Peeepholes 248.i to 248.m are like 248.d to 248.h except they apply to bitfields:
3124 // xdata struct { unsigned b0:1; unsigned b1:1; unsigned b2:1; } xport;
3125 // xport.b0=1; xport.b0=0; xport.b0=1;
3142 ; Peephole 248.i optimized or/and/or to xdata bitfield
3171 ; Peephole 248.j optimized and/or/and to xdata bitfield
3195 ; Peephole 248.k optimized or/and to xdata bitfield
3216 ; Peephole 248.l optimized and/or to xdata bitfield
3237 ; Peephole 248.m optimized xor/xor to xdata bitfield
3252 ; Peephole 249.a jump optimization
3253 } if labelRefCount %1 1
3259 ; Peephole 249.b jump optimization
3260 } if labelRefCount %1 1
3263 // This allows non-interrupt and interrupt code to safely compete
3264 // for a resource without the non-interrupt code having to disable
3266 // volatile bit resource_is_free;
3267 // if( resource_is_free ) {
3268 // resource_is_free=0; do_something; resource_is_free=1;
3275 ; Peephole 250.a using atomic test and clear
3279 } if labelRefCount %3 0
3287 ; Peephole 250.b using atomic test and clear
3291 } if labelRefCount %2 1
3294 // not before peephole 250.b
3298 ; Peephole 251.a replaced ljmp to ret with ret
3300 } if labelIsReturnOnly
3302 // not before peephole 250.b
3306 ; Peephole 251.b replaced sjmp to ret with ret
3308 } if labelIsReturnOnly
3310 // applies to shifts.c and when accessing arrays with an unsigned integer index
3311 // saving 1 byte, 2 cycles
3321 ; Peephole 252 optimized left shift
3330 // applies to: void test( char c ) { if( c ) func1(); else func2(); }
3335 ; Peephole 253.a replaced lcall/ret with ljmp
3339 // applies to: void test( char c ) { if( c ) func1(); else func2(); }
3345 ; Peephole 253.b replaced lcall/ret with ljmp
3347 } if labelRefCount %2 0
3349 // applies to f.e. scott-bool1.c
3355 ; Peephole 253.c replaced lcall with ljmp
3362 // applies to f.e. funptrs.c
3363 // saves one byte if %1 is a register or @register
3368 ; Peephole 254 optimized left shift
3373 // applies to f.e. switch.c
3386 ; Peephole 255 optimized jump table index calculation
3388 cjne a,#(%1+0x01),.+1
3397 // applies to f.e. jump tables and scott-bool1.c.
3398 // similar peepholes can be constructed for other instructions
3399 // after which a flag or a register is known (like: djnz, cjne, jnc)
3405 ; Peephole 256.a removed redundant clr c
3408 } if labelRefCount %2 0
3410 // applies to f.e. logf.c
3416 ; Peephole 256.b removed redundant clr a
3419 } if labelRefCount %2 0
3422 // unsigned char i=8; do{ } while(--i != 0);
3423 // this currently only applies if i is kept in a register
3428 ; Peephole 257 optimized decrement with compare
3433 // in_byte<<=1; if(in_bit) in_byte|=1;
3434 // helps f.e. reading data on a 3-wire (SPI) bus
3444 ; Peephole 258.a optimized bitbanging
3453 // in_byte<<=1; if(in_bit) in_byte|=1;
3463 ; Peephole 258.b optimized bitbanging
3472 // in_byte>>=1; if(in_bit) in_byte|=0x80;
3483 ; Peephole 258.c optimized bitbanging
3492 // in_byte>>=1; if(in_bit) in_byte|=0x80;
3503 ; Peephole 258.d optimized bitbanging
3512 // out_bit=out_byte&0x80; out_byte<<=1;
3513 // helps f.e. writing data on a 3-wire (SPI) bus
3522 ; Peephole 258.e optimized bitbanging
3529 // out_bit=out_byte&0x01; out_byte>>=1;
3539 ; Peephole 258.f optimized bitbanging