#ifndef __INC_CLOCKLESS_TRINKET_H #define __INC_CLOCKLESS_TRINKET_H #include "../../controller.h" #include "../../lib8tion.h" #include // for cli/se definitions FASTLED_NAMESPACE_BEGIN #if defined(FASTLED_AVR) // Scaling macro choice #ifndef TRINKET_SCALE #define TRINKET_SCALE 1 // whether or not to use dithering #define DITHER 1 #endif #if (F_CPU==8000000) #define FASTLED_SLOW_CLOCK_ADJUST // asm __volatile__ ("mov r0,r0\n\t"); #else #define FASTLED_SLOW_CLOCK_ADJUST #endif #define US_PER_TICK (64 / (F_CPU/1000000)) // Variations on the functions in delay.h - w/a loop var passed in to preserve registers across calls by the optimizer/compiler template inline void _dc(register uint8_t & loopvar); template __attribute__((always_inline)) inline void _dc_AVR(register uint8_t & loopvar) { _dc(loopvar); // The convolution in here is to ensure that the state of the carry flag coming into the delay loop is preserved asm __volatile__ ( "BRCS L_PC%=\n\t" " LDI %[loopvar], %[_LOOP]\n\tL_%=: DEC %[loopvar]\n\t BRNE L_%=\n\tBREQ L_DONE%=\n\t" "L_PC%=: LDI %[loopvar], %[_LOOP]\n\tLL_%=: DEC %[loopvar]\n\t BRNE LL_%=\n\tBSET 0\n\t" "L_DONE%=:\n\t" : [loopvar] "+a" (loopvar) : [_LOOP] "M" (_LOOP) : ); } template __attribute__((always_inline)) inline void _dc(register uint8_t & loopvar) { _dc_AVR(loopvar); } template<> __attribute__((always_inline)) inline void _dc<-6>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc<-5>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc<-4>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc<-3>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc<-2>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc<-1>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc< 0>(register uint8_t & ) {} template<> __attribute__((always_inline)) inline void _dc< 1>(register uint8_t & ) {asm __volatile__("mov r0,r0":::);} template<> __attribute__((always_inline)) inline void _dc< 2>(register uint8_t & ) {asm __volatile__("rjmp .+0":::);} template<> __attribute__((always_inline)) inline void _dc< 3>(register uint8_t & loopvar) { _dc<2>(loopvar); _dc<1>(loopvar); } template<> __attribute__((always_inline)) inline void _dc< 4>(register uint8_t & loopvar) { _dc<2>(loopvar); _dc<2>(loopvar); } template<> __attribute__((always_inline)) inline void _dc< 5>(register uint8_t & loopvar) { _dc<2>(loopvar); _dc<3>(loopvar); } template<> __attribute__((always_inline)) inline void _dc< 6>(register uint8_t & loopvar) { _dc<2>(loopvar); _dc<2>(loopvar); _dc<2>(loopvar);} template<> __attribute__((always_inline)) inline void _dc< 7>(register uint8_t & loopvar) { _dc<4>(loopvar); _dc<3>(loopvar); } template<> __attribute__((always_inline)) inline void _dc< 8>(register uint8_t & loopvar) { _dc<4>(loopvar); _dc<4>(loopvar); } template<> __attribute__((always_inline)) inline void _dc< 9>(register uint8_t & loopvar) { _dc<5>(loopvar); _dc<4>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<10>(register uint8_t & loopvar) { _dc<6>(loopvar); _dc<4>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<11>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<1>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<12>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<2>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<13>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<3>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<14>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<4>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<15>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<5>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<16>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<6>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<17>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<7>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<18>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<8>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<19>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<9>(loopvar); } template<> __attribute__((always_inline)) inline void _dc<20>(register uint8_t & loopvar) { _dc<10>(loopvar); _dc<10>(loopvar); } #define DINTPIN(T,ADJ,PINADJ) (T-(PINADJ+ADJ)>0) ? _dc(loopvar) : _dc<0>(loopvar); #define DINT(T,ADJ) if(AVR_PIN_CYCLES(DATA_PIN)==1) { DINTPIN(T,ADJ,1) } else { DINTPIN(T,ADJ,2); } #define _D1(ADJ) DINT(T1,ADJ) #define _D2(ADJ) DINT(T2,ADJ) #define _D3(ADJ) DINT(T3,ADJ) ////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////// // // Base template for clockless controllers. These controllers have 3 control points in their cycle for each bit. The first point // is where the line is raised hi. The second point is where the line is dropped low for a zero. The third point is where the // line is dropped low for a one. T1, T2, and T3 correspond to the timings for those three in clock cycles. // ////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////// #if (!defined(NO_CORRECTION) || (NO_CORRECTION == 0)) && (FASTLED_ALLOW_INTERRUPTS == 0) static uint8_t gTimeErrorAccum256ths; #endif #define FASTLED_HAS_CLOCKLESS 1 template class ClocklessController : public CPixelLEDController { static_assert(T1 >= 2 && T2 >= 2 && T3 >= 3, "Not enough cycles - use a higher clock speed"); typedef typename FastPin::port_ptr_t data_ptr_t; typedef typename FastPin::port_t data_t; CMinWait mWait; public: virtual void init() { FastPin::setOutput(); } virtual uint16_t getMaxRefreshRate() const { return 400; } protected: virtual void showPixels(PixelController & pixels) { mWait.wait(); cli(); showRGBInternal(pixels); // Adjust the timer #if (!defined(NO_CORRECTION) || (NO_CORRECTION == 0)) && (FASTLED_ALLOW_INTERRUPTS == 0) uint32_t microsTaken = (uint32_t)pixels.size() * (uint32_t)CLKS_TO_MICROS(24 * (T1 + T2 + T3)); // adust for approximate observed actal runtime (as of January 2015) // roughly 9.6 cycles per pixel, which is 0.6us/pixel at 16MHz // microsTaken += nLeds * 0.6 * CLKS_TO_MICROS(16); microsTaken += scale16by8(pixels.size(),(0.6 * 256) + 1) * CLKS_TO_MICROS(16); // if less than 1000us, there is NO timer impact, // this is because the ONE interrupt that might come in while interrupts // are disabled is queued up, and it will be serviced as soon as // interrupts are re-enabled. // This actually should technically also account for the runtime of the // interrupt handler itself, but we're just not going to worry about that. if( microsTaken > 1000) { // Since up to one timer tick will be queued, we don't need // to adjust the MS_COUNTER for that one. microsTaken -= 1000; // Now convert microseconds to 256ths of a second, approximately like this: // 250ths = (us/4) // 256ths = 250ths * (263/256); uint16_t x256ths = microsTaken >> 2; x256ths += scale16by8(x256ths,7); x256ths += gTimeErrorAccum256ths; MS_COUNTER += (x256ths >> 8); gTimeErrorAccum256ths = x256ths & 0xFF; } #if 0 // For pixel counts of 30 and under at 16Mhz, no correction is necessary. // For pixel counts of 15 and under at 8Mhz, no correction is necessary. // // This code, below, is smaller, and quicker clock correction, which drifts much // more significantly, but is a few bytes smaller. Presented here for consideration // as an alternate on the ATtiny, which can't have more than about 150 pixels MAX // anyway, meaning that microsTaken will never be more than about 4,500, which fits in // a 16-bit variable. The difference between /1000 and /1024 only starts showing // up in the range of about 100 pixels, so many ATtiny projects won't even // see a clock difference due to the approximation there. uint16_t microsTaken = (uint32_t)nLeds * (uint32_t)CLKS_TO_MICROS((24) * (T1 + T2 + T3)); MS_COUNTER += (microsTaken >> 10); #endif #endif sei(); mWait.mark(); } #define USE_ASM_MACROS // The variables that our various asm statemetns use. The same block of variables needs to be declared for // all the asm blocks because GCC is pretty stupid and it would clobber variables happily or optimize code away too aggressively #define ASM_VARS : /* write variables */ \ [count] "+x" (count), \ [data] "+z" (data), \ [b1] "+a" (b1), \ [d0] "+r" (d0), \ [d1] "+r" (d1), \ [d2] "+r" (d2), \ [loopvar] "+a" (loopvar), \ [scale_base] "+a" (scale_base) \ : /* use variables */ \ [ADV] "r" (advanceBy), \ [b0] "a" (b0), \ [hi] "r" (hi), \ [lo] "r" (lo), \ [s0] "r" (s0), \ [s1] "r" (s1), \ [s2] "r" (s2), \ [e0] "r" (e0), \ [e1] "r" (e1), \ [e2] "r" (e2), \ [PORT] "M" (FastPin::port()-0x20), \ [O0] "M" (RGB_BYTE0(RGB_ORDER)), \ [O1] "M" (RGB_BYTE1(RGB_ORDER)), \ [O2] "M" (RGB_BYTE2(RGB_ORDER)) \ : "cc" /* clobber registers */ // Note: the code in the else in HI1/LO1 will be turned into an sts (2 cycle, 2 word) opcode // 1 cycle, write hi to the port #define HI1 FASTLED_SLOW_CLOCK_ADJUST if((int)(FastPin::port())-0x20 < 64) { asm __volatile__("out %[PORT], %[hi]" ASM_VARS ); } else { *FastPin::port()=hi; } // 1 cycle, write lo to the port #define LO1 if((int)(FastPin::port())-0x20 < 64) { asm __volatile__("out %[PORT], %[lo]" ASM_VARS ); } else { *FastPin::port()=lo; } // 2 cycles, sbrs on flipping the line to lo if we're pushing out a 0 #define QLO2(B, N) asm __volatile__("sbrs %[" #B "], " #N ASM_VARS ); LO1; // load a byte from ram into the given var with the given offset #define LD2(B,O) asm __volatile__("ldd %[" #B "], Z + %[" #O "]\n\t" ASM_VARS ); // 4 cycles - load a byte from ram into the scaling scratch space with the given offset, clear the target var, clear carry #define LDSCL4(B,O) asm __volatile__("ldd %[scale_base], Z + %[" #O "]\n\tclr %[" #B "]\n\tclc\n\t" ASM_VARS ); #if (DITHER==1) // apply dithering value before we do anything with scale_base #define PRESCALE4(D) asm __volatile__("cpse %[scale_base], __zero_reg__\n\t add %[scale_base],%[" #D "]\n\tbrcc L_%=\n\tldi %[scale_base], 0xFF\n\tL_%=:\n\t" ASM_VARS); // Do the add for the prescale #define PRESCALEA2(D) asm __volatile__("cpse %[scale_base], __zero_reg__\n\t add %[scale_base],%[" #D "]\n\t" ASM_VARS); // Do the clamp for the prescale, clear carry when we're done - NOTE: Must ensure carry flag state is preserved! #define PRESCALEB4(D) asm __volatile__("brcc L_%=\n\tldi %[scale_base], 0xFF\n\tL_%=:\n\tneg %[" #D "]\n\tCLC" ASM_VARS); // Clamp for prescale, increment data, since we won't ever wrap 65k, this also effectively clears carry for us #define PSBIDATA4(D) asm __volatile__("brcc L_%=\n\tldi %[scale_base], 0xFF\n\tL_%=:\n\tadd %A[data], %[ADV]\n\tadc %B[data], __zero_reg__\n\t" ASM_VARS); #else #define PRESCALE4(D) _dc<4>(loopvar); #define PRESCALEA2(D) _dc<2>(loopvar); #define PRESCALEB4(D) _dc<4>(loopvar); #define PSBIDATA4(D) asm __volatile__( "add %A[data], %[ADV]\n\tadc %B[data], __zero_reg__\n\trjmp .+0\n\t" ASM_VARS ); #endif // 2 cycles - perform one step of the scaling (if a given bit is set in scale, add scale-base to the scratch space) #define _SCALE02(B, N) "sbrc %[s0], " #N "\n\tadd %[" #B "], %[scale_base]\n\t" #define _SCALE12(B, N) "sbrc %[s1], " #N "\n\tadd %[" #B "], %[scale_base]\n\t" #define _SCALE22(B, N) "sbrc %[s2], " #N "\n\tadd %[" #B "], %[scale_base]\n\t" #define SCALE02(B,N) asm __volatile__( _SCALE02(B,N) ASM_VARS ); #define SCALE12(B,N) asm __volatile__( _SCALE12(B,N) ASM_VARS ); #define SCALE22(B,N) asm __volatile__( _SCALE22(B,N) ASM_VARS ); // 1 cycle - rotate right, pulling in from carry #define _ROR1(B) "ror %[" #B "]\n\t" #define ROR1(B) asm __volatile__( _ROR1(B) ASM_VARS); // 1 cycle, clear the carry bit #define _CLC1 "clc\n\t" #define CLC1 asm __volatile__( _CLC1 ASM_VARS ); // 2 cycles, rortate right, pulling in from carry then clear the carry bit #define RORCLC2(B) asm __volatile__( _ROR1(B) _CLC1 ASM_VARS ); // 4 cycles, rotate, clear carry, scale next bit #define RORSC04(B, N) asm __volatile__( _ROR1(B) _CLC1 _SCALE02(B, N) ASM_VARS ); #define RORSC14(B, N) asm __volatile__( _ROR1(B) _CLC1 _SCALE12(B, N) ASM_VARS ); #define RORSC24(B, N) asm __volatile__( _ROR1(B) _CLC1 _SCALE22(B, N) ASM_VARS ); // 4 cycles, scale bit, rotate, clear carry #define SCROR04(B, N) asm __volatile__( _SCALE02(B,N) _ROR1(B) _CLC1 ASM_VARS ); #define SCROR14(B, N) asm __volatile__( _SCALE12(B,N) _ROR1(B) _CLC1 ASM_VARS ); #define SCROR24(B, N) asm __volatile__( _SCALE22(B,N) _ROR1(B) _CLC1 ASM_VARS ); ///////////////////////////////////////////////////////////////////////////////////// // Loop life cycle // dither adjustment macro - should be kept in sync w/what's in stepDithering // #define ADJDITHER2(D, E) D = E - D; #define _NEGD1(D) "neg %[" #D "]\n\t" #define _ADJD1(D,E) "add %[" #D "], %[" #E "]\n\t" #define ADJDITHER2(D, E) asm __volatile__ ( _NEGD1(D) _ADJD1(D, E) ASM_VARS); #define ADDDE1(D, E) asm __volatile__ ( _ADJD1(D, E) ASM_VARS ); // #define xstr(a) str(a) // #define str(a) #a // #define ADJDITHER2(D,E) asm __volatile__("subi %[" #D "], " xstr(DUSE) "\n\tand %[" #D "], %[" #E "]\n\t" ASM_VARS); // define the beginning of the loop #define LOOP asm __volatile__("1:" ASM_VARS ); // define the end of the loop #define DONE asm __volatile__("2:" ASM_VARS ); // 2 cycles - increment the data pointer #define IDATA2 asm __volatile__("add %A[data], %[ADV]\n\tadc %B[data], __zero_reg__\n\t" ASM_VARS ); #define IDATACLC3 asm __volatile__("add %A[data], %[ADV]\n\tadc %B[data], __zero_reg__\n\t" _CLC1 ASM_VARS ); // 1 cycle mov #define _MOV1(B1, B2) "mov %[" #B1 "], %[" #B2 "]\n\t" #define MOV1(B1, B2) asm __volatile__( _MOV1(B1,B2) ASM_VARS ); // 3 cycle mov - skip if scale fix is happening #if (FASTLED_SCALE8_FIXED == 1) #define _MOV_FIX03(B1, B2) "mov %[" #B1 "], %[scale_base]\n\tcpse %[s0], __zero_reg__\n\t" _MOV1(B1, B2) #define _MOV_FIX13(B1, B2) "mov %[" #B1 "], %[scale_base]\n\tcpse %[s1], __zero_reg__\n\t" _MOV1(B1, B2) #define _MOV_FIX23(B1, B2) "mov %[" #B1 "], %[scale_base]\n\tcpse %[s2], __zero_reg__\n\t" _MOV1(B1, B2) #else // if we haven't fixed scale8, just do the move and nop the 2 cycles that would be used to // do the fixed adjustment #define _MOV_FIX03(B1, B2) _MOV1(B1, B2) "rjmp .+0\n\t" #define _MOV_FIX13(B1, B2) _MOV1(B1, B2) "rjmp .+0\n\t" #define _MOV_FIX23(B1, B2) _MOV1(B1, B2) "rjmp .+0\n\t" #endif // 3 cycle mov + negate D for dither adjustment #define MOV_NEGD04(B1, B2, D) asm __volatile( _MOV_FIX03(B1, B2) _NEGD1(D) ASM_VARS ); #define MOV_ADDDE04(B1, B2, D, E) asm __volatile( _MOV_FIX03(B1, B2) _ADJD1(D, E) ASM_VARS ); #define MOV_NEGD14(B1, B2, D) asm __volatile( _MOV_FIX13(B1, B2) _NEGD1(D) ASM_VARS ); #define MOV_ADDDE14(B1, B2, D, E) asm __volatile( _MOV_FIX13(B1, B2) _ADJD1(D, E) ASM_VARS ); #define MOV_NEGD24(B1, B2, D) asm __volatile( _MOV_FIX23(B1, B2) _NEGD1(D) ASM_VARS ); // 2 cycles - decrement the counter #define DCOUNT2 asm __volatile__("sbiw %[count], 1" ASM_VARS ); // 2 cycles - jump to the beginning of the loop #define JMPLOOP2 asm __volatile__("rjmp 1b" ASM_VARS ); // 2 cycles - jump out of the loop #define BRLOOP1 asm __volatile__("brne 3\n\trjmp 2f\n\t3:" ASM_VARS ); // 5 cycles 2 sbiw, 3 for the breq/rjmp #define ENDLOOP5 asm __volatile__("sbiw %[count], 1\n\tbreq L_%=\n\trjmp 1b\n\tL_%=:\n\t" ASM_VARS); // NOP using the variables, forcing a move #define DNOP asm __volatile__("mov r0,r0" ASM_VARS); #define DADVANCE 3 #define DUSE (0xFF - (DADVANCE-1)) // This method is made static to force making register Y available to use for data on AVR - if the method is non-static, then // gcc will use register Y for the this pointer. static void /*__attribute__((optimize("O0")))*/ /*__attribute__ ((always_inline))*/ showRGBInternal(PixelController & pixels) { uint8_t *data = (uint8_t*)pixels.mData; data_ptr_t port = FastPin::port(); data_t mask = FastPin::mask(); uint8_t scale_base = 0; // register uint8_t *end = data + nLeds; data_t hi = *port | mask; data_t lo = *port & ~mask; *port = lo; // the byte currently being written out uint8_t b0 = 0; // the byte currently being worked on to write the next out uint8_t b1 = 0; // Setup the pixel controller pixels.preStepFirstByteDithering(); // pull the dithering/adjustment values out of the pixels object for direct asm access uint8_t advanceBy = pixels.advanceBy(); uint16_t count = pixels.mLen; uint8_t s0 = pixels.mScale.raw[RO(0)]; uint8_t s1 = pixels.mScale.raw[RO(1)]; uint8_t s2 = pixels.mScale.raw[RO(2)]; #if (FASTLED_SCALE8_FIXED==1) s0++; s1++; s2++; #endif uint8_t d0 = pixels.d[RO(0)]; uint8_t d1 = pixels.d[RO(1)]; uint8_t d2 = pixels.d[RO(2)]; uint8_t e0 = pixels.e[RO(0)]; uint8_t e1 = pixels.e[RO(1)]; uint8_t e2 = pixels.e[RO(2)]; uint8_t loopvar=0; // This has to be done in asm to keep gcc from messing up the asm code further down b0 = data[RO(0)]; { LDSCL4(b0,O0) PRESCALEA2(d0) PRESCALEB4(d0) SCALE02(b0,0) RORSC04(b0,1) ROR1(b0) CLC1 SCROR04(b0,2) SCALE02(b0,3) RORSC04(b0,4) ROR1(b0) CLC1 SCROR04(b0,5) SCALE02(b0,6) RORSC04(b0,7) ROR1(b0) CLC1 MOV_ADDDE04(b1,b0,d0,e0) MOV1(b0,b1) } { // while(--count) { // Loop beginning DNOP; LOOP; // Sum of the clock counts across each row should be 10 for 8Mhz, WS2811 // The values in the D1/D2/D3 indicate how many cycles the previous column takes // to allow things to line back up. // // While writing out byte 0, we're loading up byte 1, applying the dithering adjustment, // then scaling it using 8 cycles of shift/add interleaved in between writing the bits // out. When doing byte 1, we're doing the above for byte 2. When we're doing byte 2, // we're cycling back around and doing the above for byte 0. // Inline scaling - RGB ordering // DNOP HI1 _D1(1) QLO2(b0, 7) LDSCL4(b1,O1) _D2(4) LO1 PRESCALEA2(d1) _D3(2) HI1 _D1(1) QLO2(b0, 6) PRESCALEB4(d1) _D2(4) LO1 SCALE12(b1,0) _D3(2) HI1 _D1(1) QLO2(b0, 5) RORSC14(b1,1) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 4) SCROR14(b1,2) _D2(4) LO1 SCALE12(b1,3) _D3(2) HI1 _D1(1) QLO2(b0, 3) RORSC14(b1,4) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 2) SCROR14(b1,5) _D2(4) LO1 SCALE12(b1,6) _D3(2) HI1 _D1(1) QLO2(b0, 1) RORSC14(b1,7) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 0) switch(XTRA0) { case 4: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 3: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 2: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 1: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) } MOV_ADDDE14(b0,b1,d1,e1) _D2(4) LO1 _D3(0) HI1 _D1(1) QLO2(b0, 7) LDSCL4(b1,O2) _D2(4) LO1 PRESCALEA2(d2) _D3(2) HI1 _D1(1) QLO2(b0, 6) PSBIDATA4(d2) _D2(4) LO1 SCALE22(b1,0) _D3(2) HI1 _D1(1) QLO2(b0, 5) RORSC24(b1,1) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 4) SCROR24(b1,2) _D2(4) LO1 SCALE22(b1,3) _D3(2) HI1 _D1(1) QLO2(b0, 3) RORSC24(b1,4) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 2) SCROR24(b1,5) _D2(4) LO1 SCALE22(b1,6) _D3(2) HI1 _D1(1) QLO2(b0, 1) RORSC24(b1,7) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 0) switch(XTRA0) { case 4: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 3: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 2: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 1: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) } // Because Prescale on the middle byte also increments the data counter, // we have to do both halves of updating d2 here - negating it (in the // MOV_NEGD24 macro) and then adding E back into it MOV_NEGD24(b0,b1,d2) _D2(4) LO1 ADDDE1(d2,e2) _D3(1) HI1 _D1(1) QLO2(b0, 7) LDSCL4(b1,O0) _D2(4) LO1 PRESCALEA2(d0) _D3(2) HI1 _D1(1) QLO2(b0, 6) PRESCALEB4(d0) _D2(4) LO1 SCALE02(b1,0) _D3(2) HI1 _D1(1) QLO2(b0, 5) RORSC04(b1,1) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 4) SCROR04(b1,2) _D2(4) LO1 SCALE02(b1,3) _D3(2) HI1 _D1(1) QLO2(b0, 3) RORSC04(b1,4) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 2) SCROR04(b1,5) _D2(4) LO1 SCALE02(b1,6) _D3(2) HI1 _D1(1) QLO2(b0, 1) RORSC04(b1,7) _D2(4) LO1 RORCLC2(b1) _D3(2) HI1 _D1(1) QLO2(b0, 0) switch(XTRA0) { case 4: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 3: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 2: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) case 1: _D2(0) LO1 _D3(0) HI1 _D1(1) QLO2(b0,0) } MOV_ADDDE04(b0,b1,d0,e0) _D2(4) LO1 _D3(5) ENDLOOP5 } DONE; } #if (FASTLED_ALLOW_INTERRUPTS == 1) // stop using the clock juggler TCCR0A &= ~0x30; #endif } }; #endif FASTLED_NAMESPACE_END #endif