/* * M7memset.S: SPARC M7 optimized memset. * * Copyright (c) 2016, Oracle and/or its affiliates. All rights reserved. */ /* * M7memset.S: M7 optimized memset. * * char *memset(sp, c, n) * * Set an array of n chars starting at sp to the character c. * Return sp. * * Fast assembler language version of the following C-program for memset * which represents the `standard' for the C-library. * * void * * memset(void *sp1, int c, size_t n) * { * if (n != 0) { * char *sp = sp1; * do { * *sp++ = (char)c; * } while (--n != 0); * } * return (sp1); * } * * The algorithm is as follows : * * For small 6 or fewer bytes stores, bytes will be stored. * * For less than 32 bytes stores, align the address on 4 byte boundary. * Then store as many 4-byte chunks, followed by trailing bytes. * * For sizes greater than 32 bytes, align the address on 8 byte boundary. * if (count >= 64) { * store 8-bytes chunks to align the address on 64 byte boundary * if (value to be set is zero && count >= MIN_ZERO) { * Using BIS stores, set the first long word of each * 64-byte cache line to zero which will also clear the * other seven long words of the cache line. * } * else if (count >= MIN_LOOP) { * Using BIS stores, set the first long word of each of * ST_CHUNK cache lines (64 bytes each) before the main * loop is entered. * In the main loop, continue pre-setting the first long * word of each cache line ST_CHUNK lines in advance while * setting the other seven long words (56 bytes) of each * cache line until fewer than ST_CHUNK*64 bytes remain. * Then set the remaining seven long words of each cache * line that has already had its first long word set. * } * store remaining data in 64-byte chunks until less than * 64 bytes remain. * } * Store as many 8-byte chunks, followed by trailing bytes. * * BIS = Block Init Store * Doing the advance store of the first element of the cache line * initiates the displacement of a cache line while only using a single * instruction in the pipeline. That avoids various pipeline delays, * such as filling the miss buffer. The performance effect is * similar to prefetching for normal stores. * The special case for zero fills runs faster and uses fewer instruction * cycles than the normal memset loop. * * We only use BIS for memset of greater than MIN_LOOP bytes because a sequence * BIS stores must be followed by a membar #StoreStore. The benefit of * the BIS store must be balanced against the cost of the membar operation. */ /* * ASI_STBI_P marks the cache line as "least recently used" * which means if many threads are active, it has a high chance * of being pushed out of the cache between the first initializing * store and the final stores. * Thus, we use ASI_STBIMRU_P which marks the cache line as * "most recently used" for all but the last store to the cache line. */ #include <asm/asi.h> #include <asm/page.h> #define ASI_STBI_P ASI_BLK_INIT_QUAD_LDD_P #define ASI_STBIMRU_P ASI_ST_BLKINIT_MRU_P #define ST_CHUNK 24 /* multiple of 4 due to loop unrolling */ #define MIN_LOOP 16320 #define MIN_ZERO 512 .section ".text" .align 32 /* * Define clear_page(dest) as memset(dest, 0, PAGE_SIZE) * (can create a more optimized version later.) */ .globl M7clear_page .globl M7clear_user_page M7clear_page: /* clear_page(dest) */ M7clear_user_page: set PAGE_SIZE, %o1 /* fall through into bzero code */ .size M7clear_page,.-M7clear_page .size M7clear_user_page,.-M7clear_user_page /* * Define bzero(dest, n) as memset(dest, 0, n) * (can create a more optimized version later.) */ .globl M7bzero M7bzero: /* bzero(dest, size) */ mov %o1, %o2 mov 0, %o1 /* fall through into memset code */ .size M7bzero,.-M7bzero .global M7memset .type M7memset, #function .register %g3, #scratch M7memset: mov %o0, %o5 ! copy sp1 before using it cmp %o2, 7 ! if small counts, just write bytes bleu,pn %xcc, .wrchar and %o1, 0xff, %o1 ! o1 is (char)c sll %o1, 8, %o3 or %o1, %o3, %o1 ! now o1 has 2 bytes of c sll %o1, 16, %o3 cmp %o2, 32 blu,pn %xcc, .wdalign or %o1, %o3, %o1 ! now o1 has 4 bytes of c sllx %o1, 32, %o3 or %o1, %o3, %o1 ! now o1 has 8 bytes of c .dbalign: andcc %o5, 7, %o3 ! is sp1 aligned on a 8 byte bound? bz,pt %xcc, .blkalign ! already long word aligned sub %o3, 8, %o3 ! -(bytes till long word aligned) add %o2, %o3, %o2 ! update o2 with new count ! Set -(%o3) bytes till sp1 long word aligned 1: stb %o1, [%o5] ! there is at least 1 byte to set inccc %o3 ! byte clearing loop bl,pt %xcc, 1b inc %o5 ! Now sp1 is long word aligned (sp1 is found in %o5) .blkalign: cmp %o2, 64 ! check if there are 64 bytes to set blu,pn %xcc, .wrshort mov %o2, %o3 andcc %o5, 63, %o3 ! is sp1 block aligned? bz,pt %xcc, .blkwr ! now block aligned sub %o3, 64, %o3 ! o3 is -(bytes till block aligned) add %o2, %o3, %o2 ! o2 is the remainder ! Store -(%o3) bytes till dst is block (64 byte) aligned. ! Use long word stores. ! Recall that dst is already long word aligned 1: addcc %o3, 8, %o3 stx %o1, [%o5] bl,pt %xcc, 1b add %o5, 8, %o5 ! Now sp1 is block aligned .blkwr: andn %o2, 63, %o4 ! calculate size of blocks in bytes brz,pn %o1, .wrzero ! special case if c == 0 and %o2, 63, %o3 ! %o3 = bytes left after blk stores. set MIN_LOOP, %g1 cmp %o4, %g1 ! check there are enough bytes to set blu,pn %xcc, .short_set ! to justify cost of membar ! must be > pre-cleared lines nop ! initial cache-clearing stores ! get store pipeline moving rd %asi, %g3 ! save %asi to be restored later wr %g0, ASI_STBIMRU_P, %asi ! Primary memset loop for large memsets .wr_loop: sub %o5, 8, %o5 ! adjust %o5 for ASI store alignment mov ST_CHUNK, %g1 .wr_loop_start: stxa %o1, [%o5+8]%asi subcc %g1, 4, %g1 stxa %o1, [%o5+8+64]%asi add %o5, 256, %o5 stxa %o1, [%o5+8-128]%asi bgu %xcc, .wr_loop_start stxa %o1, [%o5+8-64]%asi sub %o5, ST_CHUNK*64, %o5 ! reset %o5 mov ST_CHUNK, %g1 .wr_loop_rest: stxa %o1, [%o5+8+8]%asi sub %o4, 64, %o4 stxa %o1, [%o5+16+8]%asi subcc %g1, 1, %g1 stxa %o1, [%o5+24+8]%asi stxa %o1, [%o5+32+8]%asi stxa %o1, [%o5+40+8]%asi add %o5, 64, %o5 stxa %o1, [%o5-8]%asi bgu %xcc, .wr_loop_rest stxa %o1, [%o5]ASI_STBI_P ! If more than ST_CHUNK*64 bytes remain to set, continue ! setting the first long word of each cache line in advance ! to keep the store pipeline moving. cmp %o4, ST_CHUNK*64 bge,pt %xcc, .wr_loop_start mov ST_CHUNK, %g1 brz,a,pn %o4, .asi_done add %o5, 8, %o5 ! restore %o5 offset .wr_loop_small: stxa %o1, [%o5+8]%asi stxa %o1, [%o5+8+8]%asi stxa %o1, [%o5+16+8]%asi stxa %o1, [%o5+24+8]%asi stxa %o1, [%o5+32+8]%asi subcc %o4, 64, %o4 stxa %o1, [%o5+40+8]%asi add %o5, 64, %o5 stxa %o1, [%o5-8]%asi bgu,pt %xcc, .wr_loop_small stxa %o1, [%o5]ASI_STBI_P ba .asi_done add %o5, 8, %o5 ! restore %o5 offset ! Special case loop for zero fill memsets ! For each 64 byte cache line, single STBI to first element ! clears line .wrzero: cmp %o4, MIN_ZERO ! check if enough bytes to set ! to pay %asi + membar cost blu %xcc, .short_set nop sub %o4, 256, %o4 .wrzero_loop: mov 64, %g3 stxa %o1, [%o5]ASI_STBI_P subcc %o4, 256, %o4 stxa %o1, [%o5+%g3]ASI_STBI_P add %o5, 256, %o5 sub %g3, 192, %g3 stxa %o1, [%o5+%g3]ASI_STBI_P add %g3, 64, %g3 bge,pt %xcc, .wrzero_loop stxa %o1, [%o5+%g3]ASI_STBI_P add %o4, 256, %o4 brz,pn %o4, .bsi_done nop .wrzero_small: stxa %o1, [%o5]ASI_STBI_P subcc %o4, 64, %o4 bgu,pt %xcc, .wrzero_small add %o5, 64, %o5 ba,a .bsi_done .asi_done: wr %g3, 0x0, %asi ! restored saved %asi .bsi_done: membar #StoreStore ! required by use of Block Store Init .short_set: cmp %o4, 64 ! check if 64 bytes to set blu %xcc, 5f nop 4: ! set final blocks of 64 bytes stx %o1, [%o5] stx %o1, [%o5+8] stx %o1, [%o5+16] stx %o1, [%o5+24] subcc %o4, 64, %o4 stx %o1, [%o5+32] stx %o1, [%o5+40] add %o5, 64, %o5 stx %o1, [%o5-16] bgu,pt %xcc, 4b stx %o1, [%o5-8] 5: ! Set the remaining long words .wrshort: subcc %o3, 8, %o3 ! Can we store any long words? blu,pn %xcc, .wrchars and %o2, 7, %o2 ! calc bytes left after long words 6: subcc %o3, 8, %o3 stx %o1, [%o5] ! store the long words bgeu,pt %xcc, 6b add %o5, 8, %o5 .wrchars: ! check for extra chars brnz %o2, .wrfin nop retl nop .wdalign: andcc %o5, 3, %o3 ! is sp1 aligned on a word boundary bz,pn %xcc, .wrword andn %o2, 3, %o3 ! create word sized count in %o3 dec %o2 ! decrement count stb %o1, [%o5] ! clear a byte b .wdalign inc %o5 ! next byte .wrword: subcc %o3, 4, %o3 st %o1, [%o5] ! 4-byte writing loop bnz,pt %xcc, .wrword add %o5, 4, %o5 and %o2, 3, %o2 ! leftover count, if any .wrchar: ! Set the remaining bytes, if any brz %o2, .exit nop .wrfin: deccc %o2 stb %o1, [%o5] bgu,pt %xcc, .wrfin inc %o5 .exit: retl ! %o0 was preserved nop .size M7memset,.-M7memset