/////////////////////////////////////////////////////////////////////////// // File : utils.c // Date : 18/10/2013 // Author : alain greiner // Copyright (c) UPMC-LIP6 /////////////////////////////////////////////////////////////////////////// // The utils.c and utils.h files are part of the GIET-VM nano-kernel. /////////////////////////////////////////////////////////////////////////// #include #include #include #include #include #include #include // This variable is allocated in boot.c file or in kernel_init.c file extern static_scheduler_t* _schedulers[X_SIZE][Y_SIZE][NB_PROCS_MAX]; /////////////////////////////////////////////////////////////////////////// // CP0 registers access functions /////////////////////////////////////////////////////////////////////////// //////////////////////////////// static_scheduler_t* _get_sched() { unsigned int ret; asm volatile( "mfc0 %0, $4,2 \n" : "=r"(ret) ); return (static_scheduler_t*)ret; } /////////////////////// unsigned int _get_epc() { unsigned int ret; asm volatile( "mfc0 %0, $14 \n" : "=r"(ret) ); return ret; } //////////////////////// unsigned int _get_bvar() { unsigned int ret; asm volatile( "mfc0 %0, $8 \n" : "=r"(ret)); return ret; } ////////////////////// unsigned int _get_cr() { unsigned int ret; asm volatile( "mfc0 %0, $13 \n" : "=r"(ret)); return ret; } ////////////////////// unsigned int _get_sr() { unsigned int ret; asm volatile( "mfc0 %0, $12 \n" : "=r"(ret)); return ret; } ////////////////////////// unsigned int _get_procid() { unsigned int ret; asm volatile ( "mfc0 %0, $15, 1 \n" :"=r" (ret) ); return (ret & 0xFFF); } //////////////////////////// unsigned int _get_proctime() { unsigned int ret; asm volatile ( "mfc0 %0, $9 \n" :"=r" (ret) ); return ret; } //////////////////////////// void _set_proctime( unsigned int val ) { asm volatile ( "mtc0 $0, $9 \n" : :"r" (val)); } ///////////////////////////////////////////// void _it_disable( unsigned int * save_sr_ptr) { unsigned int sr = 0; asm volatile( "li $3, 0xFFFFFFFE \n" "mfc0 %0, $12 \n" "and $3, $3, %0 \n" "mtc0 $3, $12 \n" : "+r"(sr) : : "$3" ); *save_sr_ptr = sr; } ////////////////////////////////////////////// void _it_restore( unsigned int * save_sr_ptr ) { unsigned int sr = *save_sr_ptr; asm volatile( "mtc0 %0, $12 \n" : : "r"(sr) : "memory" ); } ///////////////////////////////// void _set_sched(unsigned int val) { asm volatile ( "mtc0 %0, $4, 2 \n" : :"r" (val) ); } ////////////////////////////// void _set_sr(unsigned int val) { asm volatile ( "mtc0 %0, $12 \n" : :"r" (val) ); } void _cpu_sync() { asm volatile("sync" ::: "memory"); } /////////////////////////////////////////////////////////////////////////// // CP2 registers access functions /////////////////////////////////////////////////////////////////////////// //////////////////////////// unsigned int _get_mmu_ptpr() { unsigned int ret; asm volatile( "mfc2 %0, $0 \n" : "=r"(ret) ); return ret; } //////////////////////////// unsigned int _get_mmu_mode() { unsigned int ret; asm volatile( "mfc2 %0, $1 \n" : "=r"(ret) ); return ret; } //////////////////////////////////// void _set_mmu_ptpr(unsigned int val) { asm volatile ( "mtc2 %0, $0 \n" : :"r" (val) :"memory" ); } //////////////////////////////////// void _set_mmu_mode(unsigned int val) { asm volatile ( "mtc2 %0, $1 \n" : :"r" (val) :"memory" ); } //////////////////////////////////////////// void _set_mmu_dcache_inval(unsigned int val) { asm volatile ( "mtc2 %0, $7 \n" : :"r" (val) :"memory" ); } /////////////////////////////////////////////////////////////////////////// // Physical addressing related functions /////////////////////////////////////////////////////////////////////////// /////////////////////////////////////////////////////// unsigned int _physical_read( unsigned long long paddr ) { unsigned int value; unsigned int lsb = (unsigned int) paddr; unsigned int msb = (unsigned int) (paddr >> 32); unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $2, $1 \n" /* $2 <= MMU_MODE */ "andi $3, $2, 0xb \n" "mtc2 $3, $1 \n" /* DTLB off */ "mtc2 %2, $24 \n" /* PADDR_EXT <= msb */ "lw %0, 0(%1) \n" /* value <= *paddr */ "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $2, $1 \n" /* restore MMU_MODE */ : "=r" (value) : "r" (lsb), "r" (msb) : "$2", "$3" ); _it_restore(&sr); return value; } //////////////////////////////////////////////// void _physical_write( unsigned long long paddr, unsigned int value ) { unsigned int lsb = (unsigned int)paddr; unsigned int msb = (unsigned int)(paddr >> 32); unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $2, $1 \n" /* $2 <= MMU_MODE */ "andi $3, $2, 0xb \n" "mtc2 $3, $1 \n" /* DTLB off */ "mtc2 %2, $24 \n" /* PADDR_EXT <= msb */ "sw %0, 0(%1) \n" /* *paddr <= value */ "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $2, $1 \n" /* restore MMU_MODE */ "sync \n" : : "r" (value), "r" (lsb), "r" (msb) : "$2", "$3" ); _it_restore(&sr); } ///////////////////////////////////////////////////////////////// unsigned long long _physical_read_ull( unsigned long long paddr ) { unsigned int data_lsb; unsigned int data_msb; unsigned int addr_lsb = (unsigned int) paddr; unsigned int addr_msb = (unsigned int) (paddr >> 32); unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $2, $1 \n" /* $2 <= MMU_MODE */ "andi $3, $2, 0xb \n" "mtc2 $3, $1 \n" /* DTLB off */ "mtc2 %3, $24 \n" /* PADDR_EXT <= msb */ "lw %0, 0(%2) \n" /* data_lsb <= *paddr */ "lw %1, 4(%2) \n" /* data_msb <= *paddr+4 */ "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $2, $1 \n" /* restore MMU_MODE */ : "=r" (data_lsb), "=r"(data_msb) : "r" (addr_lsb), "r" (addr_msb) : "$2", "$3" ); _it_restore(&sr); return ( (((unsigned long long)data_msb)<<32) + (((unsigned long long)data_lsb)) ); } /////////////////////////////////////////////////// void _physical_write_ull( unsigned long long paddr, unsigned long long value ) { unsigned int addr_lsb = (unsigned int)paddr; unsigned int addr_msb = (unsigned int)(paddr >> 32); unsigned int data_lsb = (unsigned int)value; unsigned int data_msb = (unsigned int)(value >> 32); unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $2, $1 \n" /* $2 <= MMU_MODE */ "andi $3, $2, 0xb \n" "mtc2 $3, $1 \n" /* DTLB off */ "mtc2 %3, $24 \n" /* PADDR_EXT <= msb */ "sw %0, 0(%2) \n" /* *paddr <= value */ "sw %1, 4(%2) \n" /* *paddr+4 <= value */ "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $2, $1 \n" /* restore MMU_MODE */ "sync \n" : : "r"(data_lsb),"r"(data_msb),"r"(addr_lsb),"r"(addr_msb) : "$2", "$3" ); _it_restore(&sr); } //////////////////////////////////////////////////// void _physical_memcpy( unsigned long long dst_paddr, // dest buffer paddr unsigned long long src_paddr, // source buffer paddr unsigned int size ) // bytes { // check alignment constraints if ( (dst_paddr & 3) || (src_paddr & 3) || (size & 3) ) { _puts("\n[GIET ERROR] in _physical_memcpy() : buffer unaligned\n"); _exit(); } unsigned int src_lsb = (unsigned int)src_paddr; unsigned int src_msb = (unsigned int)(src_paddr >> 32); unsigned int dst_lsb = (unsigned int)dst_paddr; unsigned int dst_msb = (unsigned int)(dst_paddr >> 32); unsigned int iter = size>>2; unsigned int data; unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $2, $1 \n" /* $2 <= current MMU_MODE */ "andi $3, $2, 0xb \n" /* $3 <= new MMU_MODE */ "mtc2 $3, $1 \n" /* DTLB off */ "move $4, %5 \n" /* $4 < iter */ "move $5, %1 \n" /* $5 < src_lsb */ "move $6, %3 \n" /* $6 < src_lsb */ "ph_memcpy_loop: \n" "mtc2 %2, $24 \n" /* PADDR_EXT <= src_msb */ "lw %0, 0($5) \n" /* data <= *src_paddr */ "mtc2 %4, $24 \n" /* PADDR_EXT <= dst_msb */ "sw %0, 0($6) \n" /* *dst_paddr <= data */ "addi $4, $4, -1 \n" /* iter = iter - 1 */ "addi $5, $5, 4 \n" /* src_lsb += 4 */ "addi $6, $6, 4 \n" /* dst_lsb += 4 */ "bne $4, $0, ph_memcpy_loop \n" "nop \n" "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $2, $1 \n" /* restore MMU_MODE */ : "=r" (data) : "r"(src_lsb),"r"(src_msb),"r"(dst_lsb), "r"(dst_msb), "r"(iter) : "$2", "$3", "$4", "$5", "$6" ); _it_restore(&sr); } // end _physical_memcpy() //////////////////////////////////////////////// void _physical_memset( unsigned long long paddr, // dest buffer paddr unsigned int size, // bytes unsigned int data ) // written value { // check alignment constraints if ( (paddr & 3) || (size & 7) ) { _puts("\n[GIET ERROR] in _physical_memset() : buffer unaligned\n"); _exit(); } unsigned int lsb = (unsigned int)paddr; unsigned int msb = (unsigned int)(paddr >> 32); unsigned int sr; _it_disable(&sr); asm volatile( "mfc2 $8, $1 \n" /* $8 <= current MMU_MODE */ "andi $9, $8, 0xb \n" /* $9 <= new MMU_MODE */ "mtc2 $9, $1 \n" /* DTLB off */ "mtc2 %3, $24 \n" /* PADDR_EXT <= msb */ "1: \n" /* set 8 bytes per iter */ "sw %2, 0(%0) \n" /* *src_paddr = data */ "sw %2, 4(%0) \n" /* *(src_paddr+4) = data */ "addi %1, %1, -8 \n" /* size -= 8 */ "addi %0, %0, 8 \n" /* src_paddr += 8 */ "bnez %1, 1b \n" /* loop while size != 0 */ "mtc2 $0, $24 \n" /* PADDR_EXT <= 0 */ "mtc2 $8, $1 \n" /* restore MMU_MODE */ : "+r"(lsb), "+r"(size) : "r"(data), "r" (msb) : "$8", "$9", "memory" ); _it_restore(&sr); } // _physical_memset() /////////////////////////////////////////////// void _io_extended_write( unsigned int* vaddr, unsigned int value ) { unsigned long long paddr; if ( _get_mmu_mode() & 0x4 ) // MMU activated : use virtual address { *vaddr = value; } else // use paddr extension for IO { paddr = (unsigned long long)(unsigned int)vaddr + (((unsigned long long)((X_IO<current; } //////////////////////////////// unsigned int _get_thread_trdid() { static_scheduler_t* psched = (static_scheduler_t *) _get_sched(); unsigned int current = psched->current; return psched->context[current].slot[CTX_TRDID_ID]; } ////////////////////////////////////////////// unsigned int _get_thread_slot( unsigned int x, unsigned int y, unsigned int p, unsigned int ltid, unsigned int slotid ) { static_scheduler_t* psched = (static_scheduler_t*)_schedulers[x][y][p]; return psched->context[ltid].slot[slotid]; } ////////////////////////////////////// void _set_thread_slot( unsigned int x, unsigned int y, unsigned int p, unsigned int ltid, unsigned int slotid, unsigned int value ) { static_scheduler_t* psched = (static_scheduler_t*)_schedulers[x][y][p]; psched->context[ltid].slot[slotid] = value; } ///////////////////////////////////////////////////// unsigned int _get_context_slot( unsigned int slotid ) { static_scheduler_t* psched = (static_scheduler_t*)_get_sched(); unsigned int ltid = psched->current; return psched->context[ltid].slot[slotid]; } //////////////////////////////////////////// void _set_context_slot( unsigned int slotid, unsigned int value ) { static_scheduler_t* psched = (static_scheduler_t*)_get_sched(); unsigned int ltid = psched->current; psched->context[ltid].slot[slotid] = value; } ///////////////////////////////////////////////////////////////////////////// // Access functions to mapping_info data structure ///////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////// mapping_cluster_t * _get_cluster_base(mapping_header_t * header) { return (mapping_cluster_t *) ((char *) header + MAPPING_HEADER_SIZE); } ////////////////////////////////////////////////////////// mapping_pseg_t * _get_pseg_base(mapping_header_t * header) { return (mapping_pseg_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE); } ////////////////////////////////////////////////////////////// mapping_vspace_t * _get_vspace_base(mapping_header_t * header) { return (mapping_vspace_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs); } ////////////////////////////////////////////////////////// mapping_vseg_t * _get_vseg_base(mapping_header_t * header) { return (mapping_vseg_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs + MAPPING_VSPACE_SIZE * header->vspaces); } ////////////////////////////////////////////////////////////// mapping_thread_t * _get_thread_base(mapping_header_t * header) { return (mapping_thread_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs + MAPPING_VSPACE_SIZE * header->vspaces + MAPPING_VSEG_SIZE * header->vsegs); } ///////////////////////////////////////////////////////// mapping_proc_t *_get_proc_base(mapping_header_t * header) { return (mapping_proc_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs + MAPPING_VSPACE_SIZE * header->vspaces + MAPPING_VSEG_SIZE * header->vsegs + MAPPING_THREAD_SIZE * header->threads); } /////////////////////////////////////////////////////// mapping_irq_t *_get_irq_base(mapping_header_t * header) { return (mapping_irq_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs + MAPPING_VSPACE_SIZE * header->vspaces + MAPPING_VSEG_SIZE * header->vsegs + MAPPING_THREAD_SIZE * header->threads + MAPPING_PROC_SIZE * header->procs); } /////////////////////////////////////////////////////////////// mapping_periph_t *_get_periph_base(mapping_header_t * header) { return (mapping_periph_t *) ((char *) header + MAPPING_HEADER_SIZE + MAPPING_CLUSTER_SIZE * X_SIZE * Y_SIZE + MAPPING_PSEG_SIZE * header->psegs + MAPPING_VSPACE_SIZE * header->vspaces + MAPPING_VSEG_SIZE * header->vsegs + MAPPING_THREAD_SIZE * header->threads + MAPPING_PROC_SIZE * header->procs + MAPPING_IRQ_SIZE * header->irqs); } /////////////////////////////////////////////////////////////////////////// // Miscelaneous functions /////////////////////////////////////////////////////////////////////////// ////////////////////////////////////// __attribute__((noreturn)) void _exit() { unsigned int procid = _get_procid(); unsigned int x = (procid >> (Y_WIDTH + P_WIDTH)) & ((1<> P_WIDTH) & ((1<> 10) & 0x7); line_size = 2 << tmp; // iterate on cache lines for ( offset = 0; offset < buf_size; offset += line_size) { _set_mmu_dcache_inval( buf_vbase + offset ); } } ///////////////////////////////////////////// void _get_sqt_footprint( unsigned int* width, unsigned int* heigth, unsigned int* levels ) { mapping_header_t* header = (mapping_header_t *)SEG_BOOT_MAPPING_BASE; mapping_cluster_t* cluster = _get_cluster_base(header); unsigned int x; unsigned int y; unsigned int cid; unsigned int w = 0; unsigned int h = 0; // scan all clusters to compute SQT footprint (w,h) for ( x = 0 ; x < X_SIZE ; x++ ) { for ( y = 0 ; y < Y_SIZE ; y++ ) { cid = x * Y_SIZE + y; if ( cluster[cid].procs ) // cluster contains processors { if ( x > w ) w = x; if ( y > h ) h = y; } } } *width = w + 1; *heigth = h + 1; // compute SQT levels unsigned int z = (h > w) ? h : w; *levels = (z < 1) ? 1 : (z < 2) ? 2 : (z < 4) ? 3 : (z < 8) ? 4 : 5; } /////////////////////////////////////////////////////////////////////////// // Required by GCC /////////////////////////////////////////////////////////////////////////// //////////////////////////////// void* memcpy( void* dest, // dest buffer vbase const void* source, // source buffer vbase unsigned int size ) // bytes { unsigned int* idst = (unsigned int*)dest; unsigned int* isrc = (unsigned int*)source; // word-by-word copy if both buffers are word aligned if (!((unsigned int) idst & 3) && !((unsigned int) isrc & 3)) { while (size > 3) { *idst++ = *isrc++; size -= 4; } } unsigned char* cdst = (unsigned char*)idst; unsigned char* csrc = (unsigned char*)isrc; // byte-by-byte copy if size not multiple of 4 bytes while (size) { *cdst++ = *csrc++; size--; } return dest; } ///////////////////////////////// void* memset( void* dst, int value, unsigned int count ) { // word-by-word copy unsigned int* idst = dst; unsigned int data = (((unsigned char)value) ) | (((unsigned char)value) << 8) | (((unsigned char)value) << 16) | (((unsigned char)value) << 24) ; if ( ! ((unsigned int)idst & 3) ) { while ( count > 3 ) { *idst++ = data; count -= 4; } } // byte-by-byte copy unsigned char* cdst = dst; while (count--) { *cdst++ = (unsigned char)value; } return dst; } // Local Variables: // tab-width: 4 // c-basic-offset: 4 // c-file-offsets:((innamespace . 0)(inline-open . 0)) // indent-tabs-mode: nil // End: // vim: filetype=c:expandtab:shiftwidth=4:tabstop=4:softtabstop=4