📄 hipe_sparc.c
字号:
/* $Id$ */#include <stddef.h> /* offsetof() */#ifdef HAVE_CONFIG_H#include "config.h"#endif#include "global.h"#include <sys/mman.h>#include "hipe_arch.h"#include "hipe_native_bif.h" /* nbif_callemu() *//* Flush dcache and invalidate icache for a range of addresses. */void hipe_flush_icache_range(void *address, unsigned int nbytes){ char *a = (char*)address; int n = nbytes; while( n > 0 ) { hipe_flush_icache_word(a); a += 4; n -= 4; }}static void patch_sethi(Uint32 *address, unsigned int imm22){ unsigned int insn = *address; *address = (insn & 0xFFC00000) | (imm22 & 0x003FFFFF); hipe_flush_icache_word(address);}static void patch_ori(Uint32 *address, unsigned int imm10){ /* address points to an OR reg,imm,reg insn */ unsigned int insn = *address; *address = (insn & 0xFFFFE000) | (imm10 & 0x3FF); hipe_flush_icache_word(address);}static void patch_sethi_ori(Uint32 *address, Uint32 value){ patch_sethi(address, value >> 10); patch_ori(address+1, value);}void hipe_patch_load_fe(Uint32 *address, Uint32 value){ patch_sethi_ori(address, value);}int hipe_patch_insn(void *address, Uint32 value, Eterm type){ switch (type) { case am_load_mfa: case am_atom: case am_constant: case am_closure: case am_c_const: break; default: return -1; } patch_sethi_ori((Uint32*)address, value); return 0;}int hipe_patch_call(void *callAddress, void *destAddress, void *trampoline){ Uint32 relDest, newI; if (trampoline) return -1; relDest = (Uint32)((Sint32)destAddress - (Sint32)callAddress); newI = (1 << 30) | (relDest >> 2); *(Uint32*)callAddress = newI; hipe_flush_icache_word(callAddress); return 0;}/* * Memory allocator for executable code. * * This is required on x86 because some combinations * of Linux kernels and CPU generations default to * non-executable memory mappings, causing ordinary * malloc() memory to be non-executable. */static unsigned int code_bytes;static char *code_next;#if 0 /* change to non-zero to get allocation statistics at exit() */static unsigned int total_mapped, nr_joins, nr_splits, total_alloc, nr_allocs, nr_large, total_lost;static unsigned int atexit_done;static void alloc_code_stats(void){ printf("\r\nalloc_code_stats: %u bytes mapped, %u joins, %u splits, %u bytes allocated, %u average alloc, %u large allocs, %u bytes lost\r\n", total_mapped, nr_joins, nr_splits, total_alloc, nr_allocs ? total_alloc/nr_allocs : 0, nr_large, total_lost);}static void atexit_alloc_code_stats(void){ if (!atexit_done) { atexit_done = 1; (void)atexit(alloc_code_stats); }}#define ALLOC_CODE_STATS(X) do{X;}while(0)#else#define ALLOC_CODE_STATS(X) do{}while(0)#endifstatic void morecore(unsigned int alloc_bytes){ unsigned int map_bytes; char *map_hint, *map_start; /* Page-align the amount to allocate. */ map_bytes = (alloc_bytes + 4095) & ~4095; /* Round up small allocations. */ if (map_bytes < 1024*1024) map_bytes = 1024*1024; else ALLOC_CODE_STATS(++nr_large); /* Create a new memory mapping, ensuring it is executable and in the low 2GB of the address space. Also attempt to make it adjacent to the previous mapping. */ map_hint = code_next + code_bytes; if ((unsigned long)map_hint & 4095) abort(); map_start = mmap(map_hint, map_bytes, PROT_EXEC|PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS#ifdef __x86_64__ |MAP_32BIT#endif , -1, 0); if (map_start == MAP_FAILED) { perror("mmap"); abort(); } ALLOC_CODE_STATS(total_mapped += map_bytes); /* Merge adjacent mappings, so the trailing portion of the previous mapping isn't lost. In practice this is quite successful. */ if (map_start == map_hint) { ALLOC_CODE_STATS(++nr_joins); code_bytes += map_bytes; } else { ALLOC_CODE_STATS(++nr_splits); ALLOC_CODE_STATS(total_lost += code_bytes); code_next = map_start; code_bytes = map_bytes; } ALLOC_CODE_STATS(atexit_alloc_code_stats());}static void *alloc_code(unsigned int alloc_bytes){ void *res; /* Align function entries. */ alloc_bytes = (alloc_bytes + 3) & ~3; if (code_bytes < alloc_bytes) morecore(alloc_bytes); ALLOC_CODE_STATS(++nr_allocs); ALLOC_CODE_STATS(total_alloc += alloc_bytes); res = code_next; code_next += alloc_bytes; code_bytes -= alloc_bytes; return res;}void *hipe_alloc_code(Uint nrbytes, Eterm callees, Eterm *trampolines, Process *p){ if (is_not_nil(callees)) return NULL; *trampolines = NIL; return alloc_code(nrbytes);}/* called from hipe_bif0.c:hipe_bifs_make_native_stub_2() and hipe_bif0.c:hipe_make_stub() */void *hipe_make_native_stub(void *beamAddress, unsigned int beamArity){ unsigned int *code; unsigned int callEmuOffset; int i; code = alloc_code(5*sizeof(int)); /* sethi %hi(Address), %g1 */ code[0] = 0x03000000 | (((unsigned int)beamAddress >> 10) & 0x3FFFFF); /* or %g0, %o7, %l6 ! mov %o7, %l6 */ code[1] = 0xAC10000F; /* or %g1, %lo(Address), %g1 */ code[2] = 0x82106000 | ((unsigned int)beamAddress & 0x3FF); /* call callemu */ callEmuOffset = (char*)nbif_callemu - (char*)&code[3]; code[3] = (1 << 30) | ((callEmuOffset >> 2) & 0x3FFFFFFF); /* or %g0, Arity, %l7 ! mov Arity, %l7 */ code[4] = 0xAE102000 | (beamArity & 0x0FFF); /* flush I-cache as if by write_u32() */ for(i = 0; i < 5; ++i) hipe_flush_icache_word(&code[i]); return code;}void hipe_arch_print_pcb(struct hipe_process_state *p){#define U(n,x) \ printf(" % 4d | %s | 0x%08x | |\r\n", offsetof(struct hipe_process_state,x), n, (unsigned)p->x) U("nra ", nra);#undef U}
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -