diff options
author | Keith Whitwell <[email protected]> | 2005-05-18 10:04:27 +0000 |
---|---|---|
committer | Keith Whitwell <[email protected]> | 2005-05-18 10:04:27 +0000 |
commit | b745bf08cd5e772f86360267995a96e9b73384b0 (patch) | |
tree | 88a119c8d57906abec7e747548b6d15d00db3642 /src/mesa | |
parent | 2f74adc41f59db372d9d0fe58155bceb1654bd4d (diff) |
Checkpoint commit: Preliminary version of a facility to emit x86/sse code
to implement vertex emit functions for the t_vertex.c mechanism.
Diffstat (limited to 'src/mesa')
-rw-r--r-- | src/mesa/tnl/t_vertex_sse.c | 937 |
1 files changed, 937 insertions, 0 deletions
diff --git a/src/mesa/tnl/t_vertex_sse.c b/src/mesa/tnl/t_vertex_sse.c new file mode 100644 index 00000000000..b4e2c5b4748 --- /dev/null +++ b/src/mesa/tnl/t_vertex_sse.c @@ -0,0 +1,937 @@ +/* + * Copyright 2003 Tungsten Graphics, inc. + * All Rights Reserved. + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * on the rights to use, copy, modify, merge, publish, distribute, sub + * license, and/or sell copies of the Software, and to permit persons to whom + * the Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL + * TUNGSTEN GRAPHICS AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, + * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR + * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE + * USE OR OTHER DEALINGS IN THE SOFTWARE. + * + * Authors: + * Keith Whitwell <[email protected]> + */ + +#include "glheader.h" +#include "context.h" +#include "colormac.h" +#include "t_context.h" +#include "t_vertex.h" +#include "simple_list.h" + +#include <unistd.h> +#include <sys/types.h> +#include <sys/stat.h> +#include <fcntl.h> + +#define X 0 +#define Y 1 +#define Z 2 +#define W 3 + +struct x86_reg { + GLuint file:3; + GLuint idx:3; + GLuint mod:2; /* mod_REG if this is just a register */ + GLint disp:24; /* only +/- 23bits of offset - should be enough... */ +}; + +struct x86_program { + GLcontext *ctx; + + GLubyte *store; + GLubyte *csr; + + GLuint stack_offset; + + GLboolean inputs_safe; + GLboolean outputs_safe; + + struct x86_reg identity; + struct x86_reg vp0; + struct x86_reg vp1; +}; + + +#define X86_TWOB 0x0f + +/* There are more but these are all we'll use: + */ +enum x86_reg_file { + file_REG32, + file_XMM +}; + +/* Values for mod field of modr/m byte + */ +enum x86_reg_mod { + mod_INDIRECT, + mod_DISP8, + mod_DISP32, + mod_REG +}; + +enum x86_reg_name { + reg_AX, + reg_CX, + reg_DX, + reg_BX, + reg_SP, + reg_BP, + reg_SI, + reg_DI +}; + + +enum x86_cc { + cc_O, /* overflow */ + cc_NO, /* not overflow */ + cc_NAE, /* not above or equal / carry */ + cc_AE, /* above or equal / not carry */ + cc_E, /* equal / zero */ + cc_NE /* not equal / not zero */ +}; + +#define cc_Z cc_E +#define cc_NZ cc_NE + + +/* Create and manipulate registers and regmem values: + */ +static struct x86_reg make_reg( GLuint file, + GLuint idx ) +{ + struct x86_reg reg; + + reg.file = file; + reg.idx = idx; + reg.mod = mod_REG; + reg.disp = 0; + + return reg; +} + +static struct x86_reg make_disp( struct x86_reg reg, + GLint disp ) +{ + assert(reg.file == file_REG32); + + if (reg.mod == mod_REG) + reg.disp = disp; + else + reg.disp += disp; + + if (reg.disp == 0) + reg.mod = mod_INDIRECT; + else if (reg.disp <= 127 && reg.disp >= -128) + reg.mod = mod_DISP8; + else + reg.mod = mod_DISP32; + + return reg; +} + +/* Retreive a reference to one of the function arguments, taking into + * account any push/pop activity: + */ +static struct x86_reg make_fn_arg( struct x86_program *p, + GLuint arg ) +{ + return make_disp(make_reg(file_REG32, reg_SP), + p->stack_offset + arg * 4); /* ??? */ +} + + +static struct x86_reg get_identity( struct x86_program *p ) +{ + return p->identity; +} + +static struct x86_reg get_sse_temp( struct x86_program *p ) +{ + return make_reg(file_XMM, 7); /* hardwired */ +} + +static void release_temp( struct x86_program *p, + struct x86_reg reg ) +{ + assert(reg.file == file_XMM && + reg.idx == 7); +} + +/* Emit bytes to the instruction stream: + */ +static void emit_1b( struct x86_program *p, GLbyte b0 ) +{ + *(GLbyte *)(p->csr++) = b0; +} + +static void emit_1ub( struct x86_program *p, GLubyte b0 ) +{ + *(p->csr++) = b0; +} + +static void emit_2ub( struct x86_program *p, GLubyte b0, GLubyte b1 ) +{ + *(p->csr++) = b0; + *(p->csr++) = b1; +} + +static void emit_3ub( struct x86_program *p, GLubyte b0, GLubyte b1, GLubyte b2 ) +{ + *(p->csr++) = b0; + *(p->csr++) = b1; + *(p->csr++) = b2; +} + +static void emit_1i( struct x86_program *p, GLint i0 ) +{ + *(GLint *)(p->csr) = i0; + p->csr += 4; +} + + +/* Labels, jumps and fixup: + */ +static GLubyte *get_label( struct x86_program *p ) +{ + return p->csr; +} + +static void emit_jcc( struct x86_program *p, + GLuint cc, + GLubyte *label ) +{ + GLint offset = label - (get_label(p) + 2); + + if (offset <= 127 && offset >= -128) { + emit_1ub(p, 0x70 + cc); + emit_1b(p, (GLbyte) offset); + } + else { + offset = label - (get_label(p) + 5); + emit_2ub(p, 0x0f, 0x80 + cc); + emit_1i(p, offset); + } +} + +/* Always use a 32bit offset for forward jumps: + */ +static GLubyte *emit_jcc_forward( struct x86_program *p, + GLuint cc ) +{ + emit_2ub(p, 0x0f, 0x80 + cc); + emit_1i(p, 0); + return get_label(p); +} + +/* Fixup offset from forward jump: + */ +static void do_fixup( struct x86_program *p, + GLubyte *fixup ) +{ + *(int *)(fixup - 4) = get_label(p) - fixup; +} + +static void emit_push( struct x86_program *p, + struct x86_reg reg ) +{ + assert(reg.mod == mod_REG); + emit_1ub(p, 0x50 + reg.idx); + p->stack_offset += 4; +} + +static void emit_pop( struct x86_program *p, + struct x86_reg reg ) +{ + assert(reg.mod == mod_REG); + emit_1ub(p, 0x58 + reg.idx); + p->stack_offset -= 4; +} + +static void emit_inc( struct x86_program *p, + struct x86_reg reg ) +{ + assert(reg.mod == mod_REG); + emit_1ub(p, 0x40 + reg.idx); +} + +static void emit_dec( struct x86_program *p, + struct x86_reg reg ) +{ + assert(reg.mod == mod_REG); + emit_1ub(p, 0x40 + reg.idx); +} + +static void emit_ret( struct x86_program *p ) +{ + emit_1ub(p, 0xc3); +} + + + + +/* Build a modRM byte + possible displacement. No treatment of SIB + * indexing. BZZT - no way to encode an absolute address. + */ +static void emit_modrm( struct x86_program *p, + struct x86_reg reg, + struct x86_reg regmem ) +{ + GLubyte val = 0; + + assert(reg.mod == mod_REG); + + val |= regmem.mod << 6; /* mod field */ + val |= reg.idx << 3; /* reg field */ + val |= regmem.idx; /* r/m field */ + + emit_1ub(p, val); + + switch (regmem.mod) { + case mod_REG: + case mod_INDIRECT: + break; + case mod_DISP8: + emit_1b(p, regmem.disp); + case mod_DISP32: + emit_1i(p, regmem.disp); + } +} + +/* Many x86 instructions have two opcodes to cope with the situations + * where the destination is a register or memory reference + * respectively. This function selects the correct opcode based on + * the arguments presented. + */ +static void emit_op_modrm( struct x86_program *p, + GLubyte op_dst_is_reg, + GLubyte op_dst_is_mem, + struct x86_reg dst, + struct x86_reg src ) +{ + switch (dst.mod) { + case mod_REG: + emit_1ub(p, op_dst_is_reg); + emit_modrm(p, dst, src); + break; + case mod_INDIRECT: + case mod_DISP32: + case mod_DISP8: + assert(src.mod == mod_REG); + emit_1ub(p, op_dst_is_mem); + emit_modrm(p, src, dst); + break; + } +} + +static void emit_mov( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_op_modrm( p, 0x8b, 0x89, dst, src ); +} + +static void emit_xor( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_op_modrm( p, 0x33, 0x31, dst, src ); +} + +static void emit_movlps( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_1ub(p, X86_TWOB); + emit_op_modrm( p, 0x12, 0x13, dst, src ); +} + +static void emit_movhps( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_1ub(p, X86_TWOB); + emit_op_modrm( p, 0x16, 0x17, dst, src ); +} + +static void emit_movd( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_2ub(p, 0x66, X86_TWOB); + emit_op_modrm( p, 0x6e, 0x7e, dst, src ); +} + +static void emit_movss( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_2ub(p, 0xF3, X86_TWOB); + emit_op_modrm( p, 0x10, 0x11, dst, src ); +} + +static void emit_movaps( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_1ub(p, X86_TWOB); + emit_op_modrm( p, 0x28, 0x29, dst, src ); +} + +static void emit_movups( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_1ub(p, X86_TWOB); + emit_op_modrm( p, 0x10, 0x11, dst, src ); +} + +/* SSE operations often only have one format, with dest constrained to + * be a register: + */ +static void emit_mulps( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_2ub(p, X86_TWOB, 0x59); + emit_modrm( p, dst, src ); +} + +static void emit_addps( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_2ub(p, X86_TWOB, 0x58); + emit_modrm( p, dst, src ); +} + +static void emit_cvtps2dq( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_3ub(p, 0x66, X86_TWOB, 0x5B); + emit_modrm( p, dst, src ); +} + +static void emit_packssdw( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_3ub(p, 0x66, X86_TWOB, 0x6B); + emit_modrm( p, dst, src ); +} + +static void emit_packsswb( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_3ub(p, 0x66, X86_TWOB, 0x63); + emit_modrm( p, dst, src ); +} + +/* Load effective address: + */ +static void emit_lea( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src ) +{ + emit_1ub(p, 0x8d); + emit_modrm( p, dst, src ); +} + +static void emit_add_imm( struct x86_program *p, + struct x86_reg dst, + struct x86_reg src, + GLint value ) +{ + emit_lea(p, dst, make_disp(src, value)); +} + + + + +/** + * Perform a reduced swizzle: + */ +static void emit_pshufd( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0, + GLubyte x, + GLubyte y, + GLubyte z, + GLubyte w) +{ + emit_3ub(p, 0x66, X86_TWOB, 0x70); + emit_modrm(p, dest, arg0); + emit_1ub(p, (x|(y<<2)|(z<<4)|w<<6)); +} + + +static void emit_pk4ub( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_cvtps2dq(p, dest, arg0); + emit_packssdw(p, dest, dest); + emit_packsswb(p, dest, dest); +} + +static void emit_load4f_4( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movups(p, dest, arg0); +} + +static void emit_load4f_3( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + /* Have to jump through some hoops: + * + * 0 0 0 1 -- skip if reg[3] preserved over loop iterations + * c 0 0 1 + * 0 0 c 1 + * a b c 1 + */ + emit_movups(p, dest, get_identity(p)); + emit_movss(p, dest, make_disp(arg0, 8)); + emit_pshufd(p, dest, dest, Y,Z,X,W ); + emit_movlps(p, dest, arg0); +} + +static void emit_load4f_2( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + /* Pull in 2 dwords, then copy the top 2 dwords with 0,1 from id. + */ + emit_movlps(p, dest, arg0); + emit_movhps(p, dest, get_identity(p)); +} + +static void emit_load4f_1( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + /* Initialized with [0,0,0,1] from id, then pull in the single low + * word. + */ + emit_movaps(p, dest, get_identity(p)); + emit_movss(p, dest, arg0); +} + + + +static void emit_load3f_3( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + /* Over-reads by 1 dword - potential SEGV... Deal with in + * array_cache by treating size-3 arrays specially, copying to + * temporary storage if last element (how can you tell?) falls on a + * 4k boundary. + */ + if (p->inputs_safe) { + emit_movaps(p, dest, arg0); + } + else { + /* c . . . + * c c c c + * a b c c + */ + emit_movss(p, dest, make_disp(arg0, 8)); + emit_pshufd(p, dest, dest, X,X,X,X); + emit_movlps(p, dest, arg0); + } +} + +static void emit_load3f_2( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_load4f_2(p, dest, arg0); +} + +static void emit_load3f_1( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_load4f_1(p, dest, arg0); +} + +static void emit_load2f_2( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movlps(p, dest, arg0); +} + +static void emit_load2f_1( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_load4f_1(p, dest, arg0); +} + +static void emit_load1f_1( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movss(p, dest, arg0); +} + +static void (*load[4][4])( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) = { + { emit_load1f_1, + emit_load1f_1, + emit_load1f_1, + emit_load1f_1 }, + + { emit_load2f_1, + emit_load2f_2, + emit_load2f_2, + emit_load2f_2 }, + + { emit_load3f_1, + emit_load3f_2, + emit_load3f_3, + emit_load3f_3 }, + + { emit_load4f_1, + emit_load4f_2, + emit_load4f_3, + emit_load4f_4 } +}; + +static void emit_load( struct x86_program *p, + struct x86_reg temp, + GLuint sz, + struct x86_reg src, + GLuint src_sz) +{ + load[sz-1][src_sz-1](p, temp, src); +} + + +static void emit_store4f( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movups(p, dest, arg0); +} + +static void emit_store3f( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + if (p->outputs_safe) { + /* Emit the extra dword anyway. This may hurt writecombining, + * may cause other problems. + */ + emit_movups(p, dest, arg0); + } + else { + /* Alternate strategy - emit two, shuffle, emit one. + */ + struct x86_reg tmp = get_sse_temp(p); + emit_movlps(p, dest, arg0); + + emit_pshufd(p, tmp, arg0, Z, Z, Z, Z ); + emit_movss(p, make_disp(dest,8), tmp); + release_temp(p, tmp); + } +} + +static void emit_store2f( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movlps(p, dest, arg0); +} + +static void emit_store1f( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) +{ + emit_movss(p, dest, arg0); +} + + +static void (*store[4])( struct x86_program *p, + struct x86_reg dest, + struct x86_reg arg0 ) = +{ + emit_store1f, + emit_store2f, + emit_store3f, + emit_store4f +}; + +static void emit_store( struct x86_program *p, + struct x86_reg dest, + GLuint sz, + struct x86_reg temp ) + +{ + store[sz-1](p, dest, temp); +} + + +static GLint get_offset( const void *a, const void *b ) +{ + return (const char *)b - (const char *)a; +} + + + +/* Lots of hardcoding + * + * EAX -- pointer to current output vertex + * ECX -- pointer to current attribute + * + */ +static GLboolean build_vertex_emit( struct x86_program *p ) +{ + GLcontext *ctx = p->ctx; + TNLcontext *tnl = TNL_CONTEXT(ctx); + struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx); + struct tnl_clipspace_attr *a = vtx->attr; + GLuint j; + + struct x86_reg vertexEAX = make_reg(file_REG32, reg_AX); + struct x86_reg srcEDI = make_reg(file_REG32, reg_CX); + struct x86_reg countEBP = make_reg(file_REG32, reg_BP); + struct x86_reg vtxESI = make_reg(file_REG32, reg_SI); + struct x86_reg tmp = make_reg(file_XMM, 0); + struct x86_reg vp0 = make_reg(file_XMM, 1); + struct x86_reg vp1 = make_reg(file_XMM, 2); + GLubyte *fixup, *label; + + p->csr = p->store; + + /* Push a few regs? + */ + emit_push(p, srcEDI); + emit_push(p, countEBP); + emit_push(p, vtxESI); + + /* Initialize destination register. + */ + emit_mov(p, vertexEAX, make_fn_arg(p, 3)); + + /* Dereference ctx to get tnl, then vtx: + */ + emit_mov(p, vtxESI, make_fn_arg(p, 1)); + emit_mov(p, vtxESI, make_disp(vtxESI, get_offset(ctx, &ctx->swtnl_context))); + vtxESI = make_disp(vtxESI, get_offset(tnl, &tnl->clipspace)); + + /* Get vertex count, compare to zero + */ + emit_mov(p, countEBP, make_fn_arg(p, 2)); + fixup = emit_jcc_forward(p, cc_NZ); + + /* Possibly load vp0, vp1 for viewport calcs: + */ + if (vtx->need_viewport) { + emit_movups(p, vp0, make_disp(vtxESI, get_offset(vtx, &vtx->vp_scale[0]))); + emit_movups(p, vp1, make_disp(vtxESI, get_offset(vtx, &vtx->vp_xlate[0]))); + } + + /* Note address for loop jump */ + label = get_label(p); + + /* Emit code for each of the attributes. Currently routes + * everything through SSE registers, even when it might be more + * efficient to stick with regular old x86. No optimization or + * other tricks - enough new ground to cover here just getting + * things working. + */ + for (j = 0; j < vtx->attr_count; j++) { + struct x86_reg dest = make_disp(vertexEAX, vtx->attr[j].vertoffset); + struct x86_reg ptr_to_src = make_disp(vtxESI, get_offset(vtx, &vtx->attr[j].inputptr)); + + /* Load current a[j].inputptr + */ + emit_mov(p, srcEDI, ptr_to_src); + + /* Now, load an XMM reg from src, perhaps transform, then save. + * Could be shortcircuited in specific cases: + */ + switch (a[j].format) { + case EMIT_1F: + emit_load(p, tmp, 1, srcEDI, vtx->attr[j].inputsize); + emit_store(p, dest, 1, tmp); + case EMIT_2F: + emit_load(p, tmp, 2, srcEDI, vtx->attr[j].inputsize); + emit_store(p, dest, 2, tmp); + case EMIT_3F: + /* Potentially the worst case - hardcode 2+1 copying: + */ + emit_load(p, tmp, 3, srcEDI, vtx->attr[j].inputsize); + emit_store(p, dest, 3, tmp); + case EMIT_4F: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_store(p, dest, 4, tmp); + break; + case EMIT_2F_VIEWPORT: + emit_load(p, tmp, 2, srcEDI, vtx->attr[j].inputsize); + emit_mulps(p, dest, vp0); + emit_addps(p, dest, vp1); + emit_store(p, dest, 2, tmp); + break; + case EMIT_3F_VIEWPORT: + emit_load(p, tmp, 3, srcEDI, vtx->attr[j].inputsize); + emit_mulps(p, dest, vp0); + emit_addps(p, dest, vp1); + emit_store(p, dest, 3, tmp); + break; + case EMIT_4F_VIEWPORT: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_mulps(p, dest, vp0); + emit_addps(p, dest, vp1); + emit_store(p, dest, 4, tmp); + break; + case EMIT_3F_XYW: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pshufd(p, tmp, tmp, X, Y, W, Z); + emit_store(p, dest, 3, tmp); + break; + + /* Try and bond 3ub + 1ub pairs into a single 4ub operation? + */ + case EMIT_1UB_1F: + case EMIT_3UB_3F_RGB: + case EMIT_3UB_3F_BGR: + return GL_FALSE; /* add this later */ + + case EMIT_4UB_4F_RGBA: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pk4ub(p, tmp, tmp); + emit_store(p, dest, 1, tmp); + break; + case EMIT_4UB_4F_BGRA: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pshufd(p, tmp, tmp, Z, Y, X, W); + emit_pk4ub(p, tmp, tmp); + emit_store(p, dest, 1, tmp); + break; + case EMIT_4UB_4F_ARGB: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pshufd(p, tmp, tmp, W, X, Y, Z); + emit_pk4ub(p, tmp, tmp); + emit_store(p, dest, 1, tmp); + break; + case EMIT_4UB_4F_ABGR: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pshufd(p, tmp, tmp, W, Z, Y, X); + emit_pk4ub(p, tmp, tmp); + emit_store(p, dest, 1, tmp); + break; + case EMIT_4CHAN_4F_RGBA: + switch (CHAN_TYPE) { + case GL_UNSIGNED_BYTE: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_pk4ub(p, tmp, tmp); + emit_store(p, dest, 1, tmp); + break; + case GL_UNSIGNED_SHORT: + return GL_FALSE; + case GL_FLOAT: + emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize); + emit_store(p, dest, 4, tmp); + break; + default: + break; + } + default: + return GL_FALSE; /* catch any new opcodes */ + } + + /* add a[j].inputstride (hardcoded value - could just as easily + * pull the stride value from memory each time). + */ + emit_add_imm(p, srcEDI, srcEDI, a[j].inputstride); + + /* save new value of a[j].inputptr + */ + emit_mov(p, ptr_to_src, srcEDI); + + } + + /* Next vertex: + */ + emit_add_imm(p, vertexEAX, vertexEAX, vtx->vertex_size); + + /* decr count, loop if not zero + */ + emit_dec(p, countEBP); + emit_jcc(p, cc_NZ, label); + + /* Land forward jump here: + */ + do_fixup(p, fixup); + + /* Pop regs and return + */ + emit_pop(p, vtxESI); + emit_pop(p, countEBP); + emit_pop(p, srcEDI); + emit_ret(p); + + vtx->emit = (tnl_emit_func)p->store; + return GL_TRUE; +} + +void _tnl_generate_sse_emit( GLcontext *ctx ) +{ + struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx); + struct x86_program p; + + memset(&p, 0, sizeof(p)); + p.ctx = ctx; + p.store = MALLOC(1024); + + p.inputs_safe = 1; /* for now */ + p.outputs_safe = 1; /* for now */ + + if (build_vertex_emit(&p)) { + _tnl_register_fastpath( vtx, GL_TRUE ); + + { + static int i = 0; + char filename[100]; + int fd; + + sprintf(filename, "fastpath%d.o", i); + fd = creat(filename, 0600); + if (fd != -1) { + write(fd, p.store, p.csr - p.store); + close(fd); + _mesa_printf("wrote %s\n", filename); + } + } + } + else { + FREE(p.store); + } + + (void)emit_movd; + (void)emit_inc; + (void)emit_xor; +} |