summaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorKeith Whitwell <[email protected]>2005-05-18 10:04:27 +0000
committerKeith Whitwell <[email protected]>2005-05-18 10:04:27 +0000
commitb745bf08cd5e772f86360267995a96e9b73384b0 (patch)
tree88a119c8d57906abec7e747548b6d15d00db3642 /src
parent2f74adc41f59db372d9d0fe58155bceb1654bd4d (diff)
Checkpoint commit: Preliminary version of a facility to emit x86/sse code
to implement vertex emit functions for the t_vertex.c mechanism.
Diffstat (limited to 'src')
-rw-r--r--src/mesa/tnl/t_vertex_sse.c937
1 files changed, 937 insertions, 0 deletions
diff --git a/src/mesa/tnl/t_vertex_sse.c b/src/mesa/tnl/t_vertex_sse.c
new file mode 100644
index 00000000000..b4e2c5b4748
--- /dev/null
+++ b/src/mesa/tnl/t_vertex_sse.c
@@ -0,0 +1,937 @@
+/*
+ * Copyright 2003 Tungsten Graphics, inc.
+ * All Rights Reserved.
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * on the rights to use, copy, modify, merge, publish, distribute, sub
+ * license, and/or sell copies of the Software, and to permit persons to whom
+ * the Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
+ * TUNGSTEN GRAPHICS AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
+ * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
+ * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
+ * USE OR OTHER DEALINGS IN THE SOFTWARE.
+ *
+ * Authors:
+ * Keith Whitwell <[email protected]>
+ */
+
+#include "glheader.h"
+#include "context.h"
+#include "colormac.h"
+#include "t_context.h"
+#include "t_vertex.h"
+#include "simple_list.h"
+
+#include <unistd.h>
+#include <sys/types.h>
+#include <sys/stat.h>
+#include <fcntl.h>
+
+#define X 0
+#define Y 1
+#define Z 2
+#define W 3
+
+struct x86_reg {
+ GLuint file:3;
+ GLuint idx:3;
+ GLuint mod:2; /* mod_REG if this is just a register */
+ GLint disp:24; /* only +/- 23bits of offset - should be enough... */
+};
+
+struct x86_program {
+ GLcontext *ctx;
+
+ GLubyte *store;
+ GLubyte *csr;
+
+ GLuint stack_offset;
+
+ GLboolean inputs_safe;
+ GLboolean outputs_safe;
+
+ struct x86_reg identity;
+ struct x86_reg vp0;
+ struct x86_reg vp1;
+};
+
+
+#define X86_TWOB 0x0f
+
+/* There are more but these are all we'll use:
+ */
+enum x86_reg_file {
+ file_REG32,
+ file_XMM
+};
+
+/* Values for mod field of modr/m byte
+ */
+enum x86_reg_mod {
+ mod_INDIRECT,
+ mod_DISP8,
+ mod_DISP32,
+ mod_REG
+};
+
+enum x86_reg_name {
+ reg_AX,
+ reg_CX,
+ reg_DX,
+ reg_BX,
+ reg_SP,
+ reg_BP,
+ reg_SI,
+ reg_DI
+};
+
+
+enum x86_cc {
+ cc_O, /* overflow */
+ cc_NO, /* not overflow */
+ cc_NAE, /* not above or equal / carry */
+ cc_AE, /* above or equal / not carry */
+ cc_E, /* equal / zero */
+ cc_NE /* not equal / not zero */
+};
+
+#define cc_Z cc_E
+#define cc_NZ cc_NE
+
+
+/* Create and manipulate registers and regmem values:
+ */
+static struct x86_reg make_reg( GLuint file,
+ GLuint idx )
+{
+ struct x86_reg reg;
+
+ reg.file = file;
+ reg.idx = idx;
+ reg.mod = mod_REG;
+ reg.disp = 0;
+
+ return reg;
+}
+
+static struct x86_reg make_disp( struct x86_reg reg,
+ GLint disp )
+{
+ assert(reg.file == file_REG32);
+
+ if (reg.mod == mod_REG)
+ reg.disp = disp;
+ else
+ reg.disp += disp;
+
+ if (reg.disp == 0)
+ reg.mod = mod_INDIRECT;
+ else if (reg.disp <= 127 && reg.disp >= -128)
+ reg.mod = mod_DISP8;
+ else
+ reg.mod = mod_DISP32;
+
+ return reg;
+}
+
+/* Retreive a reference to one of the function arguments, taking into
+ * account any push/pop activity:
+ */
+static struct x86_reg make_fn_arg( struct x86_program *p,
+ GLuint arg )
+{
+ return make_disp(make_reg(file_REG32, reg_SP),
+ p->stack_offset + arg * 4); /* ??? */
+}
+
+
+static struct x86_reg get_identity( struct x86_program *p )
+{
+ return p->identity;
+}
+
+static struct x86_reg get_sse_temp( struct x86_program *p )
+{
+ return make_reg(file_XMM, 7); /* hardwired */
+}
+
+static void release_temp( struct x86_program *p,
+ struct x86_reg reg )
+{
+ assert(reg.file == file_XMM &&
+ reg.idx == 7);
+}
+
+/* Emit bytes to the instruction stream:
+ */
+static void emit_1b( struct x86_program *p, GLbyte b0 )
+{
+ *(GLbyte *)(p->csr++) = b0;
+}
+
+static void emit_1ub( struct x86_program *p, GLubyte b0 )
+{
+ *(p->csr++) = b0;
+}
+
+static void emit_2ub( struct x86_program *p, GLubyte b0, GLubyte b1 )
+{
+ *(p->csr++) = b0;
+ *(p->csr++) = b1;
+}
+
+static void emit_3ub( struct x86_program *p, GLubyte b0, GLubyte b1, GLubyte b2 )
+{
+ *(p->csr++) = b0;
+ *(p->csr++) = b1;
+ *(p->csr++) = b2;
+}
+
+static void emit_1i( struct x86_program *p, GLint i0 )
+{
+ *(GLint *)(p->csr) = i0;
+ p->csr += 4;
+}
+
+
+/* Labels, jumps and fixup:
+ */
+static GLubyte *get_label( struct x86_program *p )
+{
+ return p->csr;
+}
+
+static void emit_jcc( struct x86_program *p,
+ GLuint cc,
+ GLubyte *label )
+{
+ GLint offset = label - (get_label(p) + 2);
+
+ if (offset <= 127 && offset >= -128) {
+ emit_1ub(p, 0x70 + cc);
+ emit_1b(p, (GLbyte) offset);
+ }
+ else {
+ offset = label - (get_label(p) + 5);
+ emit_2ub(p, 0x0f, 0x80 + cc);
+ emit_1i(p, offset);
+ }
+}
+
+/* Always use a 32bit offset for forward jumps:
+ */
+static GLubyte *emit_jcc_forward( struct x86_program *p,
+ GLuint cc )
+{
+ emit_2ub(p, 0x0f, 0x80 + cc);
+ emit_1i(p, 0);
+ return get_label(p);
+}
+
+/* Fixup offset from forward jump:
+ */
+static void do_fixup( struct x86_program *p,
+ GLubyte *fixup )
+{
+ *(int *)(fixup - 4) = get_label(p) - fixup;
+}
+
+static void emit_push( struct x86_program *p,
+ struct x86_reg reg )
+{
+ assert(reg.mod == mod_REG);
+ emit_1ub(p, 0x50 + reg.idx);
+ p->stack_offset += 4;
+}
+
+static void emit_pop( struct x86_program *p,
+ struct x86_reg reg )
+{
+ assert(reg.mod == mod_REG);
+ emit_1ub(p, 0x58 + reg.idx);
+ p->stack_offset -= 4;
+}
+
+static void emit_inc( struct x86_program *p,
+ struct x86_reg reg )
+{
+ assert(reg.mod == mod_REG);
+ emit_1ub(p, 0x40 + reg.idx);
+}
+
+static void emit_dec( struct x86_program *p,
+ struct x86_reg reg )
+{
+ assert(reg.mod == mod_REG);
+ emit_1ub(p, 0x40 + reg.idx);
+}
+
+static void emit_ret( struct x86_program *p )
+{
+ emit_1ub(p, 0xc3);
+}
+
+
+
+
+/* Build a modRM byte + possible displacement. No treatment of SIB
+ * indexing. BZZT - no way to encode an absolute address.
+ */
+static void emit_modrm( struct x86_program *p,
+ struct x86_reg reg,
+ struct x86_reg regmem )
+{
+ GLubyte val = 0;
+
+ assert(reg.mod == mod_REG);
+
+ val |= regmem.mod << 6; /* mod field */
+ val |= reg.idx << 3; /* reg field */
+ val |= regmem.idx; /* r/m field */
+
+ emit_1ub(p, val);
+
+ switch (regmem.mod) {
+ case mod_REG:
+ case mod_INDIRECT:
+ break;
+ case mod_DISP8:
+ emit_1b(p, regmem.disp);
+ case mod_DISP32:
+ emit_1i(p, regmem.disp);
+ }
+}
+
+/* Many x86 instructions have two opcodes to cope with the situations
+ * where the destination is a register or memory reference
+ * respectively. This function selects the correct opcode based on
+ * the arguments presented.
+ */
+static void emit_op_modrm( struct x86_program *p,
+ GLubyte op_dst_is_reg,
+ GLubyte op_dst_is_mem,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ switch (dst.mod) {
+ case mod_REG:
+ emit_1ub(p, op_dst_is_reg);
+ emit_modrm(p, dst, src);
+ break;
+ case mod_INDIRECT:
+ case mod_DISP32:
+ case mod_DISP8:
+ assert(src.mod == mod_REG);
+ emit_1ub(p, op_dst_is_mem);
+ emit_modrm(p, src, dst);
+ break;
+ }
+}
+
+static void emit_mov( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_op_modrm( p, 0x8b, 0x89, dst, src );
+}
+
+static void emit_xor( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_op_modrm( p, 0x33, 0x31, dst, src );
+}
+
+static void emit_movlps( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_1ub(p, X86_TWOB);
+ emit_op_modrm( p, 0x12, 0x13, dst, src );
+}
+
+static void emit_movhps( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_1ub(p, X86_TWOB);
+ emit_op_modrm( p, 0x16, 0x17, dst, src );
+}
+
+static void emit_movd( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_2ub(p, 0x66, X86_TWOB);
+ emit_op_modrm( p, 0x6e, 0x7e, dst, src );
+}
+
+static void emit_movss( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_2ub(p, 0xF3, X86_TWOB);
+ emit_op_modrm( p, 0x10, 0x11, dst, src );
+}
+
+static void emit_movaps( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_1ub(p, X86_TWOB);
+ emit_op_modrm( p, 0x28, 0x29, dst, src );
+}
+
+static void emit_movups( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_1ub(p, X86_TWOB);
+ emit_op_modrm( p, 0x10, 0x11, dst, src );
+}
+
+/* SSE operations often only have one format, with dest constrained to
+ * be a register:
+ */
+static void emit_mulps( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_2ub(p, X86_TWOB, 0x59);
+ emit_modrm( p, dst, src );
+}
+
+static void emit_addps( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_2ub(p, X86_TWOB, 0x58);
+ emit_modrm( p, dst, src );
+}
+
+static void emit_cvtps2dq( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_3ub(p, 0x66, X86_TWOB, 0x5B);
+ emit_modrm( p, dst, src );
+}
+
+static void emit_packssdw( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_3ub(p, 0x66, X86_TWOB, 0x6B);
+ emit_modrm( p, dst, src );
+}
+
+static void emit_packsswb( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_3ub(p, 0x66, X86_TWOB, 0x63);
+ emit_modrm( p, dst, src );
+}
+
+/* Load effective address:
+ */
+static void emit_lea( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src )
+{
+ emit_1ub(p, 0x8d);
+ emit_modrm( p, dst, src );
+}
+
+static void emit_add_imm( struct x86_program *p,
+ struct x86_reg dst,
+ struct x86_reg src,
+ GLint value )
+{
+ emit_lea(p, dst, make_disp(src, value));
+}
+
+
+
+
+/**
+ * Perform a reduced swizzle:
+ */
+static void emit_pshufd( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0,
+ GLubyte x,
+ GLubyte y,
+ GLubyte z,
+ GLubyte w)
+{
+ emit_3ub(p, 0x66, X86_TWOB, 0x70);
+ emit_modrm(p, dest, arg0);
+ emit_1ub(p, (x|(y<<2)|(z<<4)|w<<6));
+}
+
+
+static void emit_pk4ub( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_cvtps2dq(p, dest, arg0);
+ emit_packssdw(p, dest, dest);
+ emit_packsswb(p, dest, dest);
+}
+
+static void emit_load4f_4( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movups(p, dest, arg0);
+}
+
+static void emit_load4f_3( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ /* Have to jump through some hoops:
+ *
+ * 0 0 0 1 -- skip if reg[3] preserved over loop iterations
+ * c 0 0 1
+ * 0 0 c 1
+ * a b c 1
+ */
+ emit_movups(p, dest, get_identity(p));
+ emit_movss(p, dest, make_disp(arg0, 8));
+ emit_pshufd(p, dest, dest, Y,Z,X,W );
+ emit_movlps(p, dest, arg0);
+}
+
+static void emit_load4f_2( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ /* Pull in 2 dwords, then copy the top 2 dwords with 0,1 from id.
+ */
+ emit_movlps(p, dest, arg0);
+ emit_movhps(p, dest, get_identity(p));
+}
+
+static void emit_load4f_1( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ /* Initialized with [0,0,0,1] from id, then pull in the single low
+ * word.
+ */
+ emit_movaps(p, dest, get_identity(p));
+ emit_movss(p, dest, arg0);
+}
+
+
+
+static void emit_load3f_3( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ /* Over-reads by 1 dword - potential SEGV... Deal with in
+ * array_cache by treating size-3 arrays specially, copying to
+ * temporary storage if last element (how can you tell?) falls on a
+ * 4k boundary.
+ */
+ if (p->inputs_safe) {
+ emit_movaps(p, dest, arg0);
+ }
+ else {
+ /* c . . .
+ * c c c c
+ * a b c c
+ */
+ emit_movss(p, dest, make_disp(arg0, 8));
+ emit_pshufd(p, dest, dest, X,X,X,X);
+ emit_movlps(p, dest, arg0);
+ }
+}
+
+static void emit_load3f_2( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_load4f_2(p, dest, arg0);
+}
+
+static void emit_load3f_1( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_load4f_1(p, dest, arg0);
+}
+
+static void emit_load2f_2( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movlps(p, dest, arg0);
+}
+
+static void emit_load2f_1( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_load4f_1(p, dest, arg0);
+}
+
+static void emit_load1f_1( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movss(p, dest, arg0);
+}
+
+static void (*load[4][4])( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 ) = {
+ { emit_load1f_1,
+ emit_load1f_1,
+ emit_load1f_1,
+ emit_load1f_1 },
+
+ { emit_load2f_1,
+ emit_load2f_2,
+ emit_load2f_2,
+ emit_load2f_2 },
+
+ { emit_load3f_1,
+ emit_load3f_2,
+ emit_load3f_3,
+ emit_load3f_3 },
+
+ { emit_load4f_1,
+ emit_load4f_2,
+ emit_load4f_3,
+ emit_load4f_4 }
+};
+
+static void emit_load( struct x86_program *p,
+ struct x86_reg temp,
+ GLuint sz,
+ struct x86_reg src,
+ GLuint src_sz)
+{
+ load[sz-1][src_sz-1](p, temp, src);
+}
+
+
+static void emit_store4f( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movups(p, dest, arg0);
+}
+
+static void emit_store3f( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ if (p->outputs_safe) {
+ /* Emit the extra dword anyway. This may hurt writecombining,
+ * may cause other problems.
+ */
+ emit_movups(p, dest, arg0);
+ }
+ else {
+ /* Alternate strategy - emit two, shuffle, emit one.
+ */
+ struct x86_reg tmp = get_sse_temp(p);
+ emit_movlps(p, dest, arg0);
+
+ emit_pshufd(p, tmp, arg0, Z, Z, Z, Z );
+ emit_movss(p, make_disp(dest,8), tmp);
+ release_temp(p, tmp);
+ }
+}
+
+static void emit_store2f( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movlps(p, dest, arg0);
+}
+
+static void emit_store1f( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 )
+{
+ emit_movss(p, dest, arg0);
+}
+
+
+static void (*store[4])( struct x86_program *p,
+ struct x86_reg dest,
+ struct x86_reg arg0 ) =
+{
+ emit_store1f,
+ emit_store2f,
+ emit_store3f,
+ emit_store4f
+};
+
+static void emit_store( struct x86_program *p,
+ struct x86_reg dest,
+ GLuint sz,
+ struct x86_reg temp )
+
+{
+ store[sz-1](p, dest, temp);
+}
+
+
+static GLint get_offset( const void *a, const void *b )
+{
+ return (const char *)b - (const char *)a;
+}
+
+
+
+/* Lots of hardcoding
+ *
+ * EAX -- pointer to current output vertex
+ * ECX -- pointer to current attribute
+ *
+ */
+static GLboolean build_vertex_emit( struct x86_program *p )
+{
+ GLcontext *ctx = p->ctx;
+ TNLcontext *tnl = TNL_CONTEXT(ctx);
+ struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
+ struct tnl_clipspace_attr *a = vtx->attr;
+ GLuint j;
+
+ struct x86_reg vertexEAX = make_reg(file_REG32, reg_AX);
+ struct x86_reg srcEDI = make_reg(file_REG32, reg_CX);
+ struct x86_reg countEBP = make_reg(file_REG32, reg_BP);
+ struct x86_reg vtxESI = make_reg(file_REG32, reg_SI);
+ struct x86_reg tmp = make_reg(file_XMM, 0);
+ struct x86_reg vp0 = make_reg(file_XMM, 1);
+ struct x86_reg vp1 = make_reg(file_XMM, 2);
+ GLubyte *fixup, *label;
+
+ p->csr = p->store;
+
+ /* Push a few regs?
+ */
+ emit_push(p, srcEDI);
+ emit_push(p, countEBP);
+ emit_push(p, vtxESI);
+
+ /* Initialize destination register.
+ */
+ emit_mov(p, vertexEAX, make_fn_arg(p, 3));
+
+ /* Dereference ctx to get tnl, then vtx:
+ */
+ emit_mov(p, vtxESI, make_fn_arg(p, 1));
+ emit_mov(p, vtxESI, make_disp(vtxESI, get_offset(ctx, &ctx->swtnl_context)));
+ vtxESI = make_disp(vtxESI, get_offset(tnl, &tnl->clipspace));
+
+ /* Get vertex count, compare to zero
+ */
+ emit_mov(p, countEBP, make_fn_arg(p, 2));
+ fixup = emit_jcc_forward(p, cc_NZ);
+
+ /* Possibly load vp0, vp1 for viewport calcs:
+ */
+ if (vtx->need_viewport) {
+ emit_movups(p, vp0, make_disp(vtxESI, get_offset(vtx, &vtx->vp_scale[0])));
+ emit_movups(p, vp1, make_disp(vtxESI, get_offset(vtx, &vtx->vp_xlate[0])));
+ }
+
+ /* Note address for loop jump */
+ label = get_label(p);
+
+ /* Emit code for each of the attributes. Currently routes
+ * everything through SSE registers, even when it might be more
+ * efficient to stick with regular old x86. No optimization or
+ * other tricks - enough new ground to cover here just getting
+ * things working.
+ */
+ for (j = 0; j < vtx->attr_count; j++) {
+ struct x86_reg dest = make_disp(vertexEAX, vtx->attr[j].vertoffset);
+ struct x86_reg ptr_to_src = make_disp(vtxESI, get_offset(vtx, &vtx->attr[j].inputptr));
+
+ /* Load current a[j].inputptr
+ */
+ emit_mov(p, srcEDI, ptr_to_src);
+
+ /* Now, load an XMM reg from src, perhaps transform, then save.
+ * Could be shortcircuited in specific cases:
+ */
+ switch (a[j].format) {
+ case EMIT_1F:
+ emit_load(p, tmp, 1, srcEDI, vtx->attr[j].inputsize);
+ emit_store(p, dest, 1, tmp);
+ case EMIT_2F:
+ emit_load(p, tmp, 2, srcEDI, vtx->attr[j].inputsize);
+ emit_store(p, dest, 2, tmp);
+ case EMIT_3F:
+ /* Potentially the worst case - hardcode 2+1 copying:
+ */
+ emit_load(p, tmp, 3, srcEDI, vtx->attr[j].inputsize);
+ emit_store(p, dest, 3, tmp);
+ case EMIT_4F:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_store(p, dest, 4, tmp);
+ break;
+ case EMIT_2F_VIEWPORT:
+ emit_load(p, tmp, 2, srcEDI, vtx->attr[j].inputsize);
+ emit_mulps(p, dest, vp0);
+ emit_addps(p, dest, vp1);
+ emit_store(p, dest, 2, tmp);
+ break;
+ case EMIT_3F_VIEWPORT:
+ emit_load(p, tmp, 3, srcEDI, vtx->attr[j].inputsize);
+ emit_mulps(p, dest, vp0);
+ emit_addps(p, dest, vp1);
+ emit_store(p, dest, 3, tmp);
+ break;
+ case EMIT_4F_VIEWPORT:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_mulps(p, dest, vp0);
+ emit_addps(p, dest, vp1);
+ emit_store(p, dest, 4, tmp);
+ break;
+ case EMIT_3F_XYW:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pshufd(p, tmp, tmp, X, Y, W, Z);
+ emit_store(p, dest, 3, tmp);
+ break;
+
+ /* Try and bond 3ub + 1ub pairs into a single 4ub operation?
+ */
+ case EMIT_1UB_1F:
+ case EMIT_3UB_3F_RGB:
+ case EMIT_3UB_3F_BGR:
+ return GL_FALSE; /* add this later */
+
+ case EMIT_4UB_4F_RGBA:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pk4ub(p, tmp, tmp);
+ emit_store(p, dest, 1, tmp);
+ break;
+ case EMIT_4UB_4F_BGRA:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pshufd(p, tmp, tmp, Z, Y, X, W);
+ emit_pk4ub(p, tmp, tmp);
+ emit_store(p, dest, 1, tmp);
+ break;
+ case EMIT_4UB_4F_ARGB:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pshufd(p, tmp, tmp, W, X, Y, Z);
+ emit_pk4ub(p, tmp, tmp);
+ emit_store(p, dest, 1, tmp);
+ break;
+ case EMIT_4UB_4F_ABGR:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pshufd(p, tmp, tmp, W, Z, Y, X);
+ emit_pk4ub(p, tmp, tmp);
+ emit_store(p, dest, 1, tmp);
+ break;
+ case EMIT_4CHAN_4F_RGBA:
+ switch (CHAN_TYPE) {
+ case GL_UNSIGNED_BYTE:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_pk4ub(p, tmp, tmp);
+ emit_store(p, dest, 1, tmp);
+ break;
+ case GL_UNSIGNED_SHORT:
+ return GL_FALSE;
+ case GL_FLOAT:
+ emit_load(p, tmp, 4, srcEDI, vtx->attr[j].inputsize);
+ emit_store(p, dest, 4, tmp);
+ break;
+ default:
+ break;
+ }
+ default:
+ return GL_FALSE; /* catch any new opcodes */
+ }
+
+ /* add a[j].inputstride (hardcoded value - could just as easily
+ * pull the stride value from memory each time).
+ */
+ emit_add_imm(p, srcEDI, srcEDI, a[j].inputstride);
+
+ /* save new value of a[j].inputptr
+ */
+ emit_mov(p, ptr_to_src, srcEDI);
+
+ }
+
+ /* Next vertex:
+ */
+ emit_add_imm(p, vertexEAX, vertexEAX, vtx->vertex_size);
+
+ /* decr count, loop if not zero
+ */
+ emit_dec(p, countEBP);
+ emit_jcc(p, cc_NZ, label);
+
+ /* Land forward jump here:
+ */
+ do_fixup(p, fixup);
+
+ /* Pop regs and return
+ */
+ emit_pop(p, vtxESI);
+ emit_pop(p, countEBP);
+ emit_pop(p, srcEDI);
+ emit_ret(p);
+
+ vtx->emit = (tnl_emit_func)p->store;
+ return GL_TRUE;
+}
+
+void _tnl_generate_sse_emit( GLcontext *ctx )
+{
+ struct tnl_clipspace *vtx = GET_VERTEX_STATE(ctx);
+ struct x86_program p;
+
+ memset(&p, 0, sizeof(p));
+ p.ctx = ctx;
+ p.store = MALLOC(1024);
+
+ p.inputs_safe = 1; /* for now */
+ p.outputs_safe = 1; /* for now */
+
+ if (build_vertex_emit(&p)) {
+ _tnl_register_fastpath( vtx, GL_TRUE );
+
+ {
+ static int i = 0;
+ char filename[100];
+ int fd;
+
+ sprintf(filename, "fastpath%d.o", i);
+ fd = creat(filename, 0600);
+ if (fd != -1) {
+ write(fd, p.store, p.csr - p.store);
+ close(fd);
+ _mesa_printf("wrote %s\n", filename);
+ }
+ }
+ }
+ else {
+ FREE(p.store);
+ }
+
+ (void)emit_movd;
+ (void)emit_inc;
+ (void)emit_xor;
+}